- August 2019
- Dec 11, 2018 AlphaGoZero - A Reflection and a Concern Dec 11, 2018
- Nov 28, 2018 The Merits of Independent Audit of AI Systems Nov 28, 2018
- Aug 23, 2018 Universal Basic Income, Capitalism and Christianity - Can We Reconcile the Three Aug 23, 2018
- Jul 23, 2018 Technology is an Autocracy - and the Risk from Externalities is Growing Jul 23, 2018
- Jun 12, 2018 The Value of Data in the Digital Age Jun 12, 2018
- Apr 23, 2018 Facebook and Cambridge Analytica - "Know Your Customer", A Higher Standard Apr 23, 2018
- Mar 22, 2018 New C-Suite Position - AI and Automation Risk Management Mar 22, 2018
- Feb 8, 2018 Personal Data - How to Control Info and Get Paid For Being You Feb 8, 2018
- Jan 9, 2018 ForHumanity AI and Automation Awards 2017 Jan 9, 2018
- Dec 1, 2017 AI and Automation - Managing the Risk to Reap the Reward Dec 1, 2017
- Nov 29, 2017 Ban Autonomous AI Currency/Capital Usage Nov 29, 2017
- Oct 30, 2017 Robot Tax (No), Sovereign Wealth Fund (Yes) Oct 30, 2017
- Oct 9, 2017 Universal Basic Income Isn't Perfect - It's a Necessity Because of the Future of Work Oct 9, 2017
- Sep 22, 2017 AI Safety - The Concept of Independent Audit Sep 22, 2017
- Sep 15, 2017 Drawing A Line - The Difference Between Medicine and Transhumanism Sep 15, 2017
- Sep 13, 2017 Replacing Your Doctors with AI and Automation Sep 13, 2017
- Sep 8, 2017 AI and Automation - Replacing YOUR Job Sep 8, 2017
- Sep 6, 2017 The Process of Technological Unemployment - How Will it Happen? Sep 6, 2017
- Aug 18, 2017 Tech In Your Body - Let The Seduction Begin Aug 18, 2017
- Aug 15, 2017 Privacy - GDPR for the US Aug 15, 2017
- Aug 15, 2017 Why Sales Jobs Will Be Automated Too... Aug 15, 2017
- Aug 12, 2017 Wake Up Call - Data Bias and Corporations Aug 12, 2017
- Aug 7, 2017 Tackling Bias in Machine Learning, AI and Humanity Aug 7, 2017
- Aug 5, 2017 Emotional Intelligence Mirage? Jobs and our Humanity Aug 5, 2017
- Aug 3, 2017 #changethediscourse Aug 3, 2017
- Jul 24, 2017 Technology + The Human Body = Insurmountable Societal Challenge? Jul 24, 2017
- Jul 18, 2017 SAFEAI - A Tool For Concerned Parents Jul 18, 2017
- Jul 9, 2017 The Amazon Impact on Retail Jobs - Defining Technological Unemployment Jul 9, 2017
- Jun 23, 2017 Universal Basic Income— A Working Road Map Part 2- Corporate Responsibility Jun 23, 2017
- Jun 19, 2017 Faith and Artificial Intelligence - Humanity's greatest schism Jun 19, 2017
- Jun 12, 2017 LIFETIME LEARNING - A Necessity Already Jun 12, 2017
- Jun 7, 2017 13 Things That UBI Can't Replace - A Reflection on Universal Basic Income Jun 7, 2017
- May 12, 2017 What Should the Partnership on AI become? May 12, 2017
- May 2, 2017 Losing a Piece of our Humanity - A Cost of Technological Self-Sufficiency May 2, 2017
- Apr 13, 2017 Transhumanism, Life Extension and my Concerns Apr 13, 2017
- Mar 31, 2017 Universal Basic Income (UBI) - A Working Road Map Mar 31, 2017
- Mar 31, 2017 Right to Privacy -2nd Bill of Rights Mar 31, 2017
- Mar 27, 2017 Challenges to Overcome for Universal Basic Income (UBI) Mar 27, 2017
- Mar 26, 2017 Right to Procreate - 2nd Bill of Rights Mar 26, 2017
- Mar 16, 2017 Strategies to Keep Your Job with AI & Automation Mar 16, 2017
- Mar 10, 2017 The Right to Mobility - Constitutional Amendment Mar 10, 2017
- Mar 1, 2017 #futureofwork why are we struggling to understand the disruption? Mar 1, 2017
- Feb 20, 2017 Living in the Age of Machines - Do we need a second Bill of Rights? Feb 20, 2017
- Feb 9, 2017 Answering Nick Bostrom's Desiderata (Part 4 of 4) Feb 9, 2017
- Feb 9, 2017 Answering Nick Bostrom's Desiderata (Part 3 of 4) Feb 9, 2017
- Feb 4, 2017 Answering Nick Bostrom's Desiderata (Part 2 of 4) Feb 4, 2017
- Jan 18, 2017 Jan 18, 2017
- Jan 14, 2017 Answering Nick Bostrom's Desiderata (Part 1 of 4) Jan 14, 2017
This article was a portion of ForHumanity’s response to the NYC Economic Development Corp’s Request for Expressions of Interest regarding their proposed Center for Responsible AI. The submission was a joint submission with The Future Society and Michelle Calabro. Independent Audit is offered, as the solution for the framwork, guidelines and governance functions of the Center. The proposal has neither been accepted or rejected at the time of this publication.
Independent Audit of AI Systems is a framework for auditing AI systems (products, services, and corporations) by examining and analyzing the downside risks associated with the ubiquitous advance of AI & Automation. It looks at 5 key areas: Privacy, Bias, Ethics, Trust and Cybersecurity. If we can make safe and responsible Artificial Intelligence & Automation profitable whilst making dangerous and irresponsible AI & Automation costly, then all of humanity wins.
The market demand for a comprehensive, implementable, third-party oversight and governance solution is enormous. There are regular and consistent calls for this type of comprehensive oversight, on a global scale. The words, ‘Trust, Assurance and Audit’ are frequently used to give humanity the comfort we need with these currently opaque and ungoverned systems. It will not only impact the world of AI in New York City, but around the world, thrusting New York into the center of all debate on Responsible AI.
Creation of the framework will be an industry-wide effort from the Audit/Assurance/Trust industry, and they will conduct an iterative, open-source, crowd-sourced dialogue with anyone who wants to engage in the process. The result will be a fully-implementable and dynamic set of rules by which companies can not only know they are being Responsible with their AI, they can prove it with an independent audit.
1.1 What is it?
A framework for auditing AI systems by examining and analyzing the downside risks associated with the ubiquitous advance of AI & Automation. It looks at 5 key areas: Privacy, Bias, Ethics, Trust and Cybersecurity.
1.2 What can get audited?
Products, services and corporations.
1.3 Why should it exist?
If we can make safe and responsible Artificial Intelligence & Automation profitable whilst making dangerous and irresponsible AI & Automation costly, then all of humanity wins.
1.4 Does the market want it?
The market demand for a comprehensive, implementable, third-party oversight and governance solution is enormous. There are regular and consistent calls for this type of comprehensive oversight, on a global scale. The words, ‘Trust, Assurance and Audit’ are frequently used to give humanity the comfort we need with these currently opaque and ungoverned systems.
1.5 Why must this happen in New York City?
Independent Audit of AI Systems is the audit/assurance/trust industry’s answer to AI governance and oversight. It is logical that it should reside in New York City, which is home to each of the Big 4’s global headquarters. Furthermore, it is crucial that an industry-wide initiative is implemented to maximize the impact of the audit rules.
1.6 How will New York City benefit from this?
We expect these rules to be adapted to local jurisdictional law in NYC and globally. This will thrust New York into the center of all debate on Responsible AI. Jobs are likely to be created as the audit process is deepened. New York City will be viewed at the leader of Responsible AI; the place to go for the most “cutting-edge” discussions. The place to go for answers.
1.7 How will humanity benefit from this?
Currently AI is generated to benefit companies, through increased profit. We are not criticizing this point-of-view, but we are suggesting that there needs to be a more balanced approach — one which considers the overall welfare of humanity. Through oversight and third-party independent governance, via Independent Audit of AI System, humanity will have a greater assurance that their interests are being considered. People can be assured that ‘best practices’ are being followed, which will increase peoples’ trust and confidence in AI systems wherever they impact people’s lives. Markets and systems will be more fair — not just in one country, as the audit rules will extend globally as well.
1.8 How will individuals benefit from this?
Individuals will benefit from better, safer products and services. They will benefit through increased education and awareness on what SAFEAI and Responsible AI mean. Individuals will be able to identify SAFEAI goods and services and increasingly have choice and control over how they manage their interactions with AIs in all facets of their lives.
1.9 Will Independent Audit of AI Systems be a self-sustainable system?
Independent Audit of AI Systems is designed to be self-sustaining over time. The SAFEAI logo will be licensed as a “seal-of-approval” to any/all companies that wish to license it. The choice is theirs, as there is not a fee to ForHumanity for the audit framework and the audits are not conducted by us.
As we develop the Independent Audit of AI Systems framework, we will spend considerable effort to build the brand as well. We want to facilitate people using the logo to impact their buying decisions for products and services. If successful, this will drive demand for the logo from those who pass audits. We will ask a licensing fee in return, with the justification being both the promotion of the brand and the ongoing maintenance of the system. The primary goal and thus the driver of licensing pricing will be the sustainability of the Independent Audit of AI systems process. It is expected that our initial donors will backstop the organization until it is self-sustaining.
2.0 Who will create it?
Creation of the framework will be an industry-wide effort from the Audit/Assurance/Trust industry, and they will conduct an iterative, open-source, crowd-sourced dialogue with anyone who wants to engage in the process. In order to be successful, it will be global and inclusive of many interested parties.
There will be a dedicated full-time staff of Core Audit Silo Heads whose responsibilities will be the following:
Research new ideas/concepts/contributors
Suggest possible rules
Solicit specific feedback on proposed rules
Track and manage dissent, striving for broad consensus
Manage discussion for decorum
Present proposed rules to Board of Directors
Hold public office for 2 hours daily, during global time zones
Facilitate dialogue on Slack with all interested participants
The Executive Director will have the following responsibilities:
Create partnerships with many like-minded stakeholders
Facilitate the quarterly update process and Board of Directors votes
Release the board votes and new rules
The Board of Directors (maximum 21 members) will consist of a diverse collection of industry professionals and academics who are experts in determining if a rule passes audit criteria.
2.1 How will the audit framework be developed?
Independent Audit of AI Systems will be a comprehensive set of best-practices covering the ‘Core Audit Silos’ of Ethics, Bias, Privacy, Trust and Cybersecurity.
Core Audit Silo Heads will host global public office hours during pre-announced, varying time slots to meet the needs of experts all around the world. They will also conduct an iterative, open-source, crowd-sourced dialogue on an ongoing basis (on a platform such as Slack) with anyone who wants to engage in the process. Corporations (who will be subject to the audit) have the opportunity to raise concerns and objections with the proposed rules. Independent Audit of AI Systems will walk the tightrope between maximum benefit for humanity and practical implementability.
Audit rules will be:
Open-Sourced and Iterated
The cohort of experts in each ‘Core Audit Silo’ will conduct micro-experiments, use-case experiments and applications to iterate and eventually achieve these standards and best practices. We will improve Responsible AI incrementally, and eventually arrive at consensus-driven rules by examining word choice, definitions, and boundaries.
The result will be a fully-implementable and dynamic set of rules by which companies can not only know they are being Responsible with their AI, they can prove it with an independent audit.
2.2 Hasn’t this been done before?
Yes. In fact, there are many precedents to learn from. The Independent Audit of AI Systems is a collaborative process designed to enhance the magnificent work already being done by experts and thought-leaders around the world.
We are seeking an outcome similar to that of the FASB and IFRS efforts in the early 70’s: the financial services industry came together and created standardized rules that were adopted by the SEC et al within a few short years. The growth comes from a level playing field, like we watched with financial accounting in the late 70’s and early 80’s.
Many groups are doing Ethical Guidelines work; the IEEE’s ‘Ethically Aligned Design’ remains the gold standard. With a singular mission to leave the world a little better today than it was yesterday, we will translate high-minded ideals into smaller, more practical baby steps toward implementing new rules. Our goal can be achieved with extensive collaboration from IEEE, NIST and others to adapt their work into an immediately implementable, auditable framework.
2.3 Once an AI System achieves SAFEAI status, what happens?
All participants will know how to be responsible with their AI.
Passing the Audit is good for 12 months.
Even as the rules change quarterly, companies have time to catch up as new ‘best-practices’ are implemented.
Compliant companies may choose to license the SAFEAI brand.
Rocker tags include Auditor, Services, Corporate, and Product.
Annotated as e.g. “SAFEAI 2019.
This post is a follow on to the previous posts on Independent Audit of AI Systems, both on Medium
The Merits of Independent Audit of AI Systems
For those who have followed ForHumanity, you know that we have been developing and promoting the concept of Independent…
and on the CACM blog
Governance and Oversight Coming to AI and Automation: Independent Audit of AI Systems
Governance and independent oversight on the design and implementation of all forms of artificial intelligence (AI) and…
We welcome all of your support, as we continue to push this concepts forward and work towards bringing oversight and governance with true accountability to the world of AI.
AlphaGoZero from DeepMind
The AI community remains “abuzz”, as DeepMind continues to announce AlphaGo Zero successes. For the uninitiated, DeepMind created a new AI/machine learning program designed to improve upon its first iteration, AlphaGo, which had successfully bested the human world champions at the ancient game of Go. Why the buzz? Well AlphaGo Zero was built “Tabula Rasa” which is the philospher Locke’s definition of the human mind at birth and it means “blank slate”. The orginal AlphaGo studied human games of Go over and over and over, millions of iterations in order to learn how to win at the game. Zero (that’s how we will refer to AlphaGo Zero, so that it is clearer) learned Tabula Rasa, meaning it was simply given the rules and objectives and taught itself. 2.5 days later it surpassed the best players in the world. 21 days later it surpassed the best AI player in AlphaGo. 40 days later it was winning 100 times out of 100. That is the speed at which “mastery of thought or narrow superintelligence” (in the game of Go) was achieved. 40 days and consider that the test was done without trying to maximize processing power and speed. Recently, the same AlphaGoZero program also mastered chess and Shogi.
Okay, that all sounds impressive, but what does it really mean. For a start here are a few disruptive things it may mean.
The designers at DeepMind have quickly convinced themselves that learning from humans is suboptimal
In this narrow space of intelligence (the game of Go, Chess and Shogi)— this is proof that learning from humans IS suboptimal
Also, in this narrow space, SuperIntelligence has been achieved
Since some believe that learning is learning (a big assumption) — why wouldn’t all learning by machines be accomplished without human input.
We now have an argument for WHY some people will want to implement Artificial General Intelligence. Or said differently, “don’t we want SuperIntelligence in as many areas as possible”?
Let me make sure that is clear for the reader. Unless someone can explain to the world how and why certain “learning” is DIFFERENT from learning the game of Go, Chess or Shogi, the brilliant minds at DeepMind have just proven to you that “learning” from humans is inefficient. That machines, when given a task to learn, will achieve human level mastery and beyond, quickly. This information will be used to demonstrate flawed human thinking time and again —the argument for replacing human thinking.
Said another way — AlphaGo Zero is PROOF that human thinking/learning is suboptimal. Let that sink in… What are some implications of THAT concept:
How fast can I get tech into my head? some will reach this conclusion as they accept the idea that human thinking is suboptimal and to remain competitive, they must “advance” as well
What is the point of human thought? another conclusion many will reach. Just let the machine do it.
Now before we fall too far off the cliff, let me be clear to say that I am certain that machine intelligence will discover things, learn things and create things that were simply not possible using the human mind alone. Those developments will drive the next few centuries of economic growth and create substantial wealth and opportunity (who that wealth accrues to is an entirely different debate, but to believe it would be equally shared is naive). As a Capitalist and non-Luddite, I believe in the advancement of these technologies and I believe they will succeed faster than most people have anticipated. The future for intelligence, discovery, productivity and exploration is vast and exciting. I am a fan.
But is that everything, is it even optimal? Seems to me that we are leaving out a lot of key issues when we measure ourselves in the “better off” category. I am not sure if these advancement are for the betterment of society. That’s the key question for me. Why are we advancing society? What about society is advancing? I know the word advancing is “loaded”, as in… of course we want to “advance” society, we should always “advance” things. We always assume that being smarter, acquiring more wealth and making everything easier is better. I think this is a falsehood, missing the mark on our total well-being.
Here are some things that we have “advanced” because of technology:
Lack of individual survivability
Polarization of society
Breakdown of community
De-Valuing of people and human life
Worship of Intelligence
General drop in our physical fitness
Loss of Faith
Of course those are only some of the negatives and yet we won’t even be able to agree that those are, in fact, all negatives. I am highlighting the point that we continue to use technology to “advance” but our measures of well-being are not being maintained and tested against the technological advancements to determine if we are ACTUALLY better off. We feel more productive - our companies certainly are. But do we have deeper, stronger relationships? Do we experience more love, contentment and joy? I am not sure that we do, so when measured properly, maybe we have stepped back decades because of technology and didn’t even realize it. That is difficult for most people to even contemplate and their knee-jerk reaction is “of course things are better” or “those negatives don’t apply to me” . I am not sure that a fair, considered, and comprehensive assessment of well-being reaches the same conclusion. John Havens, a friend at IEEE, is leading the charge on this discussion and I think it is fruitful. This Youtube link, is a good overview of his thoughts on Well-being.
On top of all of this — the abdication of thought, learning and intellectual growth to machines combined with utter reliance upon machines for physical work can be scary. I do agree that we may triumph over many of these challenges and continue to master technology for the betterment of society as a whole. However it is important to raise these concerns and to consider if we are measuring progress and advancement correctly.
I think the author Frank Hebert, who created the Dune books and franchise was on to something as he considered his future view of where technological progress and our culture was taking us.
In his Sci-Fi classic Series Dune, specifically the book, God Emperor of Dune(1981), Leto II Atreides indicates that the Butlerian Jihad had been a semi-religious social upheaval initiated by humans who felt repulsed by how guided and controlled they had become by machines: “The target of the Jihad was a machine-attitude as much as the machines,” Leto said. “Humans had set those machines to usurp our sense of beauty, our necessary selfdom out of which we make living judgments. Naturally, the machines were destroyed.”
My greatest concern about the progress of artificial intelligence and automation is that the pendulum will have to swing too far before we figure out there is a problem. The reason that quote above exists, albeit from a story of ficition, is that Frank Herbert could envision a machine-intelligence dominated society. A Jihad (which is a violent clash often associated with the passion of religious fervor), not a revival, not a rennaissance, not a shifting-of-gears, was required to stop the juggernaut that was technology and machine intelligence in Herbert’s fictional world. Works of fiction are not evidence, they aren’t even an argument, but they may inform the thoughtful on a possible outcome from excessive adoption of machine intelligence.
Moderation and comprehensive evaluation of progress has never been a trait of our species, therefore it is likely that our adoption of AI and Automation will be excessive, and, in the end, detrimental to humanity. The prescription for avoiding these issues is challenging. It is against our general nature and it may require us to break our obsession with technological “advancement” as it is currently measured. It requires a broader consensus on well-being to measure our prosperity and to decide how and when to replace ourselves with machine intelligence. AlphaGoZero, and other AIs like it, will convince many that they should replace human intelligence. Personally, I can’t imagine replacing the vast majority of human intelligence with machine intelligence, but it is the path we are on and it leaves me concerned. Are you?
For those who have followed ForHumanity, you know that we have been developing and promoting the concept of Independent Audit of AI systems. If you are new to the concept of Independent Audit, the basics can be found here:
For this post, I wanted to cover the benefits of Indepedent Audit which are maximized when four things occur:
Corporations widely adopt Independent Audit
Governments make Independent Audit mandatory, akin to the requirements related to GAAP or IFRS accounting principles
AI Safety professionals widely participates in the open-source, crowd-sourced search for best-practices
When consumers of product/services use the SAFEAI logo to inform their buying/usage decisions
It is under these assumptions that I will talk about the features and subsequent benefits which accrue to humanity. Below is a bullet point list:
Transparency — consumers of both products and services will receive an unprecendent level of transparency from companies. The transparency/disclosure we refer to here is in a few key areas, such as ethical decision-making, data usage, safety, control, explainability of algorithms, accountability of both algorithms and corporations, and bias avoidance. These areas of increased Transparency result in…
Fairer markets — when markets are transparent then decisions made in the marketplace will be better. Rewarding responsible companies and punishing irresponsible companies. To be clear, it is the market that rewards and punishes, based on transparency and choice. And…
Trust — this is the foundation of fairer markets. A marketplace which has become increasingly adversarial between companies and consumers, especially in the capture and exploitation of data, can be reversed. Trust is engendered when consumers feel that they are being provided with a valuable service AND when the price for that service is considered fair. Price, in this case, includes not only monetary compensation for services, but also impacts to privacy and personal well-being.
Opacity — This benefit accrues to corporations, but has an over-arching, related benefit, to society. Opacity is the opposite of Transparency (referenced above), so we must explain the why both are listed as features. Transparency is discussed above, but does not extend to the intellectual property of the company — specifically, the code and machines that are employing artifical intelligence. Opacity to corporations allows them to protect their intellectual property (IP). When IP can be protected, then companies will invest in products and service development, knowing that they can recoop their costs and earn a fair profit. When transparency is EXCESSIVE (often due to regulation), then investment is discouraged and in fact, cheating, copying and outright theft become commonplace. Choosing Independent Audit allows a company to protect its IP without excessive disclosure. Choosing Independent Audit, might offset the increasing call for comprehensive disclosure that legal authorities might mandate. Essentially, Independent Audit strikes the best balance between opacity and transparency, giving corporations the protection they need to justify investment while giving consumers the transparency they need to trust the companies which provide them products/services.
-Third-party verification — The value of third-party verification is an under-valued element in our world today. There are many places, where third-party verification has dramatically raised the bar of quality. Starting with financial audits and ranging to product-testers, such as Consumer Reports. These behind-the-scenes services act as a watch dog on your behalf. There are costs associated with all independent third-party reviews and those costs are passed thru to you, the consumer. The result is that companies know that they cannot cheat, they cannot cut corners, they cannot act unethically or they will be called out. These systems are not perfect for various reasons, not the least of which is a belief by humans that they can “get away with it”, but in the end, the truth will always come out sooner-or later as long as there is a watchdog in the room.
Extends beyond national boundaries — Independent Audit is a market based mechanism. The SAFEAI seal of approval will build brand recognition. Over time, it will effect consumer decision-making, causing consumers to choose products and services which have been auditted for SAFEAI. The impacts of this “seal-of-approval” are not subject to national boundries the way that laws are hindered in their effectiveness. We are not suggesting that countries should not pass laws and regulations in AI Safety, quite to the contrary, new laws and regulations will have great value. However, it is a corporation’s responsbility to avoid regulations and laws when it is legal to do so and will result in greater profitability for the company. Therefore, it is necessary to have a market-based mechanism, one that impacts profitability directly and globally, such as the Independent Audit of AI Systems and the SAFEAI seal-of-approval. When consumers use the SAFEAI logo to inform their purchase and usage decisions then we know that safe and responsible AI will be profitable while dangerous and irresponsible AI will effefcctively make a company’s products worthless. When this happens, humanity wins, but it requires humanity to participate and to pay attention to the SAFEAI logo.
On-going and Dynamic process — unlike standard financial accounting principles which may go unchanged for years at a time, “best-practices” in ethics, bias, privacy, trust and cybersecurity are likely to change regulalry. ForHumanity will maintain a dynamic, transparent and constant review process with our global open-source, crowd-sourced network to continuingly uncover “best-practices” which will update our audit process quarterly.
Transparent process — The globally, open-source, crowd-sourced process is open to all. Anyone may join the conversation. All will be heard, all input will be considered, all votes count and most importantly, all reasonable dissent will be tracked and addressed. As the audit process is created, the results will be transparent to all. Further comment and critique is encouraged so that we may refine our process to achieve the best possible results. There is no fee or membership required to participate in the input process. There are only rules of proper decorum, so that we conduct our business in a civilized manner that protects the rights and dignity of all involved.
Dedicated professionals- ForHumanity maintains a full-time staff dedicated to each of the core audit silos of ethics, bias, privacy, trust and cybersecurity. As a result, these professionals are constantly sourcing new ideas and new contacts to uncover the “best-practices” for our audit process. They maintain daily office hours and are continually reviewing input from all over the world. Our professionals are able to bridge the gap between academic thought and practical application to deliver the dynamic audit process.
Audit process may be tailored for local jurisdictions — not dissimilar to the way accounting principles matured overtime, ForHumanity expects that the audit process will become tailored. Laws, customs and regulations may dictate that the audit process be adjusted slightly on a country-by-country (or regional) basis. This will allow the process to achieve an over-arching comment on “best-practices” while being locally compliant.
Audit is good for one year — Independent Audit is conducted annually using the current audit process for “best-practices”. As these “best-practices” are likely to change fairly regularly as technology advances, this could create an onerous process of compliance for companies. The most recent audit will be good for one year, regardless of changes in “best-practices”. This standard allows the company time to comply with the ever changing “best-practices” while permitting the audit process to remain dynamic.
Results are binary- When a company submits to a SAFEAI audit, it is a pass/fail endeavor on each of the five audit silos. The process will be structured so that companies are either compliant or non-compliant, so that the consumer may know if a company is using “best-practices” for ethics, bias, privacy, trust and cybersecurity. A company may pass any number of the silos, but is not SAFEAI compliant unless all five silos are passed successfully.
Goal-Aligned with Humanity. ForHumanity exists for one purpose, to mitigate the downside risks associated with AI and Automation in order for Humanity to receive the highest possible benefits from these technological advancements. Our client-base, is you. ForHumanity is a non-profit organization designed to serve those who would buy products or use services which rely upon artificial intelligence. ForHumanity has a few sources of revenue. First, we receive funding from individuals through gracious donations. Second, we receive licensing revenue from corporations who have already completed SAFEAI audits when they choose to license the SAFEAI logo to show the world. The company does NOT pay a fee to ForHumanity for the audit and never will. We will not have the audit results tainted by a “pay-for-play” process which incentivizes successful audits. Only once they have successfully complied with the audit, then may a company pay to license the SAFEAI logo to demonstrate successfully passing the audit. Other members pay membership fees if they want to serve humanity, but fixing problems at companies that want to be SAFEAI compliant. Finally, our core members are those which conduct the audits on behalf of ForHumanity and thus for you. In the end, we strive to ensure that the goals of ForHumanity are prefectly aligned to serve… humanity.
ForHumanity believes that Independent Audit of AI Systems is crucial to mitigating the downside risks associated with the proliferation of AI and automation. But it doesn’t exist in a vacuum and it doesn’t succeed without the help of everyone. Please consider how you can become invovled and reach out.
I was born and raised in the West, steeped in Capitalism, market economies and the power of supply and demand. As I began to consider the concept of Technological Unemployment, I wrote about Capitalism having an “end-game”.
I believe that remains true. I believe that if left to its own devices, with technological advancement, Capital (as in Capital v Labor) would choose to eliminate the labor from its cost equation, resulting in 100% of profit left for Capital. You might argue that 100% capital and 0% labor is too extreme, and I agree. There will always be roles/work for humans to do, based upon the skills that humanity retains which machines cannnot replicate, even if that is limited to their “humanness”. In this piece, I am using the extreme example to highlight a risk, not predict an exact future. Capital is incentivized to eliminate labor from its cost structure. AI and Automation are capital investments that can replace labor therefore, I expect Capital to increase investment in AI and Automation which will likley result in significant unemployment, at least as it relates to jobs that pay a salary.
To complete the ideological triumvirate, I was raised and subsequently chose to be a Christian, which defines the core of my morality. I am not asking you to agree with my morality, just understand that my moral choices, come from this background, as I try to reconile these concepts. With that as foundation, I decided to host a backyard BBQ, where the pre-announced topic was Universal Basic Income, Christianity and Capitalism, reconciling the three ideologies. I invited good friends and was not attempting to make this a “comprehensive and stastically significant focus-group”, instead I wanted to just talk and debate and see if we could learn a few things and achieve some level of consensus. It was a lovely dinner, the talk and questions were challenging and while we wandered a little bit into the weeds, as all good conversations tend to, we actually did find some key points upon which we generally agreed, even if the details remained a little debateable or ambiguosly defined.
So I present for your consideration the results of this discussion. It should be noted, that the crux of the discussion was about UBI and thus what follows is a discussion about UBI, influenced by our similar capitalistic (western) backgrounds and by our shared Christian-faith. I believe this can be a useful guide for others as to how we considered some of the challenges presented by these three ideologies and where we landed. I do not expect that all will exactly share these beliefs, but rather take this as one version of the discussion for you to consider.
A few bullet points:
- Belief that we have a moral responsibility, as a community to care for the poor and those who cannot take care of themselves. This is absolute and a core principle based on our Christian faith.
- Belief that “risk and reward are linked, greater risk should equate to greater potential reward and vice versa” is a bibilical concept. It need not apply only to money and capital, but in the Parable of the Talents, failure to “invest” Talents is considered sinful. This was discussed in the context of all behavior. Taking risk, deserves reward, but may also lead to failures, which is okay. Our understanding of the parable is that we should take risks with the assets that we have - we should invest. The group voiced a concern that UBI may lead to risk-averse behavior of all kinds, notably a lack of investment. Many UBI proponents talk out of both sides of their mouth on this subject which is why we spent time on it. On one hand, they criticize those who have taken great risk, sometimes with time, effort, work/life balance sacrifice, capital or even reputation, instead frequently attributing it to inheritence or unfair exploitation. Then they suggest that a UBI will lead all people to be more entreprenuerial because their downside risk is floored with the UBI, in other words, they will take risks. Either risk and reward are linked at all levels or they are not. You can’t reward “UBI entreprenuers” with profits and begrudge the wealthy who may have already earned their profits. Not to mention those middle to upper class members who just plain worked ridiculously hard. Something that used to be called the “American Way”. The group felt that a UBI, on a mass scale, would reduce the appetite for risk amongst the mass population, even if a few were emboldened. They did not accept the premise that UBI would lead to greater entrepreneurialism.
- Belief that work and participation in your own survival is a human responsibility both to yourself and to your community. The group did not believe in a “right to survive”. They support the “right to participate in your own survival”. The group believes that the community is responsible for caring for those who are “unable to participate in their own survival”. This might be a semantic argument, but the point for us was clear. Survival is not guaranteed, it must be worked for and that is the nature of life. In fact, the idea that anyone had a guaranteed right to survive was generally considered illogical.
- The group did not require Universal to mean that 100% of people must receive the benefit fully. They were supportive of the idea that high income earners could have their basic income effectively fully taxed, which of course reduces the cost of implementation. The group felt that it should be “means-tested” on both ends. The wealthy should be taxed on their UBI to lower the cost of the program. But on the receiveing end, all should work, who are able. This is a moral decision based in the belief that providing for ourselves, our family and our communities is our responsibility. They further felt it was appropriate to determine “who is able” as a community. Implicit in this point is the “ability to work”. If work disappears, then that reduces one’s ability to work. The group flat out rejects the notion of a “right NOT to work”. That of course is not the same as “you must have a job and be receiving pay”. The group roundly supports the value of “unpaid jobs” such as stay-at-home caretakers or volunteers.
- In the context of substantial technological unemployment, the group understood and accepted the idea that Universal Basic Income might be the only option. No other alternative was offered as yet.
- There was genuine concern about UBI and unintended consequences, such as laziness, forced re-location and subsequent low-income housing concentration and negative feedback loops. Some of the group were familiar with UBI studies and their “smallness” and “terminal value”. They recognizing that behavior associated with these tests is not likely to compare to behavior in a world that MUST rely on UBI, such as the conditions that might come to pass under technological unemployment. Therefore, they reject the notion that we “know” how people would react under a comprehensive and necessary UBI program, reverting to concerns that it would not encourage work of all kinds.
- Following onto that point, one who is able, must work, whether they like the work or not. Where work is defined as “putting in effort” to participate in one’s survival or to execute the will of the community if the community is providing the support. This is different than a “job”, which is associated with pay or a salary. Stay-at-home parenting is work, and provides great benefit to the community without pay. They also reject the notion that a worker should enjoy their work. In fact, the group laughed at the idea that someone shouldn’t have to do work they don’t enjoy. They all wondered who the lucky ones were who always enjoyed their work.
- The group points to Capitalism’s excellent success in wealth creation, accepts the principle that “investment” from the wealthy creates growth and new opportunities. They also felt that the profit/return motive has made the allocation of capital generally efficient and thus generally productive. Further the group accepts that the benefit from new opportunities may be to a diminshing number of participants and that a consequence has been an increase in income inequality. One of the supporting arguments for higher taxes and potentially a UBI was the concern about rising income inequality. They did not reject the notion however that Capitalism may have an end-game — technological unemployment.
- There was considerable concern about the misuse of cash designed to provide food, clothing and shelter. One member who has had significant dealings with the poverty-stricken noted that frequently those in need, needed far more than monetary support, as mental-illness and drugs were often associated with their situation. It was suggested that a UBI payment might be used directly for food, clothing and shelter, instead of as cash to avoid misuse. To which there was varied debate, which I tabled (another version of “off into the weeds”). There were doubts about the government’s ability to provide the “right” solutions for those needs and externalities associated with that process. There was no conclusion on the best approach, cash or vouchers for services.
To wrap up our take on Universal Basic Income and trying to tie it together with Capitalism and Christianity, I would say the group was happy to consider the concept, unwilling to toss out capitalism, unwilling to accept some of the primary arguments of UBI advocates and generally unmotivated to run out and support a Universal Basic Income. They were happy to understand it better. Happy to consider the pros and cons more than they ever had and I know that awareness of the issues has been raised. Notably, I think everyone in the group is now comfortable having an opinion on the subject and how it fits into their views on life, poverty, public policy and technological unemployment. Maybe you, the reader, are a little more comfortable too. Whether you agree or disagree with the thoughts presented here, I suspect that the group’s thoughts are fairly mainstream. If you are vehemently opposed to UBI or zealously advocating for UBI, this ought to help you understand how one group thinks. Maybe it will make for a more fruitful dialogue as these challenges are considered in the future.
I suspect this is not an idea that many have considered. To date, as a society we really have not appreciated how technological change occurs and we certainly have rarely considered the governance of new change. When we have a discovery or an advancement of science, like all other inventions, it isn’t accomplished based upon the will of the majority. There is no vote, there is no consideration for society at-large (externalities). Rarely is the downside risk considered. One person sees a problem. Their own personal view of that problem and they aim to fix it. That is entreprenuerial, the foundation upon which much of Western Capitalism is built. It is also Authoritarian. One person, one rule and little or no accountability. Scary when you think about it. When you combine this “process” and lack of control with our species’ other great skill, “problem solving”, you create technological innovation and advancement which has a momentum that feels unstoppable. In fact, if you even “suggest” (which I am not) halting technological progress you get one of two response, “Luddite” or “You can’t”. That is how inexorable society views technological change.
So let me explain this in more detail, all technological advancement is about someone, somewhere seeing a problem, a weakness, a difficulty, a challenge and deciding to overcome that challenge. It’s admirable and impressive. It’s also creating problems. As a species we poorly weigh all aspects of a decision, all the pros and all the cons. Will we make money from this? Is this advancement, “cool”, Does this make my life “easier” are often the only inputs to our production/purchase decisions. There is a broad societal acceptance that “easier”, “freer” and “convenient” are universally beneficial. A simple counter-argument, of course, can be found in the gym. Your muscles would insist that “easier”, “freer” and”convenient” are not the best way for them to be strengthened or stamina to be built. They require “challenge”, “difficulty” and “strain” in order to grow and improve.
So when a new advance comes along, if it makes our life easier, even in the smallest way, we snatch it up instantly. Take, for example, Alexa and Google Home. Hugely successful products already, but was it really that difficult to type out our search query? Defenders will say things like, “now I can search while I am doing something else” or “this frees up my hands to be more productive”. And of course supporters point to the disabled for the obvious assistance to someone who is unable to type. But let’s examine the other side of the coin. What are the downside risks to such a product? Usually, not part of the sales process, I’m afraid, so you have to think carefully to compile a list. For example, our verbal search, has it caused us to lose a different skill, like the unchallenged muscle, whereby the finding of the solution was equally as important as the actual answer. But on top of that specific challenge, (the lost process and associated effort that may have strengthened the muscle, which in this case is the mind), what are some of the associated externalities to voice assistants? Let’s take a look at a few.
Is that search query worth having Amazon or Google record EVERY SINGLE WORD your family speaks within hearing distance? How about considering the fact that Amazon and Google now build personal profiles about your preferences based upon this information. Do you realize that this then limits your search results accordingly? Companies are taking CHOICE away from you and suprisingly, people don’t seem to care, in fact some like the idea. Other externalities exist as well. Recently, an Alexa recorded the conversation of a family and sent it to random contacts.
An Amazon Echo recorded a family's conversation, then sent it to a random person in their contacts…
A family in Portland, Ore., received a nightmarish phone call two weeks ago. "Unplug your Alexa devices right now," a…www.washingtonpost.com
Or this externality?
Without getting too paranoid, this last one is downright creepy and dystopian, but its potential ramifications are catastrophic if carried to the extreme. I am certain that when you decided to purchase your voice assistant, none of these externalities were factored into your buying decision.
I use this as a real life example, because our evaluation of new technology is based upon bad math. Is it cool? Is it easier? Is it profitable? and Can I afford it/afford to be without it? Nowhere in that equation are the following:
1) Does it further eliminate privacy?
2) Does it make me lazy or more dependent upon a machine?
3) Does to keep me from using all aspects of my brain as much?
4) Does it allow me to interact with actual humans less?
5) What new and different risks are associated with this product?
6) If someone wants to do me harm, does this enable them to do so in an unforseen way?
One of the chief arguments of technological advancement is that it frees us up from the mundane, routine tasks. To assume that those tasks do not have value is ludicrous on its face, but more importantly, if we are “freed” up, what are we freed up to do? Usually, we are told it is high-minded things… be entrepreneurial… be poetic… be deep thinkers… solve bigger problems… spend more time with loved ones. To be honest, I don’t see an explosion of those endeavors. A further example of our bad math…
We adopt these technological advancements often without thought about the impact it may have on our psyche, our self-worth, our ambitions, or our safety. We make these choices because they are cool, or they make something easier. With purchase decision processes that are this simple and “upside-only” considered, developers of technology have it easy to make products attractive.
This blog post has only lightly touched on malice, but all of society should be concerned about malicious intent and technology’s impact on our suceptibility. The more connected we are, the more dependent upon technology we are, the easier it is to cause mass harm. Perfect examples are recent virus attacks that spread to over 47 countries in a matter of a few hours. Sometimes the consequences are minor such as locked up computers or minor hassles we deal with like corrupted programs. Other times the hacker/criminal steals money or spies on you. Regardless of the magnitude of the impact, the ability of a criminal to “reach you” and “reach many” has been increased almost infinitely.
Here’s a final externality — how would you function without Internet? Not just for an hour or two, but permanenetly? How about without power? These are modern day conveniences that are assumed to be permenent, but how permanent are they? Do you need to consider how to operate when they are gone? Our connectivity and reliance on power make us deeply dependent and woefully unprepared for these alternatives, even if the odds of occurence are small. Hollywood frequently paints a grim picture of the dystopian existence when these conveniences are taken away, however, our ancestors existed quite nicely. Would you be prepared to survive or even thrive? The chances of these calamities are greater than zero…
Awareness of externalities is important, Consideration about downside risk is crucial and a willingness to realize that everything we do or even purchase has pros AND cons to them… The more awareness of the “cons” that you have, the better chance you have to mitigate those risks and reap a greater benefit from the upside of our technology choices. Most importantly, as a society, we will make better collective decisions on our technological progress and thwart the dangers of Technological Autocracy.
To date, data is being valued and priced by everyone EXCEPT the creators of that content — YOU. If we want to change that many things need to happen, but it begins with taking the time to figure out how a person values their own data. So let’s dissect and see if we can shed some light on this idea.
First, this process is VERY unique, because for the first time EVER, every single person on the planet has the potential to sell a product. Second, instead of being a “consuming” culture and propelling the corporate world forward, human beings are the ones in a position to profit. Third, everyone’s value judgement on data is unique, personal and unquestionable. Fourth, the opportunties for people to enrich themselves in a world of possible technological unemployment is tremendously important to the welfare of society. Finally, on top of the social ramifications, there is the obvious moral ramifications. As highlighted by the misuse of your data by corprorations, this idea of individual data ownership is morally correct.
Now we are not talking about ALL data. If you use Amazon’s services to buy something. All of the information that you create by searching, browsing and buying on the Amazon site, also belongs to them. So while I can opine on the “value” of individual data, I am certain that the legal questions around data are just beginning to be sorted out.
So with all of that in mind, let’s examine how each individual person may value the data that they can provide. Noteworthy to this discussion is that every individual has a different value function. Different people will value different things about their data. So it is vital that we appreciate that each person will price things uniquely.
However the parameter that they weigh can be summarized in a few key variables which are covered below. So lets create a list and explain each one:
- Risk of Breach — Each data item, if fallen into the wrong hands can cause harm. This is the risk of breach. This risk will be perceived differently based upon the reputation for safety of the data user, a perceived sense of associated insurance and the context of the data itself. For example, let’s consider 4 tiers of risk of breach. Tier 1 ( HARMLESS) — the contents of my dishwasher. This data might have value to someone and could not harm me if used nefariously. Tier 2 (LKELY HARMLESS)— the contents of my refrigerator. Still like to be unable to hurt me, but since people may know what I consume, one could they possibly tamper with it. Tier 3 (HARMFUL — ONLY INCONVENIENT) Examples here might include, financial breach. Where often the risk is not only yours, there is a form of insurance (bank insurance or other similar example), but it is dangerous and painful when it occurs. Tier 4 (HARMFUL — PERSONAL SAFETY) Examples here might include your exact location, your health records, access to your cybernetics and/or your genetic code.
- Risk of Privacy — How sensitive or personal do we view the data items. On this risk, I beleive that pricing is rather binary or maybe parabolic. Many data items which we can produce do not make us concerned if made public. That is, until a line is crossed, where we consider them private. My pictures, fine. My shared moments, fine. My bedroom behavior, private. So when that line is crossed, the price of the associated data rises substantially. To continue the example, a manufacturer of an everyday item, such as pots and pans, may not have to pay a privacy premium for data associated with our cooking habits. However, a manufacturer of adult toys, may have to pay a substantial premium to gain access to the bedroom behavior of a meaningful size sample of data. This is a good time to remember that these pricing mechanisms are personal, true microeconomics and everyone will value the risk of privacy differently. Even to the point where the example I just gave may be completely reversed. Bedroom behavior, no problem… but keep your eyes of of my cooking.
- Time — how easy is it to generate the data. Can I generate the data simply by existing? That data will be cheaper. Do I have to engage in a use of my time to create the data, that data will be more expensive. Would you like me to watch your commercials? more expensive. Would you like me to fill out your survey? 2 questions is cheaper than 20 questions. Time is also a function about the entire mechanism for creating, monitoring the data.
- Applicability — is the data being asked of me relevant to me. This is a question of “known” versus “unknown”. If I regularly drink a certain type of coffee, I am more likely to accept coupons, ads, sales and promotions from my coffee shop than I am from the Tea emporium around the corner. The function here is inverted, as the applicability decreases, the value of access to “me” increases from my perspective. That is not to say that it also increases for the data consumer, so with respect to applicability we have typically juxtaposed supply and demand curves. Also, if you only value data based on the supply side (what I am willing to give), then you miss out on revenue opportunities by allowing people access to your attention to “broaden your exposure”.
If the world changes to a personal data driven model, then the corporate world and the artificial intelligence world, will have to learn how to examine these key variables. The marketplace where these transactions will occur MUST be a robust mechanism for price discovery whereby many different bids and offers are being considered on a real-time basis to determine the “price/value” of data This is why I have proposed the Personal Data Exchange as a mechanism for identifying this value proposition. Exchanges are in the business of price discovery, on behalf of their listed entities, in this case, “you”.
In the end, this is the morally corrected position. For a variety of reasons it a justifiable and necessary change to a marketplace that was created largerly without your consent. Recent changes to the law, such as GDPR in Europe have begun to claw back the rights of the indidivual. But if we can get this done, it becomes a complete gamechanger. Please… get on board. Your thoughts and critiques are welcome and encouraged, ForHumanity.
Know Your Customer (KYC) is a required practice in finance. Defined as the process of a business identifying and verifying the identity of its clients. The term is also used to refer to the bank and anti-money laundering regulations which governs these activities. Many of you will not be familiar with this rule of law. It exists primarily in the financial industry and is a cousin to laws such as Anti-Money Laundering (AML) and the Patriot Act of 2001 and US Freedom Act of 2015. These laws were designed to require companies to examine who their clients were. Are they involved in illegal activities? Do they finance terrorism? What is the source of these monies? Does the customer engage in illegal activity? The idea was to prevent our financial industry from supporting or further the ability of wrong-doers to cause harm. So how does this apply to Facebook and the Cambridge Analytica issues?
I am suggesting that the Data Industry, which includes any company that sells or provides access to individual information should be held to this same standard. Facebook should have to Know Your Customer. Google should have to Know Your Customer. Doesn’t this seem reasonable? The nice part about this proposal is that it isn’t new. We don’t have to draft brand new laws to cover it, rather just modify some existing language. KYC exists for banks, now let’s expand it to social media, search engines and the sellers of big data.
Everywhere in the news today, we have questions about “who is buying ads on social media”? Was it Russians trying to influence an election? Was it neo-nazis, ANTIFA or other radical idealogues? Is it a purveyor of “fake news”? If social media outlets were required to KYC their potential clients then they will be able to weed out many of these organizations well before their propoganda reaches the eyes of their subscribers. Facebook has already stated that they want to avoid allowing groups such as these to influence their users via their platform. So it is highly reasonable to ask them to do it, or be penalized for failure to do so. Accountability is a powerful thing. Accountability means that it actuals gets done.
Speaking of “getting it done”, some of you may have seen Facebook’s complete about-face on its compliance with GDPR, moving 1.5 billion users out of Irish jurisdiction and to California where there are very limited legal restricitons. https://arstechnica.com/tech-policy/2018/04/facebook-removes-1-5-billion-users-from-protection-of-eu-privacy-law/
If you aren’t familiar with GDPR, it is Europe’s powerful new privacy law. For months, Facebook has publically stated how it would intend to comply with the law. But when push came to shove, their most recent move is to avoid the law and avoid compliance as much as possible. So flowery-language is something we often here from corporate executives on these matters, but in the end, they will still serve shareholders and profits first and foremost. So unless, these companies are forced to comply, don’t expect them to do it out of moral compunction, that’s rarely how companies operate.
Returning the the practical application of KYC, for a financial firm, this means that a salesperson has to have a reasonable relationship with their client, in order to assure that they are compliant with KYC. They need to know the client personally and be familiar with the source and usage of funds. If a financial firm fails to execute KYC and it turns out that the organization they are doing business with is charged with a crime, then the financial firm and the individuals involved would find swift ramifications, including substantive fines and potential jail time. This should apply to social media and the data industry.
Let me give you a nasty example. Have you looked at the amazing detail Facebook or Google have compiled about you? It is fairly terrifying and there are some out there (Gartner, for example) who have even predicted that your devices will know you better than your family knows you by 2022.
Now assuming this is even close to true for many of us, then imagine where that information is sold to a PhD candidate at MIT, or other reputable AI program, except that PhD student, beyond doing his AI research is funnelling that data on to hackers on the dark web, or worse, to a nation-state engaged in cyberwarfare. How easy would it be for that group to cripple a large portion of the country? Or maybe, it has already happened, with examples like Equifax and its 143 million client breach. Can you be sure that the world’s largest hacks aren’t getting their start by accessing your data from a data reseller?
To be fair, in finance, often times you are taking in the funds and controlling the activities after the fact. You know what is going on. With data, often times you are selling access or actual data to the customer and no longer have control over their activities, it might seem. But this argument simply enhances my interest in Know Your Customer, because these firms may have little idea how this data is being used or misused. Better to slow down the gravvy train than to ride it into oblivion.
Obivously the details would need to be drafted and hammered out by Congress, but I am seeking support of the broader concept and encouraging supporters to add it to the legislative agenda. ForHumanity has a fairly comprehensive set of legislative proposals at this point which we would hope would be consider in the broad concept of AI policy. Questions, thoughts and comments are always welcome. This field of AI safety remains so new that we really should have a crowd-sourced approach to identify best-practices. We welcome you to join us in the process.
Many of you will be familiar with the challenges that Facebook is facing.
The Cambridge Analytica saga is a scandal of Facebook's own making | John Harris
Big corporate scandals tend not to come completely out of the blue. As with politicians, accident-prone companies…www.theguardian.com
Internal disagreements about how data has been used. Was it sold? Was it manipulated? Was it breached? It has put the company itself at-risk and highlighted the need for a new position at the C-Suite level, one that most companies have avoided up until now. AI and Automation Risk Management.
Data is the new currency, Data is the new oil. It is the lifeblood of all Artificial Intelligence algorithms, machine and deep learning models. It is the way that our machines are being trained. It is the basis for the best and brightest to begin to uncover new tools for healthcare, new ways to protect security, new ways to sell products. In 2017, just Google and Facebook had a revenue close to $60 billion in advertising alone, all due to data. However, usage of that data is at-risk, because of perceived abuses by Facebook and others.
Data is also and more importantly about PEOPLE. Data is PERSONAL, even if there have been attempts to anonymize it. People need an advocate,inside the company, defending their legal, implied and human rights. This is a dynamic marketplace, with new rules, regulations and laws being considered and implemented all of the time. AI and Automation face some substantial challenges in their development, here is a short list and NONE of these are a priority for engineers and programmers, despite the occassional altruistic commentary to the contrary. As you will see the advancement of AI and Automation requires full-time attention.:
- Ethical machines and algorithms — There are millions and millions of decisions being made by machines and algorithms. Many of these decisions are meant to be based upon our value system as human beings. That means our Ethics need to be coded into the machines and this is no easy task with a single set of Ethics. Deriving profit from Ethics, is tenuous at best and there is certain to be a cost.
- Data and decision bias — Our society is filled with bias, our perceptions are filled with bias, thus our data is filled with bias. If we do not take steps to correct bias in the data, then our algorithmic decisions will be biased. However, correcting for bias may not be as profitable which is why it needs to be debated in the C-suite.
- Privacy — there is a significant push back forming on what is privacy online. GDPR in Europe is a substantial set of laws providing people with increased transparency, privacy and in some cases the right to be forgotten. Compliance with GDPR is one responsibility of the AI Risk Manager.
- Cybersecurity and Usage Security (a Know-Your Customer process for data uasge). Companies already engage in cybersecurity, but the responsiblity is higher when you are protecting customer data. Furthermore, companies should adopt the the finance industry standard of “Know Your Customer (KYC)”. Companies must know and understand how data is being used by clients to prevent abuses or illegal behavior.
- Safety (can the machines that are being built be controlled and avoid unintentional consequences to human life). This idea is a little farther afield for most, however now that an UBER, autonomous vehicle has been involved in a fatality, it is front and center. The AI Risk Manager’s job is to consider the possiblities of how a machine may injure a human being. Whether that be through hack, negligence, or system failure.
- Future of Work (as jobs are destroyed, how does the company treat existing employees and the displaced workers/communities) — This is the PR challenge of the role, depending on how the company chooses to engage it’s community. But imagine for a moment taking a factory with 1000 employees and automating the whole thing. that’s 1000 people directly effected. That’s a community potentially devestated, if all 1000 employees were laid off.
- Legal implications of cutting edge technology (in partnership with Legal team or outside counsel) — GDPR compliance, legal liability of machines, new regulations and their implementation. These are the domain of the AI Risk Manager in conjunction with counsel and outside counsel.
This voice is is a C-suite job and must have equal authority to the sources of revenue, in order to stand a chance of protecting the company and protecting the data, i.e. the people who create the data.
I am not here to tell you to stop using data. However if you believe that each of these companies, whose primary purpose is not compliance but instead to make profits, will always use this data prudently is naive at its best. Engineers and programmers solve problems, they have not been trained to consider existential risk, such as feelings, privacy rights, and bias. They see data of all kinds as “food” and “input” to their models. To be fair, I don’t see anything wrong with their approach. However, the company cannot let that exist unfettered. It is risking its entire reputation on using data, which is actually using PEOPLE’S PERSONAL AND OFTEN PRIVATE INFORMATION, to get results.
For many new companies and new initiatives, data is the lifeblood of the effort. It deserves to be protected and safeguarded. Beyond that, since data is about people, it deserves to be treated with respect, consideration, and fairness. Everyone is paying more and more attention to issues like this and companies must react. People and their data need an advocate in the C-Suite. I recommend the Chief of AI and Automation Risk Manager.
Here’s a quick and easy way to break some of the current monopolies that exist in the personal data market (looking at Google, Facebook and Amazon). Let people own their own data, disseminate it as they see fit and, shockingly, get paid for what is rightfully theirs.
The concept is simple, the implementation is challenging and requires a good size investment from somewhere at the outset. I’ll explain the concept first and then circle back around to the implementation at the end.
A personal data exchange. Simply explained, each person 18 and over, has the right to list themselves on a “data” exchange, just like companies list their equity on a stock exchange. Each individual would get a listing off their own on the exchange. So ideally the exchange would have hundreds of millions of listings. One, for each unique person. This exchange, which would function as a clearinghouse for YOUR data. It would be the marketplace to go to for companies, of all kinds, to retrieve the data that they require from the sellers of data — YOU. If you list John Q Public on the exchange and want to severely limit the data you provide the world, that is your choice, however, because your data is used less frequently, you will get paid very little for being John Q Public. Privacy, is your right and you should be in control of your own data.
However, if John Q Public lists his data, answers questions submitted by the marketplace, responds in a timely manner to requests and is generally open about preferences, opinions and many other pieces of information in demand, then John Q Public will find a nice revenue stream associated with this very valuable data.
The marketplace is interactive. It starts with many of the key data items that are commonly available on a simple internet search. However, as you file your “listing” on the exchange then companies will seek your information, maybe thru GPS location, maybe via your search and query results. They may also send out questionaires. If you, the listing company, find yourself uncomfortable answering certain data questions, then stop. You are in control. You release your data. How you want, when you want. Do you want your shopping tracked? If yes, then allow it. Maybe only certain shopping is included. Furthermore, this data does not have to be “public”. It can be provided directly to a client company anonymously (anonymous data is cheaper to the company and pays you less as well). As the listing entity, providing the data, you will be in control. Companies can come to the marketplace and create datasets. Ask questions of the personal listings in an attempt to grow a business, design their alogirthms or pivot a marketing strategy. They can pay for high quality information which will be crucial to their business. As a “listed entity” you would have responsibility to respond, if you want to get paid. You must answer, truthfully, completely and in a timely manner. If the data proves to be useless over time because people lie and obscure relevant information, than the marketplace breaks down and people will lose CONTROL over the data, the way that it exists today.
It is hard to know how valuable this data is. Today, it is priced at zero by the owners of the data — you. Google and Facebook make nearly $100 billion per year, selling your data — selling you. I am certain that given the choice between no Google Search engine, no facebook and $1000 in your pocket. Most people will choose the $1000. Furthermore, it is an excellent way to democratize wealth as anyone will be eligible to participate.
Now, implementation is a challenge, but not insurmountable. The absolute key is that you must launch the exchange with many, many listings. Therefore, to motivate individuals to list on the exchange, they will need to be compensated to do so. I find that money has a way of getting people’s attention. However, we aren’t talking about a small amount of money, as the minimum number of listings probably needs to be 20–30 million people from the outset to make the dataset meaningful. Furthermore, an incentive system to encourage people to help others list will be valuable. So the capital required to incentivize the listing is significant, well over $2 billion just to launch. Fortunately, there are a few large pools of investors, unbiased and correctly aligned with individuals to fund the endeavor.
Secondly, the technological challenge of this exchange is not trivial. Handling over 100 million listings of John and Sarah Q Publics is a sizable endeavor. Furthermore, the front-end will need to be extremely flexible and user friendly to ensure that all listings can easily and efficiently manage their data.
The education process is also monumental as many people do NOT value their data, furthermore they do not value their privacy. They are content to give away this asset for free, or at least in exchange for a forum to fight about politics, post some pictures, search the web, have a laugh or order some goods to be delivered in less than 2 days. However, success here breeds success and with the right incentive structure, I believe that this process will go viral. If the money is sufficient enough, then data will become an important revenue stream for individuals and households.
Finally, this has to be done at a truly institutional level. It requires the cache of the New York Stock Exchange or Chicago Mercantile Exchange to demonstrate the “seriousness” of the concept, combined with the institutional fortitude to handle the sheer magnitude required to make this successful and finally with the technology chops and delivery expertise to pull it all together and provide an exceptional experience for the listed individuals. This idea has been tried or discussed on a smaller scale, but small scale doesn’t work for the companies in the data business. This idea is go big or go home, but in terms of privacy and control of personal data, there is no better way than to put you in charge and let the market pay you directly.
The disintermediation of Google, Facebook and Amazon in the “control your data business” is just a perk…