The Partnership on AI consolidates control for a handful of corporations and their stakeholders, pitting them against Elon Musk’s OpenAI open source, non-profit that is available to everyone.

Both the Partnership on AI and Elon Musk’s OpenAI were established to advance humanity through Artificial Intelligence. While Musk’s OpenAI was devised to be open and available to the public, the Partnership on AI is more of a consolidation of power that benefits stakeholders whose findings will later be used to enact public policies.

Google/DeepMind, Microsoft, Facebook, IBM, and Amazon have collaborated to form the Partnership on AI to “formulate best practices on the challenges and opportunities within the field.” While third-party groups such as academics, non-profits, and policy specialists have been invited to be on the new Board of Directors, the Partnership on AI ultimately serves the private stakeholders, and whatever the stakeholders dictate will be relayed to the public.

Partnership on AI claims to not lobby policy makers, yet invites policy specialists to Board of Directors

Although the Partnership on AI “does not intend to lobby government or other policymaking bodies,” their internal research and conclusions will no doubt weigh heavily on “policymaking bodies.” What would be the point of consolidating all this money and power into one partnership and inviting policy specialists to the Board of Directors if no call to action on policy is implemented?

If the Partnership wants to establish itself as the main authority on the subject of AI best practices, how can policy makers not take into account its “authoritative” findings when it comes to practicing policies?

Maybe that’s the point. The Partnership on AI wouldn’t have to lobby governments or policymaking bodies, because those very same institutions would look to the Partnership as an authority they can rely-on to enact their policies.

Karl Marx didn’t have to lobby governments to have his ideas enacted; he was already dead long before the Communist Revolutions in Russia, China, Cuba, and elsewhere even took place — yet his Communist Manifesto with Frederich Engels was the framework used by these governments to enact their policies.

“Open” discussion between “closed” groups

One of the key goals outlined on the Partnership on AI’s website is “to provide a regular, structured platform for AI researchers and key stakeholders to communicate directly and openly with each other about relevant issues.”

Under the guise of creating an “open platform for discussion and engagement,” the platform is really about “open” discussion between insiders and stakeholders, not the public. The public won’t be privy to what is being discussed, only the conclusions of the Partnership — and later — the policies that are enacted after the decisions made behind closed doors.

Following up on the goal of so-called open discussion, the Partnership will then relay its findings by acting as a self-proclaimed “trusted and expert point of contact” while it creates its own educational material for the public.

In other words, the Partnership on AI is a consolidation of power that conducts its own research, gets its funding from special interests in finance, healthcare, and education, and then presents its findings to the public as the ultimate moral authority on the subject of AI.

Elon Musk’s alternative open source platform

The antithesis to the Partnership on AI is Elon Musk’s OpenAI. In naming one of his most successful business ventures Tesla Motors, Musk is an advocate of the ideologies of the late genius Nikola Tesla, who wanted to provide free, alternative energy to the public.

Tesla was blocked in his efforts by JP Morgan and Thomas Edison, who wanted to consolidate and control the energy sector by putting a restrictive monopoly on electricity in order to make a profit.

On the surface, both the Partnership on AI and OpenAI seem to serve the same goal; however, the means by which both organizations approach the subject of Artificial Intelligence ethics differ, and that has to do with whom they serve.

Read More: World’s 1st conference on ethics of sex robots launches in UK

“As a non-profit, our aim is to build value for everyone rather than shareholders,” states OpenAI’s website. This is a stark contrast to the Partnership on AI’s stakeholder-oriented research.

Because OpenAI is “free from financial obligations,” Musk’s organization is more geared towards free collaboration across institutions to focus on “positive human impact” with the idea that AI “should be an extension of individual human wills and, in the spirit of liberty, as broadly and evenly distributed as possible.”

Just as HBO and Sky Atlantic get set to debut the futuristic drama, “Westworld,” about a theme park called staffed by robotic AI and the ethical complications that arise, OpenAI has already begun to contemplate the “unknown” and potentially dangerous evolution of Artificial Intelligence as outlined on its website:

“It’s hard to fathom how much human-level AI could benefit society, and it’s equally hard to imagine how much it could damage society if built or used incorrectly.”

Read More: Can an AI collective subconscious exist? Inside HBO’s new drama Westworld

While the Partnership on AI sets out to make itself the leading moral authority on the “best practices” of AI, OpenAI seeks “to advance digital intelligence in the way that is most likely to benefit humanity as a whole, unconstrained by a need to generate financial return.”