OpenAI is the for-profit corporation OpenAI LP, whose parent organization is the non-profit organization OpenAI Inc, which conducts research in the field of artificial intelligence (AI) with the stated aim to promote and develop friendly AI in such a way as to benefit humanity as a whole. Founded in late 2015, the San Francisco-based organization aims to “freely collaborate” with other institutions and researchers by making its patents and research open to the public. The founders (notably Elon Musk and Sam Altman) are motivated in part by concerns about the existential risk from artificial general intelligence.
|Founded||December 11, 2015|
|Founders||Elon Musk, Sam Altman, Ilya Sutskever, Greg Brockman|
|Ilya Sutskever, Greg Brockman|
In October 2015, Musk, Altman and other investors announced the formation of the organization, pledging over US$1 billion to the venture.
On December 5, 2016, OpenAI released Universe, a software platform for measuring and training an AI's general intelligence across the world's supply of games, websites and other applications.
- CEO: Sam Altman, former president of the startup accelerator Y Combinator
- Ilya Sutskever, Research director, a former Google expert on machine learning
- CTO: Greg Brockman, former CTO of Stripe
Other backers of the project include:
- Reid Hoffman, LinkedIn co-founder 
- Peter Thiel, PayPal co-founder 
- Jessica Livingston, a founding partner of Y Combinator
The group started in early January 2016 with nine researchers. According to Wired, Brockman met with Yoshua Bengio, one of the “founding fathers” of the deep learning movement, and drew up a list of the “best researchers in the field”. Microsoft's Peter Lee stated that the cost of a top AI researcher exceeds the cost of a top NFL quarterback prospect. While OpenAI pays corporate-level (rather than nonprofit-level) salaries, it doesn't currently pay AI researchers salaries comparable to those of Facebook or Google. Nevertheless, Sutskever stated that he was willing to leave Google for OpenAI “partly of because of the very strong group of people and, to a very large extent, because of its mission.” Brockman stated that “the best thing that I could imagine doing was moving humanity closer to building real AI in a safe way.” OpenAI researcher Wojciech Zaremba stated that he turned down “borderline crazy” offers of two to three times his market value to join OpenAI instead.
Some scientists, such as Stephen Hawking and Stuart Russell, believed that if advanced AI someday gains the ability to re-design itself at an ever-increasing rate, an unstoppable "intelligence explosion" could lead to human extinction. Musk characterizes AI as humanity's "biggest existential threat." OpenAI's founders structured it as a non-profit so that they could focus its research on creating a positive long-term human impact.
OpenAI states that "it's hard to fathom how much human-level AI could benefit society," and that it is equally difficult to comprehend "how much it could damage society if built or used incorrectly". Research on safety cannot safely be postponed: "because of AI's surprising history, it's hard to predict when human-level AI might come within reach." OpenAI states that AI “should be an extension of individual human wills and, in the spirit of liberty, as broadly and evenly distributed as possible...”, and which sentiment has been expressed elsewhere in reference to a potentially enormous class of AI-enabled products: "Are we really willing to let our society be infiltrated by autonomous software and hardware agents whose details of operation are known only to a select few? Of course not."  Co-chair Sam Altman expects the decades-long project to surpass human intelligence.
Vishal Sikka, former CEO of Infosys, stated that an “openness” where the endeavor would “produce results generally in the greater interest of humanity” was a fundamental requirement for his support, and that OpenAI “aligns very nicely with our long-held values” and their “endeavor to do purposeful work”. Cade Metz of Wired suggests that corporations such as Amazon may be motivated by a desire to use open-source software and data to level the playing field against corporations such as Google and Facebook that own enormous supplies of proprietary data. Altman states that Y Combinator companies will share their data with OpenAI.
In 2019 OpenAI became a for profit company called OpenAI LP while staying controlled by a non-profit called OpenAI Inc, a structure that OpenAI call "capped-profit", having previously been a 501(c)(3) nonprofit organization.
Musk posed the question: “what is the best thing we can do to ensure the future is good? We could sit on the sidelines or we can encourage regulatory oversight, or we could participate with the right structure with people who care deeply about developing AI in a way that is safe and is beneficial to humanity.” Musk acknowledged that “there is always some risk that in actually trying to advance (friendly) AI we may create the thing we are concerned about”; nonetheless, the best defense is “to empower as many people as possible to have AI. If everyone has AI powers, then there's not any one person or a small set of individuals who can have AI superpower.”
Musk and Altman's counter-intuitive strategy of trying to reduce the risk that AI will cause overall harm, by giving AI to everyone, is controversial among those who are concerned with existential risk from artificial intelligence. Philosopher Nick Bostrom is skeptical of Musk's approach: “If you have a button that could do bad things to the world, you don't want to give it to everyone.” During a 2016 conversation about the technological singularity, Altman said that “we don’t plan to release all of our source code” and mentioned a plan to “allow wide swaths of the world to elect representatives to a new governance board”. Greg Brockman stated that “Our goal right now... is to do the best thing there is to do. It’s a little vague.”
Products and applicationsEdit
Gym aims to provide an easy to set up, general-intelligence benchmark with a wide variety of different environments—somewhat akin to, but broader than, the ImageNet Large Scale Visual Recognition Challenge used in supervised learning research—and that hopes to standardize the way in which environments are defined in AI research publications, so that published research becomes more easily reproducible. The project claims to provide the user with a simple interface. As of June 2017, Gym can only be used with Python. As of September 2017, the Gym documentation site was not maintained, and active work focused instead on its GitHub page.
In “RoboSumo”, virtual humanoid “metalearning” robots initially lack knowledge of how to even walk, and given the goals of learning to move around, and pushing the opposing agent out of the ring. Through this adversarial learning process, the agents learn how to adapt to changing conditions; when an agent is then removed from this virtual environment and placed in a new virtual environment with high winds, the agent braces to remain upright, suggesting it had learned how to balance in a generalized way. OpenAI's Igor Mordatch argues for that competition between agents can create an intelligence “arms race” that can increase an agent's ability to function, even outside the context of the competition.
In 2018, OpenAI launched the Debate Game, which teaches machines to debate toy problems in front of a human judge. The purpose is to research whether such an approach may assist in auditing AI decisions and in developing explainable AI.
OpenAI Five is the name of a team of five OpenAI-curated bots that are used in the competitive five-on-five video game Dota 2, who learn to play against human players at a high skill level entirely through trial-and-error algorithms. Before becoming a team of five, the first public demonstration occurred at The International 2017, the annual premiere championship tournament for the game, where Dendi, a professional Ukrainian player, lost against a bot in a live 1v1 matchup. After the match, CTO Greg Brockman explained that the bot had learned by playing against itself for two weeks of real time, and that the learning software was a step in the direction of creating software that can handle complex tasks like a surgeon. The system uses a form of reinforcement learning, as the bots learn over time by playing against themselves hundreds of times a day for months, and are rewarded for actions such as killing an enemy and taking map objectives.
By June 2018, the ability of the bots expanded to play together as a full team of five and were able to defeat teams of amateur and semi-professional players. At The International 2018, OpenAI Five played in two exhibition matches against professional players, but ended up losing both games. In April 2019, OpenAI Five defeated OG, the reigning world champions of the game at the time, 2:0 in a live exhibition match in San Francisco. The bots' final public appearance came later that month, where they played in 42,729 total games in a four-day open online competition, winning a percentage of 99.4% of games. However, it is important to note that OpenAI only learned a heavily simplified version of Dota, including only 17 out of over 100 heroes and excluding certain items as well as game mechanics.
Dactyl uses machine learning to train a robot Shadow Hand from scratch, using the same reinforcement learning algorithm code that OpenAI Five uses. The robot hand is trained entirely in physically inaccurate simulation.
GPT2 (2019) is an AI system that generates text matching its input in subject and tone. For example, when fed the first sentence of George Orwell's novel Nineteen Eighty-Four it produces plausible futuristic fiction set in China. Unlike previous OpenAI products, GPT2 initially has not been released to the public out of concerns of potential misuse, including applications for writing fake news. Much of the academic community is skeptical that GPT2 poses a significant threat. The Allen Institute for Artificial Intelligence followed up with a tool to detect "neural fake news". Other researchers, like Jeremy Howard, warn of "the technology to totally fill Twitter, email, and the web up with reasonable-sounding, context-appropriate prose, which would drown out all other speech and be impossible to filter". Eventually, OpenAI released the complete version of the GPT2 language model.
Gym Retro is a platform for reinforcement learning research on games. Gym Retro is used to conduct research on RL algorithms and study generalization. Prior research in RL has mostly focused on optimizing agents to solve single tasks. Gym Retro gives the ability to generalize between games with similar concepts but different appearances.
- Markoff, John (December 11, 2015). "Artificial-Intelligence Research Center Is Founded by Silicon Valley Investors". The New York Times. Retrieved December 12, 2015.
- Metz, Cade (July 22, 2019). "With $1 Billion From Microsoft, an A.I. Lab Wants to Mimic the Brain". The New York Times. Retrieved July 22, 2019.
- "OpenAI shifts from nonprofit to 'capped-profit' to attract capital". TechCrunch. Retrieved 2019-05-10.
- "Tech giants pledge $1bn for 'altruistic AI' venture, OpenAI". BBC News. 12 December 2015. Retrieved 19 December 2015.
- "Introducing OpenAI". OpenAI Blog. 12 December 2015.
- Lewontin, Max (14 December 2015). "Open AI: Effort to democratize artificial intelligence research?". The Christian Science Monitor. Retrieved 19 December 2015.
- Cade Metz (27 April 2016). "Inside OpenAI, Elon Musk's Wild Plan to Set Artificial Intelligence Free". Wired magazine. Retrieved 28 April 2016.
- Dave Gershgorn (27 April 2016). "Elon Musk's Artificial Intelligence Group Opens A 'Gym' To Train A.I." Popular Science. Retrieved 29 April 2016.
- Metz, Cade. "Elon Musk's Lab Wants to Teach Computers to Use Apps Just Like Humans Do". WIRED. Retrieved 31 December 2016.
- Mannes, John. "OpenAI's Universe is the fun parent every artificial intelligence deserves". TechCrunch. Retrieved 31 December 2016.
- "OpenAI - Universe". Retrieved 31 December 2016.
- Claburn, Thomas. "Elon Musk-backed OpenAI reveals Universe – a universal training ground for computers". The Register. Retrieved 31 December 2016.
- Vincent, James (February 21, 2018). "Elon Musk leaves board of AI safety group to avoid conflict of interest with Tesla". The Verge.
- Conger, Kate. "Elon Musk's Neuralink Sought to Open an Animal Testing Facility in San Francisco". Gizmodo. Retrieved 2018-10-11.
- Needleman, Sarah. "Microsoft to Invest $1 Billion in Artificial-Intelligence Startup". Wall Street Journal. Retrieved 2019-08-02.
- Bass, Dina (22 July 2019). "Microsoft to invest $1 billion in OpenAI". Los Angeles Times. Los Angeles Times. Retrieved 22 July 2019.
- "Silicon Valley investors to bankroll artificial-intelligence center". The Seattle Times. 13 December 2015. Retrieved 19 December 2015.
- Etherington, Darrell (July 22, 2019). "Microsoft invests $1 billion in OpenAI in new multiyear partnership". TechCrunch. Retrieved July 22, 2019.
- Liedtke, Michael. "Elon Musk, Peter Thiel, Reid Hoffman, others back $1 billion OpenAI research center". San Jose Mercury News. Retrieved 19 December 2015.
- "Elon Musk, Infosys, others back OpenAI with $1 bn". Business Standard India. Business Standard. IANS. 12 December 2015 2015. Retrieved 30 August 2019. Check date values in:
- Vincent, James (22 July 2019). "Microsoft invests $1 billion in OpenAI to pursue holy grail of artificial intelligence". The Verge. Retrieved 23 July 2019.
- Anthony Patch (2018-03-10), Elon Musk Artificial Intelligence could be our biggest existential threat 720p, retrieved 2018-10-31
- Mendoza, Jessica. "Tech leaders launch nonprofit to save the world from killer robots". The Christian Science Monitor.
- Glenn W. Smith (10 April 2018). "Re: Sex-Bots—Let Us Look before We Leap". Arts. doi:10.3390/arts7020015. Retrieved 30 June 2018.
- Metz, Cade (15 December 2015). "Elon Musk's Billion-Dollar AI Plan Is About Far More Than Saving the World". Wired. Retrieved 19 December 2015.
“Altman said they expect this decades-long project to surpass human intelligence.”
- Vishal Sikka (14 December 2015). "OpenAI: AI for All". InfyTalk. Infosys. Archived from the original on 22 December 2015. Retrieved 22 December 2015.
- Levy, Steven (December 11, 2015). "How Elon Musk and Y Combinator Plan to Stop Computers From Taking Over". Medium/Backchannel. Retrieved December 11, 2015.
“Elon Musk: ...we came to the conclusion that having a 501(c)(3)... would probably be a good thing to do”
- Brockman, Greg (April 3, 2017). "Yes, we're a 501(c)(3). As you mention in /r/ControlProblem, we will file our 990 later this year as required. Not yet sure of exact date".
- "Sam Altman's Manifest Destiny". The New Yorker (10 October 2016). Retrieved 4 October 2016.
- Greg Brockman; John Schulman (27 April 2016). "OpenAI Gym Beta". OpenAI Blog. OpenAI. Retrieved 29 April 2016.
- "OpenAI Gym". GitHub. Retrieved 8 May 2017.
- Brockman, Greg (12 Sep 2017). "Yep, the Github repo has been the focus of the project for the past year. The Gym site looks cool but hasn't been maintained". @gdb. Retrieved 2017-11-07.
- "AI Sumo Wrestlers Could Make Future Robots More Nimble". Wired. 11 October 2017. Retrieved 2 November 2017.
- "OpenAI's Goofy Sumo-Wrestling Bots Are Smarter Than They Look". MIT Technology Review. Retrieved 2 November 2017.
- Greene, Tristan (2018-05-04). "OpenAI's Debate Game teaches you and your friends how to lie like robots". The Next Web. Retrieved 2018-05-31.
- "Why Scientists Think AI Systems Should Debate Each Other". Fast Company. 8 May 2018. Retrieved 2 June 2018.
- Savov, Vlad. "My favorite game has been invaded by killer AI bots and Elon Musk hype". The Verge. Retrieved June 25, 2018.
- Frank, Blair Hanley. "OpenAI's bot beats top Dota 2 player so badly that he quits". Venture Beat. Archived from the original on August 12, 2017. Retrieved August 12, 2017.
- "Dota 2". blog.openai.com. Retrieved 12 August 2017.
- "More on Dota 2". blog.openai.com. Retrieved 16 August 2017.
- Simonite, Tom. "Can Bots Outwit Humans in One of the Biggest Esports Games?". Wired. Retrieved June 25, 2018.
- Kahn, Jeremy. "A Bot Backed by Elon Musk Has Made an AI Breakthrough in Video Game World". Bloomberg. Retrieved June 27, 2018.
- Clifford, Catherine. "Bill Gates says gamer bots from Elon Musk-backed nonprofit are 'huge milestone' in A.I." CNBC. Retrieved June 29, 2018.
- "OpenAI Five Benchmark". blog.openai.com. Retrieved 25 August 2018.
- Simonite, Tom. "Can Bots Outwit Humans in One of the Biggest Esports Games?". Wired. Retrieved 25 June 2018.
- Vincent, James. "AI bots trained for 180 years a day to beat humans at Dota 2". The Verge. Retrieved 25 June 2018.
- Savov, Vlad. "The OpenAI Dota 2 bots just defeated a team of former pros". The Verge. Retrieved August 7, 2018.
- Simonite, Tom. "Pro Gamers Fend off Elon Musk-Backed AI Bots—for Now". Wired. Retrieved 25 August 2018.
- Quach, Katyanna. "Game over, machines: Humans defeat OpenAI bots once again at video games Olympics". The Register. Retrieved 25 August 2018.
- "The International 2018: Results". blog.openai.com. Retrieved 25 August 2018.
- Statt, Nick. "OpenAI's Dota 2 AI steamrolls world champion e-sports team with back-to-back victories". The Verge. Retrieved 20 July 2019.
- "How to Train Your OpenAI Five". OpenAI Blog. Retrieved 20 July 2019.
- Wiggers, Kyle. "OpenAI's Dota 2 bot defeated 99.4% of players in public matches". Venture Beat. Retrieved 22 April 2019.
- "Learning Dexterity". Openai.com. OpenAI. Retrieved 27 August 2018.
- Ryan, Mae (2018). "How Robot Hands Are Evolving to Do What Ours Can". Retrieved 1 September 2018.
- Hern, Alex (14 February 2019). "New AI fake text generator may be too dangerous to release, say creators". The Guardian. Retrieved 14 February 2019.
- Schwartz, Oscar (4 July 2019). "Could 'fake text' be the next global political threat?". The Guardian. Retrieved 16 July 2019.
- Vincent, James (14 February 2019). "OpenAI's new multitalented AI writes, translates, and slanders". The Verge. Retrieved 16 July 2019.
- "GPT-2: 1.5B Release". OpenAI. 2019-11-05. Retrieved 2019-11-14.