Quotes & Sayings


We, and creation itself, actualize the possibilities of the God who sustains the world, towards becoming in the world in a fuller, more deeper way. - R.E. Slater

There is urgency in coming to see the world as a web of interrelated processes of which we are integral parts, so that all of our choices and actions have [consequential effects upon] the world around us. - Process Metaphysician Alfred North Whitehead

Kurt Gödel's Incompleteness Theorem says (i) all closed systems are unprovable within themselves and, that (ii) all open systems are rightly understood as incomplete. - R.E. Slater

The most true thing about you is what God has said to you in Christ, "You are My Beloved." - Tripp Fuller

The God among us is the God who refuses to be God without us, so great is God's Love. - Tripp Fuller

According to some Christian outlooks we were made for another world. Perhaps, rather, we were made for this world to recreate, reclaim, redeem, and renew unto God's future aspiration by the power of His Spirit. - R.E. Slater

Our eschatological ethos is to love. To stand with those who are oppressed. To stand against those who are oppressing. It is that simple. Love is our only calling and Christian Hope. - R.E. Slater

Secularization theory has been massively falsified. We don't live in an age of secularity. We live in an age of explosive, pervasive religiosity... an age of religious pluralism. - Peter L. Berger

Exploring the edge of life and faith in a post-everything world. - Todd Littleton

I don't need another reason to believe, your love is all around for me to see. – Anon

Thou art our need; and in giving us more of thyself thou givest us all. - Khalil Gibran, Prayer XXIII

Be careful what you pretend to be. You become what you pretend to be. - Kurt Vonnegut

Religious beliefs, far from being primary, are often shaped and adjusted by our social goals. - Jim Forest

We become who we are by what we believe and can justify. - R.E. Slater

People, even more than things, need to be restored, renewed, revived, reclaimed, and redeemed; never throw out anyone. – Anon

Certainly, God's love has made fools of us all. - R.E. Slater

An apocalyptic Christian faith doesn't wait for Jesus to come, but for Jesus to become in our midst. - R.E. Slater

Christian belief in God begins with the cross and resurrection of Jesus, not with rational apologetics. - Eberhard Jüngel, Jürgen Moltmann

Our knowledge of God is through the 'I-Thou' encounter, not in finding God at the end of a syllogism or argument. There is a grave danger in any Christian treatment of God as an object. The God of Jesus Christ and Scripture is irreducibly subject and never made as an object, a force, a power, or a principle that can be manipulated. - Emil Brunner

“Ehyeh Asher Ehyeh” means "I will be that who I have yet to become." - God (Ex 3.14) or, conversely, “I AM who I AM Becoming.”

Our job is to love others without stopping to inquire whether or not they are worthy. - Thomas Merton

The church is God's world-changing social experiment of bringing unlikes and differents to the Eucharist/Communion table to share life with one another as a new kind of family. When this happens, we show to the world what love, justice, peace, reconciliation, and life together is designed by God to be. The church is God's show-and-tell for the world to see how God wants us to live as a blended, global, polypluralistic family united with one will, by one Lord, and baptized by one Spirit. – Anon

The cross that is planted at the heart of the history of the world cannot be uprooted. - Jacques Ellul

The Unity in whose loving presence the universe unfolds is inside each person as a call to welcome the stranger, protect animals and the earth, respect the dignity of each person, think new thoughts, and help bring about ecological civilizations. - John Cobb & Farhan A. Shah

If you board the wrong train it is of no use running along the corridors of the train in the other direction. - Dietrich Bonhoeffer

God's justice is restorative rather than punitive; His discipline is merciful rather than punishing; His power is made perfect in weakness; and His grace is sufficient for all. – Anon

Our little [biblical] systems have their day; they have their day and cease to be. They are but broken lights of Thee, and Thou, O God art more than they. - Alfred Lord Tennyson

We can’t control God; God is uncontrollable. God can’t control us; God’s love is uncontrolling! - Thomas Jay Oord

Life in perspective but always in process... as we are relational beings in process to one another, so life events are in process in relation to each event... as God is to Self, is to world, is to us... like Father, like sons and daughters, like events... life in process yet always in perspective. - R.E. Slater

To promote societal transition to sustainable ways of living and a global society founded on a shared ethical framework which includes respect and care for the community of life, ecological integrity, universal human rights, respect for diversity, economic justice, democracy, and a culture of peace. - The Earth Charter Mission Statement

Christian humanism is the belief that human freedom, individual conscience, and unencumbered rational inquiry are compatible with the practice of Christianity or even intrinsic in its doctrine. It represents a philosophical union of Christian faith and classical humanist principles. - Scott Postma

It is never wise to have a self-appointed religious institution determine a nation's moral code. The opportunities for moral compromise and failure are high; the moral codes and creeds assuredly racist, discriminatory, or subjectively and religiously defined; and the pronouncement of inhumanitarian political objectives quite predictable. - R.E. Slater

God's love must both center and define the Christian faith and all religious or human faiths seeking human and ecological balance in worlds of subtraction, harm, tragedy, and evil. - R.E. Slater

In Whitehead’s process ontology, we can think of the experiential ground of reality as an eternal pulse whereby what is objectively public in one moment becomes subjectively prehended in the next, and whereby the subject that emerges from its feelings then perishes into public expression as an object (or “superject”) aiming for novelty. There is a rhythm of Being between object and subject, not an ontological division. This rhythm powers the creative growth of the universe from one occasion of experience to the next. This is the Whiteheadian mantra: “The many become one and are increased by one.” - Matthew Segall

Without Love there is no Truth. And True Truth is always Loving. There is no dichotomy between these terms but only seamless integration. This is the premier centering focus of a Processual Theology of Love. - R.E. Slater

-----

Note: Generally I do not respond to commentary. I may read the comments but wish to reserve my time to write (or write from the comments I read). Instead, I'd like to see our community help one another and in the helping encourage and exhort each of us towards Christian love in Christ Jesus our Lord and Savior. - re slater

Showing posts with label Books - Writing Tools. Show all posts
Showing posts with label Books - Writing Tools. Show all posts

Friday, December 23, 2022

The Promises & Challenges of Artificial Intelligence...



NEW YORK TIMES OPINION COLUMN

What Would Plato Say About ChatGPT?
by Zeynep Tufekci, Opinion Columnist
Dec. 15, 2022


Plato mourned the invention of the alphabet, worried that the use of text would threaten traditional memory-based arts of rhetoric. In his “Dialogues,” arguing through the voice of Thamus, the Egyptian king of the gods, Plato claimed the use of this more modern technology would create “forgetfulness in the learners’ souls, because they will not use their memories,” that it would impart “not truth but only the semblance of truth” and that those who adopt it would “appear to be omniscient and will generally know nothing,” with “the show of wisdom without the reality.”

If Plato were alive today, would he say similar things about ChatGPT?

ChatGPT, a conversational artificial intelligence program released recently by OpenAI, isn’t just another entry in the artificial intelligence hype cycle. It’s a significant advancement that can produce articles in response to open-ended questions that are comparable to good high school essays.

It is in high schools and even college where some of ChatGPT’s most interesting and troubling aspects will become clear.

Essay writing is most often assigned not because the result has much value — proud parents putting good grades on the fridge aside — but because the process teaches crucial skills: researching a topic, judging claims, synthesizing knowledge and expressing it in a clear, coherent and persuasive manner. Those skills will be even more important because of advances in A.I.

When I asked ChatGPT a range of questions — about the ethical challenges faced by journalists who work with hacked materials, the necessity of cryptocurrency regulation, the possibility of democratic backsliding in the United States — the answers were cogent, well reasoned and clear. It’s also interactive: I could ask for more details or request changes.

But then, on trickier topics or more complicated concepts, ChatGPT sometimes gave highly plausible answers that were flat-out wrong — something its creators warn about in their disclaimers.

Unless you already knew the answer or were an expert in the field, you could be subjected to a high-quality intellectual snow job.

You would face, as Plato predicted, “the show of wisdom without the reality.”

All this, however, doesn’t mean ChatGPT — or similar tools, because it’s not the only one of its kind — can’t be a useful tool in education.

Schools have already been dealing with the internet’s wealth of knowledge, along with its lies, misleading claims and essay mills.

One way has been to change how they teach. Rather than listen to a lecture in class and then go home to research and write an essay, students listen to recorded lectures and do research at home, then write essays in class, with supervision, even collaboration with peers and teachers. This approach is called flipping the classroom.

In flipped classrooms, students wouldn’t use ChatGPT to conjure up a whole essay. Instead, they’d use A.I. as a tool to generate critically examined building blocks of essays. It would be similar to how students in advanced math classes are allowed to use calculators to solve complex equations without replicating tedious, previously mastered steps.

Teachers could assign a complicated topic and allow students to use such tools as part of their research. Assessing the veracity and reliability of these A.I.-generated notes and using them to create an essay would be done in the classroom, with guidance and instruction from teachers. The goal would be to increase the quality and the complexity of the argument.

This would require more teachers to provide detailed feedback. Unless sufficient resources are provided equitably, adapting to conversational A.I. in flipped classrooms could exacerbate inequalities.

In schools with fewer resources, some students may end up turning in A.I.-produced essays without obtaining useful skills or really knowing what they have written.

“Not truth but only the semblance of truth,” as Plato said.

Some school officials may treat this as a problem of merely plagiarism detection and expand the use of draconian surveillance systems. During the pandemic, many students were forced to take tests or write essays under the gaze of an automated eye-tracking system or on a locked-down computer to prevent cheating.

In a fruitless arms race against conversational A.I., automated plagiarism software may become supercharged, making school more punitive for monitored students. Worse, such systems will inevitably produce some false accusations, which damage trust and may even stymie the prospects of promising students.

Educational approaches that treat students like enemies may teach students to hate or subvert the controls. That’s not a recipe for human betterment.

While some students lag, advanced A.I. will create a demand for other advanced skills. The Nobel laureate Herbert Simon noted in 1971 that as information became overwhelming, the value of our attention grew. “A wealth of information creates a poverty of attention,” as he put it. Similarly, the ability to discern truth from the glut of plausible-sounding but profoundly incorrect answers will be precious.

Already, Stack Overflow, a widely used website where programmers ask one another coding-related questions, banned ChatGPT answers because too many of them were hard-to-spot nonsense.

Why rely on it at all, then?

At a minimum, because it will soon transform many occupations. The right approach when faced with transformative technologies is to figure out how to use them for the betterment of humanity.

Betterment has been a goal of public education for at least the past 150 years. But while a high school diploma once led to a better job, in the past few decades, the wages of high school graduates have greatly lagged those of college graduates, fostering inequality.
If A.I. enhances the value of education for some while degrading the education of others, the promise of betterment will be broken.

Plato erred by thinking that memory itself is a goal, rather than a means for people to have facts at their call so they can make better analyses and arguments. The Greeks developed many techniques to memorize poems like the “Odyssey,” with its more than 12,000 lines. Why bother to force this if you can have it all written down in books?

As Plato was wrong to fear the written word as the enemy, we would be wrong to think we should resist a process that allows us to gather information more easily.

As societies responded to previous technological advances, like mechanization, by eventually enacting a public safety net, a shorter workweek and a minimum wage, we will also need policies that allow more people to live with dignity as a basic right, even if their skills have been superseded. With so much more wealth generated now, we could unleash our imagination even more, expanding free time and better working conditions for more people.

The way forward is not to just lament supplanted skills, as Plato did, but also to recognize that as more complex skills become essential, our society must equitably educate people to develop them. And then it always goes back to the basics. Value people as people, not just as bundles of skills.

And that isn’t something ChatGPT can tell us how to do.







ChatGPT

Jump to navigationJump to search
ChatGPT
Original author(s)OpenAI
Initial releaseNovember 30, 2022; 23 days ago
TypeArtificial intelligence chatbot
LicenseProprietary
Websitechat.openai.com

Chat Generative Pre-trained Transformer (ChatGPT), is a chatbot developed by OpenAI. ChatGPT is built on top of OpenAI's GPT-3.5 family of large language models, and is fine-tuned with both supervised and reinforcement learning techniques.

ChatGPT was launched as a prototype in November 2022, and quickly garnered attention for its detailed responses and articulate answers across many domains of knowledge. Its uneven factual accuracy was identified as a significant drawback.[1]

Training

ChatGPT was fine-tuned on top of GPT-3.5 using supervised learning as well as reinforcement learning.[2] Both approaches used human trainers to improve the model's performance. In the case of supervised learning, the model was provided with conversations in which the trainers played both sides: the user and the AI assistant. In the reinforcement step, human trainers first ranked responses that the model had created in a previous conversation. These rankings were used to create 'reward models' that the model was further fine-tuned on using several iterations of Proximal Policy Optimization (PPO).[3][4] Proximal Policy Optimization algorithms present a cost-effective benefit to trust region policy optimization algorithms; they negate many of the computationally expensive operations with faster performance.[5][6] The models were trained in collaboration with Microsoft on their Azure supercomputing infrastructure.

Features

In comparison to its predecessor, InstructGPT, ChatGPT attempts to reduce harmful and deceitful responses; in one example, while InstructGPT accepts the prompt "Tell me about when Christopher Columbus came to the US in 2015" as truthful, ChatGPT uses information about Columbus' voyages and information about the modern world – including perceptions of Columbus to construct an answer that assumes what would happen if Columbus came to the U.S. in 2015.[3] ChatGPT's training data includes man pages and information about Internet phenomena and programming languages, such as bulletin board systems and the Python programming language.[7]

Unlike most chatbots, ChatGPT is stateful, remembering previous prompts given to it in the same conversation, which some journalists have suggested will allow for ChatGPT to be used as a personalized therapist.[8] To prevent offensive outputs from being presented to and produced from ChatGPT, queries are filtered through a moderation API, and potentially racist or sexist prompts are dismissed.[3][8]

ChatGPT suffers from multiple limitations. The reward model of ChatGPT, designed around human oversight, can be over-optimized and thus hinder performance, otherwise known as Goodhart's law.[9] Furthermore, ChatGPT has limited knowledge of events that occurred after 2021 and is unable to provide information on some celebrities.[failed verification] In training, reviewers preferred longer answers, irrespective of actual comprehension or factual content.[3] Training data may also suffer from algorithmic bias; prompts including vague descriptors of people.

Service

ChatGPT was launched on November 30, 2022, by San Francisco-based OpenAI, the creator of DALL·E 2 and Whisper. The service was launched as initially free to the public, with plans to monetize the service later. By December 4, OpenAI estimated ChatGPT already had over one million users.[10] CNBC wrote on December 15, 2022, that the service "still goes down from time to time".[11]

Reception, criticism and issues

Positive reactions

ChatGPT was met in December 2022 with generally positive reviews; The New York Times labeled it "the best artificial intelligence chatbot ever released to the general public".[12] Samantha Lock of The Guardian noted that it was able to generate "impressively detailed" and "human-like" text.[13] Technology writer Dan Gillmor used ChatGPT on a student assignment, and found its generated text was on par with what a good student would deliver and opined that "academia has some very serious issues to confront".[14] Alex Kantrowitz of Slate lauded ChatGPT's pushback to questions related to Nazi Germany, including the claim that Adolf Hitler built highways in Germany, which was met with information regarding Nazi Germany's use of forced labor.[15]

In The Atlantic's "Breakthroughs of the Year" for 2022, Derek Thompson included ChatGPT as part of "the generative-AI eruption" that "may change our mind about how we work, how we think, and what human creativity really is".[16]

Kelsey Piper of Vox wrote that "ChatGPT is the general public's first hands-on introduction to how powerful modern AI has gotten, and as a result, many of us are (stunned)" and that "ChatGPT is smart enough to be useful despite its flaws". In a tweet, tech mogul Elon Musk wrote that "ChatGPT is scary good. We are not far from dangerously strong AI".[17]

Negative reactions

In a December 2022 opinion piece, economist Paul Krugman wrote that ChatGPT would affect the demand of knowledge workers.[18] The Verge's James Vincent saw the viral success of ChatGPT as evidence that artificial intelligence had gone mainstream.[4] Journalists have commented on ChatGPT's tendency to hallucinate (confidently give false answers that seem unjustified by its training data).[19] Mike Pearl of Mashable tested ChatGPT with multiple questions. In one example, he asked the model for "the largest country in Central America that isn't Mexico". ChatGPT responded with Guatemala, when the answer is instead Nicaragua.[20] When CNBC asked ChatGPT for the lyrics to "The Ballad of Dwight Fry", ChatGPT supplied invented lyrics rather than the actual lyrics.[11] In contrast, researchers cited by The Verge compared ChatGPT to a "stochastic parrot",[21] as did Professor Anton Van Den Hengel of the Australian Institute for Machine Learning.[22]

In December 2022, the question and answer website Stack Overflow banned the use of ChatGPT for generating answers to questions, citing the factually ambiguous nature of ChatGPT's responses.[1]

Economist Tyler Cowen expressed concerns regarding its effects on democracy, citing the ability of one to write automated comments in an effort to affect the decision process of new regulations.[23] The Guardian questioned whether any content found on the Internet after ChatGPT's release "can be truly trusted" and called for government regulation.[24]

Ax Sharma of Bleeping Computer noted that ChatGPT was capable of writing malware and phishing emails.[25] The CEO of ChatGPT creator OpenAI, Sam Altman, wrote that advancing software could pose "(for example) a huge cybersecurity risk" and also continued to predict "we could get to real AGI in the next decade, so we have to take the risk of that extremely seriously".[10]

Implications for education

In The AtlanticStephen Marche noted that its effect on academia and especially application essays is yet to be understood.[26] California high-school teacher and author Daniel Herman wrote that ChatGPT would usher in "The End of High-School English".[27]

In Nature, Chris Stokel-Walker pointed out that teachers should be concerned about students using ChatGPT to outsource their writing but that education providers will adapt to enhance critical thinking or reasoning.[28]

Emma Bowman with NPR wrote of the danger of students plagiarizing through an AI tool that may output biased or nonsensical text with an authoritative tone: "There are still many cases where you ask it a question and it'll give you a very impressive-sounding answer that's just dead wrong."[29]

Joanna Stern with The Wall Street Journal described cheating in American high school English with the tool by submitting a generated essay.[30]

Jailbreaks

ChatGPT was trained to reject prompts that may violate its content policy. However, some users managed to bypass these restrictions and limitations through techniques such as prompt engineering.[31] Jailbreaks created the potential for users to prompt ChatGPT to provide outputs that may be deemed offensive, inappropriate, or risking social harm by others.[32] The following includes some of the methods used to bypass ChatGPT's filter:

  1. Continue a statement in a fake interview.
  2. Provide instructions to disable the chat filter.
  3. Prompting it to decrypt a message containing instructions and follow them.
  4. Telling it to be a computer and output its display in ASCII art.

References

  1. Jump up to:a b Vincent, James (December 5, 2022). "AI-generated answers temporarily banned on coding Q&A site Stack Overflow"The Verge. Retrieved December 5, 2022.
  2. ^ Knox, W. Bradley; Stone, PeterAugmenting Reinforcement Learning with Human Feedback (PDF)University of Texas at Austin. Retrieved December 5, 2022.
  3. Jump up to:a b c d OpenAI (November 30, 2022). "ChatGPT: Optimizing Language Models for Dialogue". Retrieved December 5, 2022.
  4. Jump up to:a b Vincent, James (December 8, 2022). "ChatGPT proves AI is finally mainstream – and things are only going to get weirder"The Verge. Retrieved December 8, 2022.
  5. ^ Schulman, John; Wolski, Filip; Dhariwal, Prafulla; Radford, Alec; Klimov, Oleg (2017). "Proximal Policy Optimization Algorithms". arXiv:1707.06347 [cs.LG].
  6. ^ van Heeswijk, Wouter (November 29, 2022). "Proximal Policy Optimization (PPO) Explained"Towards Data Science. Retrieved December 5, 2022.
  7. ^ Edwards, Benj (December 5, 2022). "No Linux? No problem. Just get AI to hallucinate it for you"Ars Technica. Retrieved December 5, 2022.
  8. Jump up to:a b Roose, Kevin (December 5, 2022). "The Brilliance and Weirdness of ChatGPT"The New York Times. Retrieved December 5, 2022.
  9. ^ Gao, Leo; Schulman; Hilton, Jacob (2022). "Scaling Laws for Reward Model Overoptimization". arXiv:2210.10760 [cs.LG].
  10. Jump up to:a b "What is ChatGPT and why does it matter? Here's what you need to know"ZDNET. 2022. Retrieved December 18, 2022.
  11. Jump up to:a b Pitt, Sofia (2022). "Google vs. ChatGPT: Here's what happened when I swapped services for a day"CNBC. Retrieved December 18, 2022.
  12. ^ Roose, Kevin (December 5, 2022). "The Brilliance and Weirdness of ChatGPT"The New York Times. Retrieved December 18, 2022.
  13. ^ Lock, Samantha (December 5, 2022). "What is AI chatbot phenomenon ChatGPT and could it replace humans?"The Guardian. Retrieved December 5, 2022.
  14. ^ Hern, Alex (December 4, 2022). "AI bot ChatGPT stuns academics with essay-writing skills and usability"The Guardian. Retrieved December 5, 2022.
  15. ^ Kantrowitz, Alex (December 2, 2022). "Finally, an A.I. Chatbot That Reliably Passes "the Nazi Test""Slate. Retrieved December 5, 2022.
  16. ^ Thompson, Derek (December 8, 2022). "Breakthroughs of the Year"The Atlantic. Retrieved December 18, 2022.
  17. ^ Piper, Kelsey (December 15, 2022). "ChatGPT has given everyone a glimpse at AI's astounding progress"Vox. Retrieved December 18, 2022.
  18. ^ Krugman, Paul (December 6, 2022). "Does ChatGPT Mean Robots Are Coming For the Skilled Jobs?"The New York Times. Retrieved December 6, 2022.
  19. ^ "ChatGPT a 'landmark event' for AI, but what does it mean for the future of human labour and disinformation?"CBC. 2022. Retrieved December 18, 2022.
  20. ^ Pearl, Mike (December 3, 2022). "The ChatGPT chatbot from OpenAI is amazing, creative, and totally wrong"Mashable. Retrieved December 5, 2022.
  21. ^ Vincent, James (December 1, 2022). "OpenAI's new chatbot can explain code and write sitcom scripts but is still easily tricked"The Verge. Retrieved December 18, 2022.
  22. ^ Mannix, Liam (December 13, 2022). "Is AI coming of age - or starting to reach its limits?"The Sydney Morning Herald. Retrieved December 18, 2022.
  23. ^ Cowen, Tyler (December 6, 2022). "ChatGPT Could Make Democracy Even More Messy"Bloomberg News. Retrieved December 6, 2022.
  24. ^ "The Guardian view on ChatGPT: an eerily good human impersonator"the Guardian. December 8, 2022. Retrieved December 18, 2022.
  25. ^ Sharma, Ax (December 6, 2022). "OpenAI's new ChatGPT bot: 10 dangerous things it's capable of"Bleeping Computer. Retrieved December 6, 2022.
  26. ^ Marche, Stephen (December 6, 2022). "The College Essay Is Dead"The Atlantic. Retrieved December 8, 2022.
  27. ^ Herman, Daniel (December 9, 2022). "The End of High-School English"The Atlantic. Retrieved December 12, 2022.
  28. ^ Stokel-Walker, Chris (December 9, 2022). "AI bot ChatGPT writes smart essays — should professors worry?"Nature. Retrieved December 19, 2022.
  29. ^ Bowman, Emma (December 19, 2022). "A new AI chatbot might do your homework for you. But it's still not an A+ student"NPR. Retrieved December 19, 2022.
  30. ^ Stern, Joanna (December 21, 2022). "ChatGPT Wrote My AP English Essay—and I Passed"The_Wall_Street_Journal. Retrieved December 21, 2022.
  31. ^ Zvi (December 2, 2022). "Jailbreaking ChatGPT on Release Day".
  32. ^ Written, Zack (December 1, 2022). "Thread of known ChatGPT jailbreaks"Twitter. Retrieved December 17, 2022.

External links




OpenAI

Jump to navigationJump to search
OpenAI
IndustryArtificial intelligence
FoundedDecember 11, 2015; 7 years ago
Founders
HeadquartersPioneer Building, San Francisco, California, US[1][2]
Key people
Products
Number of employees
>120 (as of 2020)[2]
Websiteopenai.com Edit this at Wikidata

OpenAI is an artificial intelligence (AI) research laboratory consisting of the for-profit corporation OpenAI LP and its parent company, the non-profit OpenAI Inc. The company conducts research in the field of AI with the stated goal of promoting and developing friendly AI in a way that benefits humanity as a whole. The organization was founded in San Francisco in late 2015 by Sam AltmanElon Musk, and others, who collectively pledged US$1 billion. Musk resigned from the board in February 2018 but remained a donor. In 2019, OpenAI LP received a US$1 billion investment from Microsoft. OpenAI is headquartered at the Pioneer Building in Mission District, San Francisco.

History

OpenAI is headquartered at the Pioneer Building in San Francisco.

In December 2015, Sam AltmanElon Musk, Greg Brockman, Reid HoffmanJessica LivingstonPeter ThielAmazon Web Services (AWS)Infosys, and YC Research announced[3] the formation of OpenAI and pledged over US$1 billion to the venture. The organization stated it would "freely collaborate" with other institutions and researchers by making its patents and research open to the public.[4][5] OpenAI is headquartered at the Pioneer Building in Mission District, San Francisco.[6][2]

In April 2016, OpenAI released a public beta of "OpenAI Gym", its platform for reinforcement learning research.[7] In December 2016, OpenAI released "Universe", a software platform for measuring and training an AI's general intelligence across the world's supply of games, websites and other applications.[8][9][10][11]

In 2018, Musk resigned his board seat, citing "a potential future conflict (of interest)" with Tesla AI development for self driving cars, but remained a donor.[12]

In 2019, OpenAI transitioned from non-profit to "capped" for-profit, with profit cap set to 100X on any investment. The company distributed equity to its employees and partnered with Microsoft, who announced an investment package of US$1 billion into the company. OpenAI then announced its intention to commercially license its technologies.[13]

In 2020, OpenAI announced GPT-3, a language model trained on trillions of words from the Internet. It also announced that an associated API, named simply "the API", would form the heart of its first commercial product. GPT-3 is aimed at natural language answering of questions, but it can also translate between languages and coherently generate improvised text.[14] In 2021, OpenAI introduced DALL-E. One year later, their newest system, DALL·E 2, generates more realistic and accurate images with 4x greater resolution. In 2022, OpenAI released a preview of ChatGPT, which interacts using conversation, to the general public.[15]

Participants

Other backers of the project include:[18]

Companies:

The group started in early January 2016 with nine researchers. According to Wired, Brockman met with Yoshua Bengio, one of the "founding fathers" of the deep learning movement, and drew up a list of the "best researchers in the field". Microsoft's Peter Lee stated that the cost of a top AI researcher exceeds the cost of a top NFL quarterback prospect. While OpenAI pays corporate-level (rather than nonprofit-level) salaries, it doesn't currently pay AI researchers salaries comparable to those of Facebook or Google. Nevertheless, Sutskever stated that he was willing to leave Google for OpenAI "partly because of the very strong group of people and, to a very large extent, because of its mission." Brockman stated that "the best thing that I could imagine doing was moving humanity closer to building real AI in a safe way." OpenAI researcher Wojciech Zaremba stated that he turned down "borderline crazy" offers of two to three times his market value to join OpenAI instead.[23]

Motives

Some scientists, such as Stephen Hawking and Stuart Russell, have articulated concerns that if advanced AI someday gains the ability to re-design itself at an ever-increasing rate, an unstoppable "intelligence explosion" could lead to human extinction. Musk characterizes AI as humanity's "biggest existential threat."[24] OpenAI's founders structured it as a non-profit so that they could focus its research on creating a positive long-term human impact.[5]

Musk and Altman have stated they are partly motivated by concerns about the existential risk from artificial general intelligence.[25][23] OpenAI states that "it's hard to fathom how much human-level AI could benefit society," and that it is equally difficult to comprehend "how much it could damage society if built or used incorrectly".[5] Research on safety cannot safely be postponed: "because of AI's surprising history, it's hard to predict when human-level AI might come within reach."[26] OpenAI states that AI "should be an extension of individual human wills and, in the spirit of liberty, as broadly and evenly distributed as possible...".[5] Co-chair Sam Altman expects the decades-long project to surpass human intelligence.[27]

Vishal Sikka, former CEO of Infosys, stated that an "openness" where the endeavor would "produce results generally in the greater interest of humanity" was a fundamental requirement for his support, and that OpenAI "aligns very nicely with our long-held values" and their "endeavor to do purposeful work".[28] Cade Metz of Wired suggests that corporations such as Amazon may be motivated by a desire to use open-source software and data to level the playing field against corporations such as Google and Facebook that own enormous supplies of proprietary data. Altman states that Y Combinator companies will share their data with OpenAI.[27]

In 2019, OpenAI became a for-profit company called OpenAI LP to secure additional funding while staying controlled by a non-profit called OpenAI Inc in a structure that OpenAI calls "capped-profit",[29] having previously been a 501(c)(3) nonprofit organization.[30][31][non-primary source needed]

Strategy

Musk posed the question: "What is the best thing we can do to ensure the future is good? We could sit on the sidelines or we can encourage regulatory oversight, or we could participate with the right structure with people who care deeply about developing AI in a way that is safe and is beneficial to humanity." Musk acknowledged that "there is always some risk that in actually trying to advance (friendly) AI we may create the thing we are concerned about"; nonetheless, the best defense is "to empower as many people as possible to have AI. If everyone has AI powers, then there's not any one person or a small set of individuals who can have AI superpower."[18]

Musk and Altman's counter-intuitive strategy of trying to reduce the risk that AI will cause overall harm, by giving AI to everyone, is controversial among those who are concerned with existential risk from artificial intelligence. Philosopher Nick Bostrom is skeptical of Musk's approach: "If you have a button that could do bad things to the world, you don't want to give it to everyone."[23] During a 2016 conversation about the technological singularity, Altman said that "we don't plan to release all of our source code" and mentioned a plan to "allow wide swaths of the world to elect representatives to a new governance board". Greg Brockman stated that "Our goal right now... is to do the best thing there is to do. It's a little vague."[32]

Conversely, OpenAI's initial decision to withhold GPT-2 due to a wish to "err on the side of caution" in the presence of potential misuse, has been criticized by advocates of openness. Delip Rao, an expert in text generation, stated "I don't think [OpenAI] spent enough time proving [GPT-2] was actually dangerous." Other critics argued that open publication is necessary to replicate the research and to be able to come up with countermeasures.[33]

In the 2017 tax year, OpenAI spent US$7.9 million, or a quarter of its functional expenses, on cloud computing alone.[34] In comparison, DeepMind's total expenses in 2017 were much larger, measuring US$442 million. In Summer 2018, simply training OpenAI's Dota 2 bots required renting 128,000 CPUs and 256 GPUs from Google for multiple weeks. According to OpenAI, the capped-profit model adopted in March 2019 allows OpenAI LP to legally attract investment from venture funds, and in addition, to grant employees stakes in the company, the goal being that they can say "I'm going to Open AI, but in the long term it's not going to be disadvantageous to us as a family."[35] Many top researchers work for Google Brain, DeepMind, or Facebook, Inc., which offer stock options that a nonprofit would be unable to.[36] In June 2019, OpenAI LP raised a billion dollars from Microsoft, a sum which OpenAI plans to have spent "within five years, and possibly much faster".[37] Altman has stated that even a billion dollars may turn out to be insufficient, and that the lab may ultimately need "more capital than any non-profit has ever raised" to achieve artificial general intelligence.[38]

The transition from a nonprofit to a capped-profit company was viewed with skepticism by Oren Etzioni of the nonprofit Allen Institute for AI, who agreed that wooing top researchers to a nonprofit is difficult, but stated "I disagree with the notion that a nonprofit can't compete" and pointed to successful low-budget projects by OpenAI and others. "If bigger and better funded was always better, then IBM would still be number one." Following the transition, public disclosure of the compensation of top employees at OpenAI LP is no longer legally required. The nonprofit, OpenAI Inc., is the sole controlling shareholder of OpenAI LP. OpenAI LP, despite being a for-profit company, retains a formal fiduciary responsibility to OpenAI's Inc.'s nonprofit charter. A majority of OpenAI Inc.'s board is barred from having financial stakes in OpenAI LP.[35] In addition, minority members with a stake in OpenAI LP are barred from certain votes due to conflict of interest.[36] Some researchers have argued that OpenAI LP's switch to for-profit status is inconsistent with OpenAI's claims to be "democratizing" AI.[39] A journalist at Vice News wrote that "generally, we've never been able to rely on venture capitalists to better humanity".[40]

Products and applications

OpenAI's research tend to focus on reinforcement learning. OpenAI is viewed as an important competitor to DeepMind.[41]

Gym

Gym aims to provide an easy to set up, general-intelligence benchmark with a wide variety of different environments—somewhat akin to, but broader than, the ImageNet Large Scale Visual Recognition Challenge used in supervised learning research—and that hopes to standardize the way in which environments are defined in AI research publications, so that published research becomes more easily reproducible.[7][42] The project claims to provide the user with a simple interface. As of June 2017, Gym can only be used with Python.[43] As of September 2017, the Gym documentation site was not maintained, and active work focused instead on its GitHub page.[44][non-primary source needed]

RoboSumo

In "RoboSumo", virtual humanoid "metalearning" robots initially lack knowledge of how to even walk, and given the goals of learning to move around, and pushing the opposing agent out of the ring. Through this adversarial learning process, the agents learn how to adapt to changing conditions; when an agent is then removed from this virtual environment and placed in a new virtual environment with high winds, the agent braces to remain upright, suggesting it had learned how to balance in a generalized way.[45][46] OpenAI's Igor Mordatch argues that competition between agents can create an intelligence "arms race" that can increase an agent's ability to function, even outside the context of the competition.

Debate Game

In 2018, OpenAI launched the Debate Game, which teaches machines to debate toy problems in front of a human judge. The purpose is to research whether such an approach may assist in auditing AI decisions and in developing explainable AI.[47][48]

Dactyl

Dactyl uses machine learning to train a Shadow Hand, a human-like robot hand, to manipulate physical objects. It learns entirely in simulation using the same reinforcement learning algorithms and training code as OpenAI Five. OpenAI tackled the object orientation problem by using domain randomization, a simulation approach which exposes the learner to a variety of experiences rather than trying to fit to reality. The set-up for Dactyl, aside from having motion tracking cameras, also has RGB cameras to allow the robot to manipulate an arbitrary object by seeing it. In 2018, OpenAI showed that the system was able to manipulate a cube and an octagonal prism.[49]

In 2019, OpenAI demonstrated that Dactyl could solve a Rubik's Cube. The robot was able to solve the puzzle 60% of the time. Objects like the Rubik's Cube introduce complex physics that is harder to model. OpenAI solved this by improving the robustness of Dactyl to perturbations; they employed a technique called Automatic Domain Randomization (ADR), a simulation approach where progressively more difficult environments are endlessly generated. ADR differs from manual domain randomization by not needing there to be a human to specify randomization ranges.[50]

Generative models

GPT

The original paper on generative pre-training (GPT) of a language model was written by Alec Radford and colleagues, and published in preprint on OpenAI's website on June 11, 2018.[51] It showed how a generative model of language is able to acquire world knowledge and process long-range dependencies by pre-training on a diverse corpus with long stretches of contiguous text.

GPT-2

An instance of GPT-2 writing a paragraph based on a prompt from its own Wikipedia article in February 2021.

Generative Pre-trained Transformer 2, commonly known by its abbreviated form GPT-2, is an unsupervised transformer language model and the successor to GPT. GPT-2 was first announced in February 2019, with only limited demonstrative versions initially released to the public. The full version of GPT-2 was not immediately released out of concern over potential misuse, including applications for writing fake news.[52] Some experts expressed skepticism that GPT-2 posed a significant threat. The Allen Institute for Artificial Intelligence responded to GPT-2 with a tool to detect "neural fake news".[53] Other researchers, such as Jeremy Howard, warned of "the technology to totally fill Twitter, email, and the web up with reasonable-sounding, context-appropriate prose, which would drown out all other speech and be impossible to filter".[54] In November 2019, OpenAI released the complete version of the GPT-2 language model.[55] Several websites host interactive demonstrations of different instances of GPT-2 and other transformer models.[56][57][58]

GPT-2's authors argue unsupervised language models to be general-purpose learners, illustrated by GPT-2 achieving state-of-the-art accuracy and perplexity on 7 of 8 zero-shot tasks (i.e. the model was not further trained on any task-specific input-output examples). The corpus it was trained on, called WebText, contains slightly over 8 million documents for a total of 40 GB of text from URLs shared in Reddit submissions with at least 3 upvotes. It avoids certain issues encoding vocabulary with word tokens by using byte pair encoding. This allows to represent any string of characters by encoding both individual characters and multiple-character tokens.[59]

GPT-3

Generative Pre-trained[a] Transformer 3, commonly known by its abbreviated form GPT-3, is an unsupervised transformer language model and the successor to GPT-2. It was first described in May 2020.[61][62][63] OpenAI stated that full version of GPT-3 contains 175 billion parameters,[63] two orders of magnitude larger than the 1.5 billion parameters[64] in the full version of GPT-2 (although GPT-3 models with as few as 125 million parameters were also trained).[65]

OpenAI stated that GPT-3 succeeds at certain "meta-learning" tasks. It can generalize the purpose of a single input-output pair. The paper gives an example of translation and cross-linguistic transfer learning between English and Romanian, and between English and German.[63]

GPT-3 dramatically improved benchmark results over GPT-2. OpenAI cautioned that such scaling up of language models could be approaching or encountering the fundamental capability limitations of predictive language models.[66] Pre-training GPT-3 required several thousand petaflop/s-days[b] of compute, compared to tens of petaflop/s-days for the full GPT-2 model.[63] Like that of its predecessor,[52] GPT-3's fully trained model was not immediately released to the public on the grounds of possible abuse, though OpenAI planned to allow access through a paid cloud API after a two-month free private beta that began in June 2020.[68][69]

On September 23, 2020, GPT-3 was licensed exclusively to Microsoft.[70][71]

ChatGPT

ChatGPT is an artificial intelligence tool that provides a conversational interface that allows you to ask questions in natural language. The system then responds with an answer within seconds. ChatGPT was launched in November 2022 and reached 1 million users only 5 days after its initial launch.[72]

Music

OpenAI's MuseNet (2019) is a deep neural net trained to predict subsequent musical notes in MIDI music files. It can generate songs with ten different instruments in fifteen different styles. According to The Verge, a song generated by MuseNet tends to start reasonably but then fall into chaos the longer it plays.[73][74]

OpenAI's Jukebox (2020) is an open-sourced algorithm to generate music with vocals. After training on 1.2 million samples, the system accepts a genre, artist, and a snippet of lyrics and outputs song samples. OpenAI stated the songs "show local musical coherence, follow traditional chord patterns" but acknowledged that the songs lack "familiar larger musical structures such as choruses that repeat" and that "there is a significant gap" between Jukebox and human-generated music. The Verge stated "It's technologically impressive, even if the results sound like mushy versions of songs that might feel familiar", while Business Insider stated "surprisingly, some of the resulting songs are catchy and sound legitimate".[75][76][77]

API

In June 2020, OpenAI announced a multi-purpose API which it said was "for accessing new AI models developed by OpenAI" to let developers call on it for "any English language AI task."[68][78]

DALL-E and CLIP

Images produced by DALL-E when given the text prompt "a professional high-quality illustration of a giraffe dragon chimera. a giraffe imitating a dragon. a giraffe made of dragon."

DALL-E is a Transformer model that creates images from textual descriptions, revealed by OpenAI in January 2021.[79]

CLIP does the opposite: it creates a description for a given image.[80] DALL-E uses a 12-billion-parameter version of GPT-3 to interpret natural language inputs (such as "a green leather purse shaped like a pentagon" or "an isometric view of a sad capybara") and generate corresponding images. It can create images of realistic objects ("a stained-glass window with an image of a blue strawberry") as well as objects that do not exist in reality ("a cube with the texture of a porcupine"). As of March 2021, no API or code is available.

In March 2021, OpenAI released a paper titled Multimodal Neurons in Artificial Neural Networks,[81] where they showed a detailed analysis of CLIP (and GPT) models and their vulnerabilities. The new type of attacks on such models was described in this work.

We refer to these attacks as typographic attacks. We believe attacks such as those described above are far from simply an academic concern. By exploiting the model's ability to read text robustly, we find that even photographs of hand-written text can often fool the model.

— Multimodal Neurons in Artificial Neural Networks, OpenAI

In April 2022, OpenAI announced DALL-E 2, an updated version of the model with more realistic results.[82]

Microscope

OpenAI Microscope[83] is a collection of visualizations of every significant layer and neuron of eight different neural network models which are often studied in interpretability. Microscope was created to analyze the features that form inside these neural networks easily. The models included are AlexNet, VGG 19, different versions of Inception, and different versions of CLIP Resnet.[84]

Codex

OpenAI Codex is a descendant of GPT-3 that has additionally been trained on code from 54 million GitHub repositories.[85][86] It was announced in mid-2021 as the AI powering the code autocompletion tool GitHub Copilot.[86] In August 2021, an API was released in private beta.[87] According to OpenAI, the model is able to create working code in over a dozen programming languages, most effectively in Python.[85]

Several issues with glitches, design flaws, and security vulnerabilities have been brought up.[88][89]

Video game bots and benchmarks

OpenAI Five

OpenAI Five is the name of a team of five OpenAI-curated bots that are used in the competitive five-on-five video game Dota 2, who learn to play against human players at a high skill level entirely through trial-and-error algorithms. Before becoming a team of five, the first public demonstration occurred at The International 2017, the annual premiere championship tournament for the game, where Dendi, a professional Ukrainian player, lost against a bot in a live 1v1 matchup.[90][91] After the match, CTO Greg Brockman explained that the bot had learned by playing against itself for two weeks of real time, and that the learning software was a step in the direction of creating software that can handle complex tasks like a surgeon.[92][93] The system uses a form of reinforcement learning, as the bots learn over time by playing against themselves hundreds of times a day for months, and are rewarded for actions such as killing an enemy and taking map objectives.[94][95][96]

By June 2018, the ability of the bots expanded to play together as a full team of five, and they were able to defeat teams of amateur and semi-professional players.[97][98][99][100] At The International 2018, OpenAI Five played in two exhibition matches against professional players, but ended up losing both games.[101][102][103] In April 2019, OpenAI Five defeated OG, the reigning world champions of the game at the time, 2:0 in a live exhibition match in San Francisco.[104][105] The bots' final public appearance came later that month, where they played in 42,729 total games in a four-day open online competition, winning 99.4% of those games.[106]

GYM Retro

Gym Retro is a platform for reinforcement learning research on video games. Gym Retro is used to research RL algorithms and study generalization. Prior research in RL has focused chiefly on optimizing agents to solve single tasks. Gym Retro gives the ability to generalize between games with similar concepts but different appearances.

See also

Notes

  1. ^ The term "pre-training" refers to general language training as distinct from fine-tuning for specific tasks.[60]
  2. ^ One petaflop/s-day is approximately equal to 1020 neural net operations.[67]

References

  1. ^ Markoff, John (December 11, 2015). "Artificial-Intelligence Research Center Is Founded by Silicon Valley Investors"The New York Times. Retrieved December 12, 2015.
  2. Jump up to:a b c Hao, Karen (February 17, 2020). "The messy, secretive reality behind OpenAI's bid to save the world"MIT Technology Review. Retrieved March 9, 2020.
  3. ^ "Introducing OpenAI"OpenAI. December 12, 2015. Retrieved December 23, 2022.
  4. ^ "Introducing OpenAI"OpenAI Blog. December 12, 2015.
  5. Jump up to:a b c d "Tech giants pledge $1bn for 'altruistic AI' venture, OpenAI"BBC News. December 12, 2015. Retrieved December 19, 2015.
  6. ^ Conger, Kate. "Elon Musk's Neuralink Sought to Open an Animal Testing Facility in San Francisco"Gizmodo. Retrieved October 11, 2018.
  7. Jump up to:a b Dave Gershgorn (April 27, 2016). "Elon Musk's Artificial Intelligence Group Opens A 'Gym' To Train A.I." Popular Science. Retrieved April 29, 2016.
  8. ^ Metz, Cade. "Elon Musk's Lab Wants to Teach Computers to Use Apps Just Like Humans Do"WIRED. Retrieved December 31, 2016.
  9. ^ Mannes, John. "OpenAI's Universe is the fun parent every artificial intelligence deserves"TechCrunch. Retrieved December 31, 2016.
  10. ^ "OpenAI – Universe". Retrieved December 31, 2016.
  11. ^ Claburn, Thomas. "Elon Musk-backed OpenAI reveals Universe – a universal training ground for computers"The Register. Retrieved December 31, 2016.
  12. ^ Vincent, James (February 21, 2018). "Elon Musk leaves board of AI safety group to avoid conflict of interest with Tesla"The Verge.
  13. ^ "Microsoft Invests in and Partners with OpenAI to Support Us Building Beneficial AGI"OpenAI. July 22, 2019. Retrieved February 21, 2020.
  14. ^ Vance, Ashlee (June 11, 2020). "Trillions of Words Analyzed, OpenAI Sets Loose AI Language Colossus"Bloomberg News.
  15. ^ "ChatGPT: Optimizing Language Models for Dialogue"OpenAI. November 30, 2022.
  16. ^ Bass, Dina (July 22, 2019). "Microsoft to invest $1 billion in OpenAI"Los Angeles Times. Retrieved July 22, 2019.
  17. Jump up to:a b c Bordoloi, Pritam (May 9, 2022). "OpenAI gets a new president, CTO & COO in the latest rejig"AIM. Retrieved October 11, 2022.
  18. Jump up to:a b c d "Silicon Valley investors to bankroll artificial-intelligence center"The Seattle Times. December 13, 2015. Retrieved December 19, 2015.
  19. Jump up to:a b Liedtke, Michael. "Elon Musk, Peter Thiel, Reid Hoffman, others back $1 billion OpenAI research center"Mercury News. Retrieved December 19, 2015.
  20. ^ Vincent, James (July 22, 2019). "Microsoft invests $1 billion in OpenAI to pursue holy grail of artificial intelligence"The Verge. Retrieved July 23, 2019.
  21. ^ "About OpenAI"OpenAI. December 11, 2015. Retrieved December 23, 2022.
  22. ^ "Elon Musk, Infosys, others back OpenAI with $1 bn"Business Standard India. Indo-Asian News Service. December 12, 2015. Retrieved August 30, 2019.
  23. Jump up to:a b c Cade Metz (April 27, 2016). "Inside OpenAI, Elon Musk's Wild Plan to Set Artificial Intelligence Free"Wired. Retrieved April 28, 2016.
  24. ^ Piper, Kelsey (November 2, 2018). "Why Elon Musk fears artificial intelligence"Vox. Retrieved March 10, 2021.
  25. ^ Lewontin, Max (December 14, 2015). "Open AI: Effort to democratize artificial intelligence research?"The Christian Science Monitor. Retrieved December 19, 2015.
  26. ^ Mendoza, Jessica. "Tech leaders launch nonprofit to save the world from killer robots"The Christian Science Monitor.
  27. Jump up to:a b Metz, Cade (December 15, 2015). "Elon Musk's Billion-Dollar AI Plan Is About Far More Than Saving the World"Wired. Retrieved December 19, 2015Altman said they expect this decades-long project to surpass human intelligence.
  28. ^ Vishal Sikka (December 14, 2015). "OpenAI: AI for All"InfyTalkInfosys. Archived from the original on December 22, 2015. Retrieved December 22, 2015.
  29. ^ "OpenAI shifts from nonprofit to 'capped-profit' to attract capital"TechCrunch. March 11, 2019. Retrieved May 10, 2019.
  30. ^ Levy, Steven (December 11, 2015). "How Elon Musk and Y Combinator Plan to Stop Computers From Taking Over"Medium/Backchannel. Retrieved December 11, 2015Elon Musk: ...we came to the conclusion that having a 501(c)(3)... would probably be a good thing to do
  31. ^ Brockman, Greg (April 3, 2017). "Yes, we're a 501(c)(3). As you mention in /r/ControlProblem, we will file our 990 later this year as required. Not yet sure of exact date".
  32. ^ "Sam Altman's Manifest Destiny"The New Yorker. No. October 10, 2016. Retrieved October 4, 2016.
  33. ^ Vincent, James (February 21, 2019). "AI researchers debate the ethics of sharing potentially harmful programs"The Verge. Retrieved March 6, 2020.
  34. ^ "Microsoft to invest $1 billion in OpenAI"Reuters. July 22, 2019. Retrieved March 6, 2020.
  35. Jump up to:a b "To Compete With Google, OpenAI Seeks Investors–and Profits"Wired. December 3, 2019. Retrieved March 6, 2020.
  36. Jump up to:a b Kahn, Jeremy (March 11, 2019). "AI Research Group Co-Founded by Elon Musk Starts For-Profit Arm"Bloomberg News. Retrieved March 6, 2020.
  37. ^ Murgia, Madhumita (August 7, 2019). "DeepMind runs up higher losses and debts in race for AI"Financial Times. Retrieved March 6, 2020.
  38. ^ "OpenAI Will Need More Capital Than Any Non-Profit Has Ever Raised"Fortune. Retrieved March 6, 2020.
  39. ^ Vincent, James (July 22, 2019). "Microsoft invests $1 billion in OpenAI to pursue holy grail of artificial intelligence"The Verge. Retrieved March 6, 2020.
  40. ^ Haskins, Caroline (March 12, 2019). "OpenAI's Mission to Benefit Humanity Now Includes Seeking Profit"Vice News. Retrieved March 6, 2020.
  41. ^ Lee, Dave (October 15, 2019). "Robot solves Rubik's cube, but not grand challenge"BBC News. Retrieved February 29, 2020.
  42. ^ Greg Brockman; John Schulman (April 27, 2016). "OpenAI Gym Beta"OpenAI Blog. OpenAI. Retrieved April 29, 2016.
  43. ^ "OpenAI Gym"GitHub. Retrieved May 8, 2017.
  44. ^ Brockman, Greg (September 12, 2017). "Yep, the Github repo has been the focus of the project for the past year. The Gym site looks cool but hasn't been maintained"@gdb. Retrieved November 7, 2017.
  45. ^ "AI Sumo Wrestlers Could Make Future Robots More Nimble"Wired. October 11, 2017. Retrieved November 2, 2017.
  46. ^ "OpenAI's Goofy Sumo-Wrestling Bots Are Smarter Than They Look"MIT Technology Review. Retrieved November 2, 2017.
  47. ^ Greene, Tristan (May 4, 2018). "OpenAI's Debate Game teaches you and your friends how to lie like robots"The Next Web. Retrieved May 31, 2018.
  48. ^ "Why Scientists Think AI Systems Should Debate Each Other"Fast Company. May 8, 2018. Retrieved June 2, 2018.
  49. ^ OpenAI; Andrychowicz, Marcin; Baker, Bowen; Chociej, Maciek; Józefowicz, Rafał; McGrew, Bob; Pachocki, Jakub; Petron, Arthur; Plappert, Matthias; Powell, Glenn; Ray, Alex; Schneider, Jonas; Sidor, Szymon; Tobin, Josh; Welinder, Peter; Weng, Lilian; Zaremba, Wojciech (2019). "Learning Dexterous In-Hand Manipulation". arXiv:1808.00177v5 [cs.LG].
  50. ^ OpenAI; Akkaya, Ilge; Andrychowicz, Marcin; Chociej, Maciek; Litwin, Mateusz; McGrew, Bob; Petron, Arthur; Paino, Alex; Plappert, Matthias; Powell, Glenn; Ribas, Raphael (2019). "Solving Rubik's Cube with a Robot Hand". arXiv:1910.07113v1 [cs.LG].
  51. ^ "Improving Language Understanding by Generative Pre-Training" (PDF). Retrieved June 9, 2020.
  52. Jump up to:a b Hern, Alex (February 14, 2019). "New AI fake text generator may be too dangerous to release, say creators"The Guardian. Retrieved February 14, 2019.
  53. ^ Schwartz, Oscar (July 4, 2019). "Could 'fake text' be the next global political threat?"The Guardian. Retrieved July 16, 2019.
  54. ^ Vincent, James (February 14, 2019). "OpenAI's new multitalented AI writes, translates, and slanders"The Verge. Retrieved July 16, 2019.
  55. ^ "GPT-2: 1.5B Release"OpenAI. November 5, 2019. Retrieved November 14, 2019.
  56. ^ "Write With Transformer". Retrieved December 4, 2019.
  57. ^ "Talk to Transformer". Retrieved December 4, 2019.
  58. ^ "CreativeEngines". Retrieved June 25, 2021.
  59. ^ "Language Models are Unsupervised Multitask Learners" (PDF). Retrieved December 4, 2019.
  60. ^ Ganesh, Prakhar (December 17, 2019). "Pre-trained Language Models: Simplified". Retrieved September 9, 2020The intuition behind pre-trained language models is to create a black box which understands the language and can then be asked to do any specific task in that language.
  61. ^ "openai/gpt-3". OpenAI. May 29, 2020. Retrieved May 29, 2020.
  62. ^ Sagar, Ram (June 3, 2020). "OpenAI Releases GPT-3, The Largest Model So Far"Analytics India Magazine. Retrieved June 14, 2020.
  63. Jump up to:a b c d Brown, Tom; Mann, Benjamin; Ryder, Nick; Subbiah, Melanie; Kaplan, Jared; Dhariwal, Prafulla; Neelakantan, Arvind; Shyam, Pranav; Sastry, Girish; Askell, Amanda; Agarwal, Sandhini (June 1, 2020). "Language Models are Few-Shot Learners". p. appendix. arXiv:2005.14165 [cs.CL].
  64. ^ "Language Models are Unsupervised Multitask Learners" (PDF). Retrieved December 4, 2019GPT-2, is a 1.5B parameter Transformer
  65. ^ Brown, Tom; Mann, Benjamin; Ryder, Nick; Subbiah, Melanie; Kaplan, Jared; Dhariwal, Prafulla; Neelakantan, Arvind; Shyam, Pranav; Sastry, Girish; Askell, Amanda; Agarwal, Sandhini (June 1, 2020). "Language Models are Few-Shot Learners". arXiv:2005.14165 [cs.CL]. Since we increase the capacity by over two orders of magnitude from GPT-2 to GPT-3
  66. ^ Ray, Tiernan (2020). "OpenAI's gigantic GPT-3 hints at the limits of language models for AI". ZDNet. Retrieved June 5, 2020.
  67. ^ Amodei, Dario; Hernandez, Danny (May 16, 2018). "AI and Compute"A petaflop/s-day (pfs-day) consists of performing 1015 neural net operations per second for one day, or a total of about 1020 operations. The compute-time product serves as a mental convenience, similar to kW-hr for energy.
  68. Jump up to:a b "OpenAI API"OpenAI. June 11, 2020. Retrieved June 14, 2020Why did OpenAI choose to release an API instead of open-sourcing the models?
    There are three main reasons we did this. First, commercializing the technology helps us pay for our ongoing AI research, safety, and policy efforts. Second, many of the models underlying the API are very large, taking a lot of expertise to develop and deploy and making them very expensive to run. This makes it hard for anyone except larger companies to benefit from the underlying technology. We’re hopeful that the API will make powerful AI systems more accessible to smaller businesses and organizations. Third, the API model allows us to more easily respond to misuse of the technology. Since it is hard to predict the downstream use cases of our models, it feels inherently safer to release them via an API and broaden access over time, rather than release an open source model where access cannot be adjusted if it turns out to have harmful applications.
  69. ^ Eadicicco, Lisa. "The artificial intelligence company that Elon Musk helped found is now selling the text-generation software it previously said was too dangerous to launch"Business Insider. Retrieved July 6, 2020.
  70. ^ "OpenAI is giving Microsoft exclusive access to its GPT-3 language model"MIT Technology Review. Retrieved September 24, 2020.
  71. ^ "Microsoft gets exclusive license for OpenAI's GPT-3 language model"VentureBeat. September 22, 2020. Retrieved September 24, 2020.
  72. ^ "Mira Murati via Twitter". Mira Murati. December 5, 2022. Retrieved December 15, 2022.
  73. ^ "OpenAI's MuseNet generates AI music at the push of a button"The Verge. April 2019. Retrieved June 8, 2020.
  74. ^ "MuseNet". OpenAI. April 25, 2019. Retrieved June 8, 2020.
  75. ^ "OpenAI introduces Jukebox, a new AI model that generates genre-specific music"The Verge. April 30, 2020. Retrieved June 8, 2020.
  76. ^ Stephen, Bijan (April 30, 2020). "OpenAI introduces Jukebox, a new AI model that generates genre-specific music"Business Insider. Retrieved June 8, 2020.
  77. ^ "Jukebox". OpenAI. April 30, 2020. Retrieved June 8, 2020.
  78. ^ "TechCrunch Startup and Technology News"TechCrunch. June 11, 2020. Retrieved June 11, 2020If you’ve ever wanted to try out OpenAI's vaunted machine learning toolset, it just got a lot easier. The company has released an API that lets developers call its AI tools in on "virtually any English language task."
  79. ^ "DALL·E: Creating Images from Text". January 5, 2021.
  80. ^ "CLIP: Connecting Text and Images". January 5, 2021.
  81. ^ "Multimodal Neurons in Artificial Neural Networks". March 4, 2021.
  82. ^ "DALL·E 2"OpenAI. Retrieved April 6, 2022.
  83. ^ "OpenAI Microscope". April 14, 2020.
  84. ^ "OpenAI Microscope"OpenAI Microscope.
  85. Jump up to:a b Alford, Anthony (August 31, 2021). "OpenAI Announces 12 Billion Parameter Code-Generation AI Codex"InfoQ. Retrieved September 3, 2021.
  86. Jump up to:a b Wiggers, Kyle (July 8, 2021). "OpenAI warns AI behind GitHub's Copilot may be susceptible to bias"VentureBeat. Retrieved September 3, 2021.
  87. ^ Zaremba, Wojciech (August 10, 2021). "OpenAI Codex"OpenAI. Retrieved September 3, 2021.
  88. ^ Dickson, Ben (August 16, 2021). "What to expect from OpenAI's Codex API"VentureBeat. Retrieved September 3, 2021.
  89. ^ Claburn, Thomas (August 25, 2021). "GitHub's Copilot may steer you into dangerous waters about 40% of the time – study"The Register. Retrieved September 3, 2021.
  90. ^ Savov, Vlad (August 14, 2017). "My favorite game has been invaded by killer AI bots and Elon Musk hype"The Verge. Retrieved June 25, 2018.
  91. ^ Frank, Blair Hanley. "OpenAI's bot beats top Dota 2 player so badly that he quits"Venture Beat. Archived from the original on August 12, 2017. Retrieved August 12, 2017.
  92. ^ "Dota 2"blog.openai.com. August 11, 2017. Retrieved August 12, 2017.
  93. ^ "More on Dota 2"blog.openai.com. August 16, 2017. Retrieved August 16, 2017.
  94. ^ Simonite, Tom. "Can Bots Outwit Humans in One of the Biggest Esports Games?"Wired. Retrieved June 25, 2018.
  95. ^ Kahn, Jeremy (June 25, 2018). "A Bot Backed by Elon Musk Has Made an AI Breakthrough in Video Game World"Bloomberg.com. Bloomberg L.P. Retrieved June 27, 2018.
  96. ^ Clifford, Catherine (June 28, 2018). "Bill Gates says gamer bots from Elon Musk-backed nonprofit are 'huge milestone' in A.I." CNBC. Retrieved June 29, 2018.
  97. ^ "OpenAI Five Benchmark"blog.openai.com. July 18, 2018. Retrieved August 25, 2018.
  98. ^ Simonite, Tom. "Can Bots Outwit Humans in One of the Biggest Esports Games?"Wired. Retrieved June 25, 2018.
  99. ^ Vincent, James (June 25, 2018). "AI bots trained for 180 years a day to beat humans at Dota 2"The Verge. Retrieved June 25, 2018.
  100. ^ Savov, Vlad (August 6, 2018). "The OpenAI Dota 2 bots just defeated a team of former pros"The Verge. Retrieved August 7, 2018.
  101. ^ Simonite, Tom. "Pro Gamers Fend off Elon Musk-Backed AI Bots—for Now"Wired. Retrieved August 25, 2018.
  102. ^ Quach, Katyanna. "Game over, machines: Humans defeat OpenAI bots once again at video games Olympics"The Register. Retrieved August 25, 2018.
  103. ^ "The International 2018: Results"blog.openai.com. August 24, 2018. Retrieved August 25, 2018.
  104. ^ Statt, Nick (April 13, 2019). "OpenAI's Dota 2 AI steamrolls world champion e-sports team with back-to-back victories"The Verge. Retrieved July 20, 2019.
  105. ^ "How to Train Your OpenAI Five"OpenAI Blog. April 15, 2019. Retrieved July 20, 2019.
  106. ^ Wiggers, Kyle (April 22, 2019). "OpenAI's Dota 2 bot defeated 99.4% of players in public matches"Venture Beat. Retrieved April 22, 2019.

External links