If you are looking for an answer to the question What is Artificial Intelligence? and you only have a minute, then here's the definition the Association for the Advancement of Artificial Intelligence offers on its home page: "the scientific understanding of the mechanisms underlying thought and intelligent behavior and their embodiment in machines."
However, if you are fortunate enough to have more than a minute, then please get ready to embark upon an exciting journey exploring AI (but beware, it could last a lifetime) …
On May 2020, OpenAI introduced GPT-3 which is the third iteration of GPT language generation model series. The model boasted of a capacity of 175 billion parameters, more than 10 times than any other language model created before it. And to say that the model was a significant improvement would be an understatement. It could write essays on any topic without any inconsistencies in the output text. When trained on code samples, it could generate small code snippets (which were actually functional!) by getting a description of the task in English by the user. A group of developers trained it to update financial statements on Microsoft Excel based on casual description of transactions.
The machine learning company OpenAI is developing models that improve computer vision and can produce original images from a text prompt. Why it matters: The new models are the latest steps in ongoing efforts to create machine learning systems that exhibit elements of general intelligence, while performing tasks that are actually useful in the real world -- without breaking the bank on computing power. What's happening: OpenAI today is announcing two new systems that attempt to do for images what its landmark GPT-3 model did last year for text generation. What they're saying: "Last year, we were able to make substantial progress on text with GPT-3, but the thing is that the world isn't just built on text," says Sutskever. "This is a step towards the grander goal of building a neural network that can work in both images and text."
You've probably never wondered what a knight made of spaghetti would look like, but here's the answer anyway--courtesy of a clever new artificial intelligence program from OpenAI, a company in San Francisco. The program, DALL-E, released earlier this month, can concoct images of all sorts of weird things that don't exist, like avocado armchairs, robot giraffes, or radishes wearing tutus. OpenAI generated several images, including the spaghetti knight, at WIRED's request. DALL-E is a version of GPT-3, an AI model trained on text scraped from the web that's capable of producing surprisingly coherent text. DALL-E was fed images and accompanying descriptions; in response, it can generate a decent mashup image.
When prompted to generate "a mural of a blue pumpkin on the side of a building," OpenAI's new deep ... [ ] learning model DALL-E produces this series of original images. OpenAI has done it again. Earlier this month, OpenAI--the research organization behind last summer's much-hyped language model GPT-3--released a new AI model named DALL-E. While it has generated less buzz than GPT-3 did, DALL-E has even more profound implications for the future of AI. In a nutshell, DALL-E takes text captions as input and produces original images as output. For instance, when fed phrases as diverse as "a pentagonal green clock," "a sphere made of fire" or "a mural of a blue pumpkin on the side of a building," DALL-E is able to generate shockingly accurate visual renderings.
As progress accelerates towards AGI, the number of people who realize the significance of each new breakthrough decreases. This is the AGI Significance Paradox. There is a very old metaphor that you can boil a frog in water without it jumping out when you gradually increase the temperature. The fable goes that the frog does not have the internal models to recognize that there is a change in the water temperature. A cold-blooded creature like the frog is thought to have its temperature regulated only by the external environment.
An artist can draw a baby daikon radish wearing a tutu and walking a dog, even if they've never seen one before. But this kind of visual mashup has long been a trickier task for computers. Now, a new artificial-intelligence model can create such images with clarity -- and cuteness. This week nonprofit research company OpenAI released DALL-E, which can generate a slew of impressive-looking, often surrealistic images from written prompts such as "an armchair in the shape of an avocado" or "a painting of a capybara sitting in a field at sunrise." (And yes, the name DALL-E is a portmanteau referencing surrealist artist Salvador Dalí and animated sci-fi film "WALL-E.") A new AI model from OpenAI, DALL-E, can create pictures from the text prompt "an illustration of a baby daikon radish in a tutu walking a dog".
OpenAI's DALL-E is a neural network that is really quite good at generating images from complex text descriptions. Capable of visualizing a range of surreal concepts, like an avocado chair or a baby radish in a tutu walking a dog, DALL-E has wide implications for the future of AI. The creative ability of such models is essential in measuring AI intelligence and how well it understands the human world.
A neural network uses text captions to create outlandish images – such as armchairs in the shape of avocados – demonstrating it understands how language shapes visual culture. OpenAI, an artificial intelligence company that recently partnered with Microsoft, developed the neural network, which it calls DALL-E. It is a version of the company's GPT-3 language model that can create expansive written works based on short text prompts, but DALL-E produces images instead. "The world isn't just text," says Ilya Sutskever, co-founder of OpenAI. "Humans don't just talk: we also see. A lot of important context comes from looking."