In August 1955, a group of scientists applied for US$13,500 in funding to organize a summer workshop at Dartmouth College, New Hampshire. The area they proposed to explore was artificial intelligence (IA).
Although the request for funding was modest, the researchers conjecture was not: “Every aspect of learning or any other characteristic of intelligence can, in principle, be described with such precision that a machine can be designed to simulate it”.
Since those humble beginnings, movies and media have romanticized AI or portrayed it as a villain. Yet for most people, AI has remained a talking point and not part of a conscious lived experience.
AI has arrived in our lives
At the end of last month, AI, in the form of ChatGPTbroke free from sci-fi speculation and research labs and onto mainstream desktop computers and phones.
It’s called “generative AI” – suddenly a cleverly worded prompt can produce an essay or put together a recipe and shopping list, or create an Elvis Presley-style poem .
Although ChatGPT was the most spectacular entrant in a year of generative AI success, similar systems showed even greater potential for creating new content, with text-to-image prompts used to create images vibrant that have even won art competitions.
AI may not have enough to live on yet awareness or a theory of the mind popular in sci-fi movies and novels, but it at least comes close to disrupting what we think artificial intelligence systems can do.
Researchers working closely with these systems have swooned under the perspective of sensitivity, as in the case of Google’s LaMDA Large Language Model (LLM). An LLM is a model that has been trained to process and generate natural language.
Generative AI has also raised concerns about plagiarism, the exploitation of original content used to create models, ethics of manipulation of information and breach of trust, and even “the end of the program“.
At the center of it all is the question that has become increasingly urgent since the Dartmouth Summer Institute: Does AI differ from human intelligence?
What does “AI” really mean?
To qualify as an AI, a system must exhibit a certain level of learning and adaptation. For this reason, decision-making systems, automation and statistics are not AI.
AI is generally defined in two categories: artificial narrow intelligence (ANI) and artificial general intelligence (AGI). To date, AGI does not exist.
The main challenge in creating a general AI is to adequately model the world with all the knowledge, in a consistent and useful way. It’s a colossal undertaking, to say the least.
Most of what we call AI today has narrow intelligence – where a particular system solves a particular problem. Unlike human intelligence, such narrow AI intelligence is only effective in the domain in which it was trained: fraud detection, facial recognition or social recommendations, for example.
AGI, however, would work like humans do. For now, the most notable example of trying to achieve this is the use of neural networks and “deep learning” trained on large amounts of data.
Neural networks are inspired by the functioning of the human brain. unlike most machine learning Models that perform calculations on training data, neural networks work by feeding each data point one by one through an interconnected network, adjusting the parameters each time.
As more and more data passes through the network, the parameters stabilize; the end result is the “trained” neural network, which can then produce the desired output on new data – for example, recognizing whether an image contains a cat or a dog.
The significant leap forward in AI today is due to technological improvements in the way we can train large neural networks, readjusting large numbers of parameters with each run thanks to the capabilities of large cloud computing infrastructures. For example, GPT-3 (the AI system that powers ChatGPT) is a large neural network with 175 billion parameters.
What does AI need to work?
AI needs three things to be successful.
First, it needs high-quality, unbiased, and massive data. Researchers who build neural networks use the large datasets resulting from the digitization of society.
Co-Pilot, to augment human programmers, derives its data from billions of lines of code shared on GitHub. ChatGPT and other major language models use the billions of websites and text documents stored online.
Text-image tools, such as Stable Diffusion, DALLE-2, and Midjourney, use image-text pairs from datasets such as LAION-5B. AI models will continue to evolve in sophistication and impact as we further digitize our lives and provide them with alternative data sources, such as simulated data or data from game settings like that Minecraft.
AI also needs IT infrastructure for effective training. As computers become more powerful, models that now require intensive effort and large-scale computation could, in the near future, be managed locally. Stable Diffusion, for example, can already be run on local computers rather than cloud environments.
The third need for AI is the improvement of models and algorithms. Data-driven systems continue to advance rapidly in domain after domain once considered the territory of human cognition.
However, as the world around us is constantly changing, AI systems must be constantly retrained using new data. Without this crucial step, AI systems will produce answers that are factually incorrect or do not take into account new information that has emerged since their formation.
Neural networks are not the only approach to AI. Another important camp in artificial intelligence research is symbolic AI – instead of digesting huge sets of data, it relies on rules and knowledge similar to the human process of forming internal symbolic representations of particular phenomena.
But the balance of power has tilted heavily towards data-driven approaches over the past decade, with the “founding fathers” of modern deep learning recently received the Turing Awardthe equivalent of the Nobel Prize in Computer Science.
Data, computation and algorithms form the foundation of the future of AI. All indicators show that rapid progress will be made in all three categories in the foreseeable future.
George Siemensco-director, professor, Center for Change and Complexity in Learning, University of South Australia
This article is republished from The conversation under Creative Commons license. Read it original article.