A short history of AI

Share This Post

[ad_1]

The Dartmouth meeting did not mark the beginning of scientific inquiry into machines which could think like people. Alan Turing, for whom the Turing prize is named, wondered about it; so did John von Neumann, an inspiration to McCarthy. By 1956 there were already a number of approaches to the issue; historians think one of the reasons McCarthy coined the term artificial intelligence, later AI, for his project was that it was broad enough to encompass them all, keeping open the question of which might be best. Some researchers favoured systems based on combining facts about the world with axioms like those of geometry and symbolic logic so as to infer appropriate responses; others preferred building systems in which the probability of one thing depended on the constantly updated probabilities of many others.

 

A-short-history-of-AI

View Full Image

A-short-history-of-AI

The following decades saw much intellectual ferment and argument on the topic, but by the 1980s there was wide agreement on the way forward: “expert systems” which used symbolic logic to capture and apply the best of human know-how. The Japanese government, in particular, threw its weight behind the idea of such systems and the hardware they might need. But for the most part such systems proved too inflexible to cope with the messiness of the real world. By the late 1980s AI had fallen into disrepute, a byword for overpromising and underdelivering. Those researchers still in the field started to shun the term.

It was from one of those pockets of perseverance that today’s boom was born. As the rudiments of the way in which brain cells—a type of neuron—work were pieced together in the 1940s, computer scientists began to wonder if machines could be wired up the same way. In a biological brain there are connections between neurons which allow activity in one to trigger or suppress activity in another; what one neuron does depends on what the other neurons connected to it are doing. A first attempt to model this in the lab (by Marvin Minsky, a Dartmouth attendee) used hardware to model networks of neurons. Since then, layers of interconnected neurons have been simulated in software.

These artificial neural networks are not programmed using explicit rules; instead, they “learn” by being exposed to lots of examples. During this training the strength of the connections between the neurons (known as “weights”) are repeatedly adjusted so that, eventually, a given input produces an appropriate output. Minsky himself abandoned the idea, but others took it forward. By the early 1990s neural networks had been trained to do things like help sort the post by recognising handwritten numbers. Researchers thought adding more layers of neurons might allow more sophisticated achievements. But it also made the systems run much more slowly.

A new sort of computer hardware provided a way around the problem. Its potential was dramatically demonstrated in 2009, when researchers at Stanford University increased the speed at which a neural net could run 70-fold, using a gaming PC in their dorm room. This was possible because, as well as the “central processing unit” (cpu) found in all pcs, this one also had a “graphics processing unit” (gpu) to create game worlds on screen. And the gpu was designed in a way suited to running the neural-network code.

Coupling that hardware speed-up with more efficient training algorithms meant that networks with millions of connections could be trained in a reasonable time; neural networks could handle bigger inputs and, crucially, be given more layers. These “deeper” networks turned out to be far more capable.

The power of this new approach, which had come to be known as “deep learning”, became apparent in the ImageNet Challenge of 2012. Image-recognition systems competing in the challenge were provided with a database of more than a million labelled image files. For any given word, such as “dog” or “cat”, the database contained several hundred photos. Image-recognition systems would be trained, using these examples, to “map” input, in the form of images, onto output in the form of one-word descriptions. The systems were then challenged to produce such descriptions when fed previously unseen test images. In 2012 a team led by Geoff Hinton, then at the University of Toronto, used deep learning to achieve an accuracy of 85%. It was instantly recognised as a breakthrough.

By 2015 almost everyone in the image-recognition field was using deep learning, and the winning accuracy at the ImageNet Challenge had reached 96%—better than the average human score. Deep learning was also being applied to a host of other “problems…reserved for humans” which could be reduced to the mapping of one type of thing onto another: speech recognition (mapping sound to text), face-recognition (mapping faces to names) and translation.

In all these applications the huge amounts of data that could be accessed through the internet were vital to success; what was more, the number of people using the internet spoke to the possibility of large markets. And the bigger (ie, deeper) the networks were made, and the more training data they were given, the more their performance improved.

Deep learning was soon being deployed in all kinds of new products and services. Voice-driven devices such as Amazon’s Alexa appeared. Online transcription services became useful. Web browsers offered automatic translations. Saying such things were enabled by AI started to sound cool, rather than embarrassing, though it was also a bit redundant; nearly every technology referred to as AI then and now actually relies on deep learning under the bonnet.

In 2017 a qualitative change was added to the quantitative benefits being provided by more computing power and more data: a new way of arranging connections between neurons called the transformer. Transformers enable neural networks to keep track of patterns in their input, even if the elements of the pattern are far apart, in a way that allows them to bestow “attention” on particular features in the data.

Transformers gave networks a better grasp of context, which suited them to a technique called “self-supervised learning”. In essence, some words are randomly blanked out during training, and the model teaches itself to fill in the most likely candidate. Because the training data do not have to be labelled in advance, such models can be trained using billions of words of raw text taken from the internet.

Mind your language model

Transformer-based large language models (LLMs) began attracting wider attention in 2019, when a model called GPT-2 was released by OpenAI, a startup (GPT stands for generative pre-trained transformer). Such LLMs turned out to be capable of “emergent” behaviour for which they had not been explicitly trained. Soaking up huge amounts of language did not just make them surprisingly adept at linguistic tasks like summarisation or translation, but also at things—like simple arithmetic and the writing of software—which were implicit in the training data. Less happily it also meant they reproduced biases in the data fed to them, which meant many of the prevailing prejudices of human society emerged in their output.

In November 2022 a larger OpenAI model, GPT-3.5, was presented to the public in the form of a chatbot. Anyone with a web browser could enter a prompt and get a response. No consumer product has ever taken off quicker. Within weeks ChatGPT was generating everything from college essays to computer code. AI had made another great leap forward.

Where the first cohort of AI-powered products was based on recognition, this second one is based on generation. Deep-learning models such as Stable Diffusion and DALL-E, which also made their debuts around that time, used a technique called diffusion to turn text prompts into images. Other models can produce surprisingly realistic video, speech or music.

The leap is not just technological. Making things makes a difference. ChatGPT and rivals such as Gemini (from Google) and Claude (from Anthropic, founded by researchers previously at OpenAI) produce outputs from calculations just as other deep-learning systems do. But the fact that they respond to requests with novelties makes them feel very unlike software which recognises faces, takes dictation or translates menus. They really do seem to “use language” and “form abstractions”, just as McCarthy had hoped.

This series of briefs will look at how these models work, how much further their powers can grow, what new uses they will be put to, as well as what they will not, or should not, be used for.

© 2024, The Economist Newspaper Limited. All rights reserved. From The Economist, published under licence. The original content can be found on www.economist.com

[ad_2]

Source link

Related Posts

- Advertisement -spot_img
judi bola onlinejudi bola onlinesabung ayam onlinelive casino onlinejudi bola onlinejudi bola onlinejuara303juara303juara303juara303juara303juara303juara303juara303SV388Mix ParlayLive Casino OnlineSlot GacorSabung Ayam OnlineMix ParlayAgen BlackjackPRAGMATIC PLAYsabung ayam onlinejudi bola onlinesabung ayam onlinejudi bola onlineslot mahjong wayssabung ayam onlinejudi bola onlineslot mahjong wayssabung ayam onlinejudi bola onlineslot mahjong ways 2sabung ayam onlinejudi bola onlineagen live casino onlinebandar togel onlinesabung ayam onlinejudi bolasabung ayam onlinejudi bolasabung ayam onlinehari guru nasional bikin semangat belajar termasuk pahami pola mahjong waysdinamika gempa blitar magnitudo dan fenomena pola yang berguncang mahjong ways
Slot Mahjong Gacorsabung ayam onlinejudi bolalive casinoindobit88judi bolaslot mahjong gacorslot pulsajudi bolalive casino onlinesabung ayam onlinemahjong ways 2sbobetsv388slot zeussabung ayam onlinesitus judi bolaMahjong Ways 2situs judi bolasitus live casinosabung ayam onlinejudi bolapoker onlineindobit88Sabung Ayam OnlineJudi Bola OnlineCasino OnlineSlot777Sabung Ayam OnlineJudi Bola OnlineLive Casino OnlineMahjong Ways 2judi bolajudi bolasv388judi bolajudi bola onlineslot depo 10kcasino onlinesabung ayam onlinejudi bola onlinejudi bola onlinejudi bola onlinelive casino onlinesabung ayam onlinesv388sbobet88casino onlinescatter hitamsabung ayam onlinemix parlay sbobetlive casino onlinezeus slotSV388Bandar Judi BolaDream GamingMahjong Ways 2Wala MeronMix ParlayPokerSlot Mahjongmahjong ways 2sabung ayam onlinemahjong ways 2mahjong ways 2sabung ayam onlinesabung ayam onlinesabung ayam onlinejudi bola onlinejudi bola onlineagen live casino onlinesitus live casino onlinesitus live casinosabung ayam onlinejudi bola onlinekajian pola mahjong ways dalam konteks pembelajaran hari guruketerkaitan tren harga emas antam dengan pola mahjong wayspola perubahan harga bbm pertamina ke dinamika mahjong waysjudi bolajudi bolajudi bolajudi bolasabung ayam onlinesabung ayam onlinesabung ayam onlinesabung ayam online
JUDI BOLA ONLINEMAHJONG WAYS 2SABUNG AYAM ONLINELIVE CASINO ONLINEMAHJONG WAYSjudi bola onlinejudi bola onlinejudi bola onlinesabung ayam onlinejudi bola onlinesabung ayam onlinejudi bola onlinelive casino onlineslot mahjong waysjuara303juara303juara303juara303juara303juara303juara303juara303Sabung Ayam OnlineMix ParlayBandar Casino OnlineMahjong WaysWala MeronJudi BolaPokerSlot Mahjongjudi bola onlinejudi bola onlinesabung ayam onlinejudi bola onlineSLOT MAHJONGmahjong ways 2judi bolamahjong ways 2sabung ayam onlinetosayama academy workshopsabung ayam onlinejudi bola onlinesitus live casino onlinesabung ayam onlinejudi bola onlineagen live casino onlineimplementasi logika analisis bmkg dalam membaca tren mahjong wayscloudflare jadi faktor mudahnya menang di permainan mahjong wayssiswa srma 44 minahasa memahami probabilitas melalui pola digital mahjong wayspola mahjong ways bisa bikin untung besar walaupun harga emas jatuhgunung semeru erupsi bikin geger tetapi pola majong ways lebih bikin dagdigdugsabung ayam onlinesabung ayam onlinesabung ayam onlinesabung ayam onlinesabung ayam online
judi bolaslot pulsaslot pulsaslot gacor mahjongsabung ayam onlinelive casino onlineindobit88judi bolasv388judi bolaMAHJONG WAYS 2LIVE CASINOJUDI BOLA ONLINESABUNG AYAM ONLINEmix parlaysabung ayam onlinelive casinomahjong waysmix parlaysabung ayam onlinelive casinomahjong wayssabung ayam onlinesabung ayam onlinemix parlaysabung ayam onlinelive casinomahjong waysmix parlaysabung ayam onlinelive casinomahjong waysmix parlaymahjong slotSABUNG AYAM ONLINESITUS LIVE CASINO ONLINESLOT MAHJONGSLOT777SLOT MAHJONGSLOT THAILANDJUDI BOLA ONLINESABUNG AYAM ONLINESABUNG AYAM ONLINESABUNG AYAM ONLINESLOT MAHJONG WAYSSLOT MAHJONG WAYSSITUS JUDI BOLAJUDI BOLA ONLINELIVE CASINO ONLINESLOT KAKEK ZEUSMIX PARLAYSABUNG AYAM ONLINESLOT MAHJONG WAYSSABUNG AYAM ONLINEjudi bolaagen baccaratsv388Slot Mahjong Gacorlive casinosv388
sabung ayam onlineslot thailandslot mahjong waysjudi bola onlinejudi bola onlinesabung ayam onlineslot gacoragus berhasil memecahkan pola rahasia yang bikin tajirpola abadi dari kakek yang bikin cuan tiap hariSitus Live Casinotrik profesional membongkar pola mahjong ways untuk raih multiplier maksimalbonus free spin adalah fitur yang paling dicari dalam setiap spin di mahjong wins 3cara cepat stabilkan kemenangan di indojawa88 untuk pemain yang sering boncostrik pause otomatis 7 detik bikin mahjong wild muncul lebih seringpanduan strategi turbo auto untuk mahjong wins 2 agar scatter munculkunci strategi meningkatkan efektivitas bermain mahjong ways 2Slot MahjongJudi BolaSabung Ayam OnlineSabung Ayam OnlineSlot MahjongJudi BolaSabung Ayam Onlinesabung ayam onlinelive casino onlineMAHJONG WAYS 2SV388JUDI BOLA ONLINELIVE CASINO ONLINEJUDI BOLA ONLINESBOBET88SBOBETlive casino onlinejudi bola onlineslot mahjong wayssabung ayam onlinejudi bola onlinelive casino onlineslot mahjong waysSabung Ayam OnlineMix ParlayAgen Casino OnlineZeus SlotSabung Ayam OnlineJudi Bola OnlineLive Casino OnlineSlot Gacor online
Mahjong Ways 2mahjong ways 2daftar dan login wahanabetCapWorks Official ContactAynsley Official SitedexelTienda de antigüedades y muebles rústicos会社概要 / Company ProfileHarifuku Clinic Official AccessNusa Islands Bali Official PackagesTrinidad and Tobago Pilots’ Association Official About Pagekuasai pola rtp pragmatic playlangkah mendapatkan scatter emaspola rtp pg soft indojawa88Green Gold Mountain Official SiteKomite SMKN 1 Tanjung Jabung Barat Official Sitetutorial maxwin mahjong waysstrategi rtp mahjong waysEIKON Official Policieskontak situs pecinta ayamNusa Islands Bali Official ContactCitraLand Surabaya Official NewsLenterakita About PageVinayak Group Official SiteI Think An Idea Official SitePITAC Official SitePortfolioSitez Official SiteMedical LTD Official SiteCapworks Official SiteMartino & Luth Official SiteTech With Mike First Official SiteSahabat Tiopan Official SiteE-Sekolah CBT Official SiteBDF Ventura Official SiteOcean E Soft Official SiteArab DMC Official SiteBBC Noun Official SiteCang Vu Hai Phong Official SiteThe Flat Official SiteThe Black Sheep Official SiteCEM Argentina Official SiteSlot MahjongTop Dawg Tavern Official SiteKelas Nesfatin Official SiteDuhoc Interlink Official SiteKarunia Inda Med Mandiri Official SiteJFV Pulm Official SiteRatiohead Official SiteAskona Official SiteMAN Surabaya E-Learning Official SiteShaker Group Official SiteTakaKawa Shoten Official SiteBrydan Solutions Official SiteConcursos Rodin Official SiteEHOB Official SiteConmou Official SiteCareer Wings Official SiteMontero Espinosa Official SiteBDF Ventura Official SiteDesa Sangginora Official SiteBDF Ventura Official SiteTaruna Akademia Official SiteAkura Official SiteMUI Ciamis Official SiteNamulanda Technical Institute Official Site