"The AI revolution is about to spread way beyond chatbots.
From new plastic-eating bacteria and cancer cures to autonomous helper robots and self-driving cars, the generative-AI technology that gained prominence as the engine of ChatGPT is poised to change our lives in ways that make talking bots look like mere distractions.
While we tend to equate the current artificial-intelligence boom with computers that can write, talk, code and make pictures, most of those forms of expression are built on an underlying technology called a "transformer" that has far broader applications.
First announced in a 2017 paper from Google researchers, transformers are a kind of AI algorithm that lets computers understand the underlying structure of any heap of data -- be it words, driving data, or the amino acids in a protein -- so that it can generate its own similar output.
The transformer paved the way for OpenAI to launch ChatGPT two years ago, and a range of companies are now working on how to use the innovation in new ways, from Waymo and its robot taxis to a biology startup called EvolutionaryScale, whose AI systems are designing new protein molecules.
The applications of this breakthrough are so broad that in the seven years since the Google research was published, it has been cited in other scientific papers more than 140,000 times.
Modern AI has long been good at recognizing patterns in information. But previous approaches put serious limits on what more it could do. With language, for example, most AI systems could only process words one at a time, and evaluate them only in the sequence they were read, which limited their ability to understand what those words meant.
The Google researchers who wrote that seminal 2017 paper were focused on the process of translating languages. They realized that an AI system that could digest all the words in a piece of writing, and put more weight on the meanings of some words than others -- in other words, read in context -- could make much better translations.
For example, in the sentence "I arrived at the bank after crossing the river," a transformer-based AI that knows the sentence ends in "river" instead of "road" can translate "bank" as a stretch of land, not a place to put your money.
That level of contextual understanding enables transformer-based AI systems to not only recognize patterns, but predict what could plausibly come next -- and thus generate their own new information. And that ability can extend to data other than words.
"In a sense, the models are discovering the latent structure of the data," says Alexander Rives, chief scientist of EvolutionaryScale, which he co-founded last year after working on AI for Meta Platforms, the parent company of Facebook.
EvolutionaryScale is training its AI on the published sequences of every protein the company's researchers can get their hands on. Using that data, and with no assistance from human engineers, his AI is able to determine the relationship between a given sequence of molecular building blocks, and how the protein that it creates functions in the world.
Earlier research related to this topic, which was more focused on the structure of proteins rather than their function, is the reason that Google AI chief Demis Hassabis shared the 2024 Nobel Prize in chemistry. The system he and his team developed, called AlphaFold, is also based on transformers.
Already, EvolutionaryScale has created one proof-of-concept molecule. It's a protein that functions like the one that makes some jellyfish light up, but its AI-invented sequence is radically different than anything nature has yet to invent.
The company's eventual goal is to enable all sorts of companies -- from pharmaceutical makers producing new drugs to synthetic chemistry companies working on new enzymes -- to come up with substances that would be impossible without their technology. That could include bacteria equipped with novel enzymes that could digest plastic, or new drugs tailored to individuals' particular cancers.
Meanwhile, Karol Hausman's goal is to create a universal AI that can power any robot. Hausman's San Francisco-based startup, Physical Intelligence, is less than a year old, and Hausman himself used to work at Google's AI wing, DeepMind. His company starts with a variant of the same large language model you use when you access ChatGPT. The newest of these language models also incorporate and can work with images. They are key to how Hausman's robots operate.
In a recent demonstration, a Physical Intelligence-powered pair of robot arms does what is, believe it or not, one of the hardest tasks in all of robotics: folding laundry. Clothes can take on any shape, and require surprising flexibility and dexterity to handle, so roboticists can't script the sequence of actions that will tell a robot exactly how to move its limbs to retrieve and fold laundry.
Physical Intelligence's system can remove clothes from a dryer and neatly fold them using a system that learned how to do this task on its own, with no input from humans other than a mountain of data for it to digest. That demonstration, and others like it, was impressive enough that earlier this month the company raised $400 million from investors including Jeff Bezos and OpenAI.
In October, researchers at the Massachusetts Institute of Technology announced they're pursuing a similar transformer-based strategy to create robot brains that can take in vast amounts of data from a variety of sources, and then operate flexibly in a wide range of environments. In one instance, they made several films of a regular robotic arm putting dog food into a bowl, then used the videos to train a separate AI-powered robot to do the same.
As in robotics, researchers and companies working on self-driving cars are figuring out how to use transformer-based "visual language models" that can take in and connect not just language but images too. California-based Nuro and London-based Wayve, as well as Waymo, owned by Google's parent company, are among the companies working with these models.
This is a departure from pre-transformer approaches to self-driving, which used a mix of human-written instructions and older types of AI to process sensor data to identify objects on the road. The new transformer-based models are essentially a shortcut to giving self-driving systems the kind of general knowledge about the world that was previously very difficult to grant them.
Powerful as they can be, these systems still have limits and unpredictability that mean they won't be able to completely automate people's jobs, says Dettmer.
The AI at the heart of EvolutionaryScale, for example, can suggest new molecules for humans to try in the lab, but humans still have to synthesize and test them.
And transformer-based models are far from reliable enough to take over driving completely.
Physical Intelligence's system that taught itself to fold laundry would have to relearn that process in a way that's specific to your home before it can take over the job from you. That would require a huge amount of engineers' time, as well as money to train the model.
"I want to make sure I set expectations," says Hausman, the CEO. "As proud as we are of our accomplishment, we are still at the beginning."" [1]
1. EXCHANGE --- Keywords: A Powerful AI Breakthrough Is About to Transform the World --- The technology driving ChatGPT is capable of so much more. Mims, Christopher. Wall Street Journal, Eastern edition; New York, N.Y.. 16 Nov 2024: B.2.
Komentarų nėra:
Rašyti komentarą