Meta develops an AI language bot that can use external software tools
reach outside the box — Meta develops an AI language bot that can use external software tools With Toolformer, an LLM can improve its abilities by calling APIs to external programs.
Benj Edwards – Feb 15, 2023 11:56 pm UTC Enlarge / An artist’s impression of a robot hand using a desktop calculator.Aurich Lawson | Getty Images reader comments 80 with Share this story Share on Facebook Share on Twitter Share on Reddit
Language models like ChatGPT have revolutionized the field of natural language processing, but they still struggle with some basic tasks such as arithmetic and fact-checking. Last Thursday, researchers from Meta revealed Toolformer, an AI language model that can teach itself to use external tools such as search engines, calculators, and calendars without sacrificing its core language modeling abilities. Further ReadingOpenAI invites everyone to test ChatGPT, a new AI-powered chatbotwith amusing results
The key to Toolformer is that it can use APIs (application programming interfaces), which are a set of protocols that allow different applications to communicate with one another, often in a seamless and automated manner. During training, researchers gave Toolformer a small set of human-written examples demonstrating how each API is used and then allowed it to annotate a large language modeling dataset with potential API calls. It did this in a “self-supervised” way, meaning that it could learn without needing explicit human guidance.
The model learned to predict each text-based API call as if they were any other form of text. When in operationgenerating text as the result of a human inputit can insert the calls when needed. Moreover, Toolformer can “decide” for itself which tool to use for the proper context and how to use it.
This API-calling ability enables Toolformer to use external software tools like search engines, calculators, language translators, and factual references. For example, large language models (LLM) are well-known for not being particularly good at arithmetic. Toolformer can work around that limitation by using a calculator program. Or if someone wanted an LLM-based assistant to add a date to their calendar, Toolformer could handle that task by using an API link to a calendar app. Advertisement An illustration provided by Meta researcher Timo Schick shows an example of Toolformer making an API call to the calendar app. Timo Schick An illustration provided by Meta researcher Timo Schick shows an example of Toolformer making an API call to the calculator app. Timo Schick An illustration provided by Meta researcher Timo Schick shows an example of Toolformer making an API call to an external factual reference. Timo Schick
Toolformer is based on a pre-trained GPT-J model with 6.7 billion parameters . Experiments conducted by the researchers on various tool-using tasks seem to demonstrate that Toolformer achieves far stronger performance than the much larger GPT-3 model, which contains 175 billion parameters. Further ReadingAI-powered Bing Chat loses its mind when fed Ars Technica article
This isn’t the first time researchers have attempted to make up for limitations in language models. In fact, the recent Bing Chat model making the news this week can perform web searches on its own when needed, and others have attempted integrations with browsers, calculators, and search engines. According to Meta’s researchers, most existing approaches to integrating tools into language models have relied on large amounts of human annotations or have been limited to specific task-specific settings. By contrast, Toolformer can learn to use a range of tools in a generalized way that does not require specialized training for specific tasks.
With techniques like those found in Toolformer, we’re looking at a potential future where LLMs augmented with the ability to use external apps will become far more versatile and reliable assistants (ostensibly). But the ability to perform API calls also might increase an LLM’s capability to cause harm to user data (in apps) or create trouble in the outside world (through a web browser or communications tools)abilities that they might accidentally invoke while providing an answer. reader comments 80 with Share this story Share on Facebook Share on Twitter Share on Reddit Benj Edwards Benj Edwards is an AI and Machine Learning Reporter for Ars Technica. For over 16 years, he has written about technology and tech history for sites such as The Atlantic, Fast Company, PCMag, PCWorld, Macworld, How-To Geek, and Wired. In 2005, he created Vintage Computing and Gaming. He also hosted The Culture of Tech podcast and contributes to Retronauts. Mastodon: benjedwards@mastodon.social Twitter @benjedwards Advertisement Channel Ars Technica ← Previous story Next story → Related Stories Today on Ars