Microsoft has announced that it is investigating how human-robot interactions occur, using ChatGPT artificial intelligence (AI) technology on these machines.
Powered by OpenAI, ChatGPT is a ‘chatbot’ trained to carry on a text conversation that is based on the GPT 3.5 language model and has the ability to generate and link ideas, as well as to remember conversations.
In the last month, the company has been integrating this technology into its different services, starting with the Azure OpenAI Service in mid-January, to later take it to your Bing search engine and your Edge browser.
Its latest iteration has materialized in the preview of the new AI-powered Bing in mobile apps and its Edge browser, as well as the search engine version for Skype.
Now him Technological manufacturer has announced that it is testing this technology in robotsto “make natural interactions possible” with humans and has explained what he expects from this integration.
“Our goal with this research is to see if ChatGPT can think beyond the text and reason about the physical world to help with robotics tasks,” Microsoft said in a statement, where it has qualified that it seeks “help people interact with robots more easilywithout the need to learn complex programming languages or details about robotic systems.”
In this sense, he has established as a challenge to teach this ‘chatbot’ to solve problems taking into account factors such as the context of the operating environment and the way in which “physical actions can change the state of the world”.
The company has recognized that this OpenAI product has capabilities to perform tasks on its own, but that has developed a set of design principles that can be used to guide language models towards solving robotics tasks.
On the other hand, he has indicated that ChatGPT can “unlock a new paradigm of robotics”, in which slow, expensive and inefficient processes of rewriting the code of the robot’s behavior usually follow one another.
In this way, this ‘chatbot’ has the ability to provide high-level feedback to the so-called large language model (LLM), while monitoring the performance of the robot and generating code to different robotics scenarios.
To boost the LLM, Microsoft has built a series of design principles intended to write directions for robotics tasks.. To do this, it has defined a set of robotic application programming interfaces (APIs).
Once the text message has been written for ChatGPT to interpret, in which the objective of the task that the machine must perform is described, it is explicitly indicated what it must do, what it must not do or how the ‘chatbot’ must formulate your answers.
He developer or engineer awaits the response or output of the ChatGPT code using a simulator and, if necessary, uses natural language to assess the AI on the quality and security of its responses.
Once the model has responded in the desired way and the solution is viable, the final code of the artificial intelligence chat is implemented in the robot.
ChatGPT cases in robots
The technology company has shared a series of experiments to show what are the actions that robots can respond to using the model developed by OpenAI.
First, it taught him the steps to follow to control a real drone, and he was able to write complex code structures, such as a zigzag pattern to visualize the shelves of a bookcase.
ChatGPT was also used with the Microsoft AirSim simulator, a scenario in which the model “was able to analyze user intent” and the detailed signals to control a drone with precision.
This language was even tested in a manipulation scenario with a robotic arm. After analyzing the guidelines, ChatGPT was able to link a series of skills learned to, among other activities, stack wooden blocks.
The developers also chose to give these robots objectives such as finding objects, using detection functions. He also carried out additional experiments to evaluate if the ‘chatbot’ can decide where to direct the machine based on what it should look for.
*With information from Europa Press