Use ChatGPT to control NPC, the action is as realistic as a normal person! Stanford Google's new research bombing field, endowing large models with accurate memory...

Posted by Xiao Xiao from Aufeisi Qubit
| Public Account QbitAI

Letting 25 AIs live together, they simulated human actions!

0723d27e0b0273f3632227ce301ad2bd.gif

Not only do they get up, brush their teeth, eat, and sleep like people, but they are also forced to "go out to work" and even celebrate festivals together in their spare time.

All this, from a new study from Stanford and Google -

They did a very new experiment, created 25 AI NPCs, each with different identities and action decisions, and let them live together in a sandbox environment, relying on the ChatGPT large model to complete action decisions .

6eb5d4435c96697ebcfe83c9fd8dbcef.png

It turns out that these AI NPCs will not only live like people, such as writers to write, shopkeepers to run stores, but they will also interact with each other and even have dialogues:

345c424ba656e4004103ffc19c6aaf7c.png

As soon as the research was released, it exploded. Ethan Mollick, a professor at the Wharton School of Business at Penn University, said:

These actions are more like actions that normal humans would make than "role-playing".

d3ece1b665f3b8ecdc8e2f5475cdf25c.png

OpenAI scientist Andrej Karpathy even said that "this dwarfs open world games":

5adc5cff163f997d8305fc7aa3675215.png

So, what kind of "AI society" does this research construct?

When AI Plays The Sims

Judging from the demo display, these 25 AIs have different identities, personalities and ages:

c030e2768c9861ef42e2d9aa81f5961f.png

They will live in a communal community, which is very detailed, including residential areas, universities, parks, cafes, etc.:

0df5952911bae183389f827a6efb5358.png

It can be seen that each AI has its own basic itinerary, including waking up, washing, eating, sleeping, etc. According to different identity settings, it will take different actions to jointly maintain the operation of the community.

Among them, the emoji in the demo simply represent the ongoing operations of AI, such as sending and receiving letters and eating:

58b8a81019c656fb5746b486eb19307b.png

But among them, there will also be interactive behaviors:

fc47bb4ce7d0cf2e2be32cc0da376ac8.png

It is not the language of AI, but the way humans speak:

da684adec2628a7320990427b8587aa1.png

Not only that, these AIs that can interact with each other can even listen to the requirements of human players and take corresponding actions.

For example, the authors of the paper mentioned that they told one of the AI ​​NPCs as a human being that they hoped to hold a "Valentine's Day Party" event in the community.

Unexpectedly, this AI NPC not only enthusiastically agreed to the requests of the authors, but spontaneously began to organize this event, including the time, place and people present:

26e8502b71fe6799dbc5b5fd0433497c.png

Seeing this, some netizens have already expressed "too crazy"——

For players, rather than follow the steps, they may really need to persuade NPCs or think and act in order to complete the task in the game.

573a687e7f50e9554a1d1a5c4b32e10f.png

There are also netizens who have opened their minds and thought of a more "evil" gameplay:

Imagine you killed an NPC, then drove back to the city and watched them hold his funeral...

9a08d921a7518c77ea8634168953c7b1.png

So, how do these AIs act spontaneously and build a self-running AI interactive society?

Control Action Decisions with Large Models

Specifically, Stanford and Google have AIs take action and interact with other AIs in two steps .

The first step is to construct a variety of AI "personal settings" and let them take daily actions that conform to the settings.

Here we take John Lin, one of the AI ​​characters, as an example. You can find his setting list in the Demo.

It can be seen that his settings are very detailed, including basic information (name, age, location, time), attention settings (field of vision, etc.), personal circumstances (personality, family background, work, etc.).

739765fd521b78249ce0baf142e6a094.png

Using this as a reference, the AI ​​named John Lin can form a fixed daily living habit, including getting up early to wash up routinely, then going to work, and finally going home to sleep.

abb531e6aa97ddd917f4792df59882a1.png

But in the process of his life, he will definitely communicate with other AIs or meet new people, so——

The second step is to let the AI ​​character respond to the external environment and the actions of other people to generate interactions .

Thus, this research builds an architecture that empowers AI with memory, retrieval memory, perception, feedback, and interaction.

e200d9ed79a1911b8a8e8790df536093.png

Judging from the architecture diagram, AI will first perceive the surrounding environment, record events that occur within a certain field of vision (attention), and add them to its own memory stream (memory stream);

Subsequently, whether it is what the AI ​​plans to do (plan) or give feedback (reflect) on what happened outside, it will be affected by this memory flow. The specific way is to retrieve the corresponding events from the memory and rely on large Model ChatGPT to make decisions ;

(Of course, whether it is planning or feedback, the corresponding event will also be added to the memory flow and become a new memory)

Finally, the results of these decisions will be translated into concrete solutions, prompting AI to think and take concrete actions.

The study mentioned that the reason why the large model + memory architecture is adopted is that although the large model already has very good action logic, whether it is GPT-4 or other large models, long-term planning and maintaining a clear memory are still a problem. kind of challenge.

Based on this, a memory system is needed as a "reference" for the large model, so that it has a clear list of events that can be traced back when making decisions.

efbe5327cdeae3fe28d9a8774aad2776.png

Still, the study raises some questions about AI safety .

Including how to prevent its impact on society, and at the same time being used by people with intentions to generate Deepfake and "customized version" AI fraud, etc.——

For example, the AI ​​tea seller who chats with you in the future may not only remember everything you said, every birthday, but can even use this as a reference to generate more "intimate" answers, so that "her" looks more like a person...

291657d2217a0e80d3cfd7d371d3c50a.png

One More Thing

It is worth mentioning that a domestic company called Hyperparameter Technology has already made a similar AI NPC system GAEA .

Like the Stanford Google study, GAEA also has a memory system and relies on a large language model to help NPCs make action decisions.

8a339ecdfcfb68e025e03c6eff052a98.png

However, there are some differences in the incentive mechanisms that drive AI to act. Compared to this study relying on large models to complete action decisions:

b27f9857af9d4c27094ed31cf41c4048.png

The AI ​​in GAEA will also rely on more specific values ​​to further affect the NPC's action decisions (hunger value, mood value, etc.):

9971f094164eda709ed3f1094c12191d.png

For this kind of AI NPC technology, what brain-opening gameplay do you think of?

Demo address:
https://reverie.herokuapp.com/arXiv_Demo/#

Paper address:
https://arxiv.org/abs/2304.03442

Reference link:
[1]https://twitter.com/emollick/status/1645499660402925576
[2]https://twitter.com/karpathy/status/1645485475996790784

Guess you like

Origin blog.csdn.net/QbitAI/article/details/130096800
Recommended