Open source "imitation" ChatGPT, it works? UC Berkeley thesis, persuasion, or move forward?

7fd50ef431529e9368e8e9eedb95660c.png

Original: Tan Jing+ChatGPT

From "ancient" to the present, the world of AI is a world where open source leads development.

Although Stable Diffusion, as an open source image generation model, has brought image generation to a new level, the emergence of ChatGPT seems to have shaken some people's beliefs.

Because ChatGPT is closed source, and it is the most advanced.

Open source fans shouted to the sky:

When will the open source large model (LLM) catch up with ChatGPT?

At this time, a university research team with a big brain hole seriously answered related questions with an academic paper:

"Imitate the ChatGPT model, will it work?"

30eb41e1b0204afde6627244641522c8.png

On May 25, 2023, the research team of the University of California, Berkeley submitted an academic paper on the international open source paper website arxiv.

The title of the thesis is "The False Promise of Imitating Proprietary LLMs".

My translation is "Imitating Language Models, Who Is Responsible for Blank Checks".

The English word imitating in this article is translated as imitation , here is a small model to learn the output of ChatGPT. Cannot be translated as "imitation".

They start by asking:

The open source LLMs large model is currently not as capable as the closed source, what should I do?

Second, come up with action ideas:

In the latest study, the research team attempted to improve the capabilities of large open-source models by making them mimic closed-source language models (e.g., ChatGPT).

Then, describe the observation:

Preliminary results look good, and the ability to imitate large models is comparable to ChatGPT in some respects. However, after in-depth analysis, it is found that these "imitators" only imitate to the surface and lack core capabilities.

Fourth, the conclusions are as follows:

The most effective way is to face the challenge of ChatGPT and fundamentally improve the capabilities of the open source model, rather than simply imitating it.

History has repeatedly proved, and academic papers have proved that there is no shortcut to copy in scientific research, only down-to-earth and hard work.

The thesis is over.

f0a911b4d176c57a7925ac2bf35e2ddc.png

Teacher Tan, I adhere to the concept that "AI will only eliminate people who can't use AI", and announce that the academic advisor behind this article is ChatGPT.

I had a simple conversation with it, and it gave me great, paid AI academic support, and that is a thank you in my heart:

First, Ctrl+C .

Paste the abstract of the paper "The False Promise of Imitating Proprietary LLMs" with one click.

Second, let it translate .

"An emerging method for cheaply improving weaker language models is to fine-tune on the output of more powerful models. This approach aims to cheaply imitate the capabilities of proprietary models using weaker open-source models. In In this work, we provide a rigorous analysis of this approach.

"We first fine-tuned a series of language models imitating ChatGPT. These models have different base model sizes (1.5 billion--13 billion), data sources, and imitation data volumes (0.3M--150M tokens). Then, we use Models are evaluated by crowd raters and authoritative NLP benchmarks."

"Initially, we were surprised by the quality of the imitation model's output—they seemed to be better at following instructions, and crowd workers rated their output as competitive with ChatGPT. However, when performing more targeted automated evaluations, we It is found that the imitation model hardly narrows the gap between the basic language model and ChatGPT on tasks where the imitation data is not large.

"We demonstrate that these performance differences can be overlooked by human evaluators because imitation models are good at imitating ChatGPT's style, but not its factuality. Overall, we conclude that model imitation is a false promise : There is a significant capability gap between open-source and closed-source language models, which current methods can only bridge by using large amounts of imitation data or more capable underlying language models. Therefore, we believe that the highest leverage action for improving open-source models It is about facing the difficult challenge of developing a better underlying language model, rather than taking the shortcut of imitating a proprietary system."

8d97bbf7c9ee5f154a9da815e0bbc525.png

The joys and sorrows of academics and ordinary people are not the same. I just feel that it is long-winded. As a distinguished paying user, I let go of my reserve and say frankly:

"Can you make the above academic text more understandable, to the level that can be published as an official account?"

ChatGPT answered me:

In a new research approach, a way to improve the quality of language models with a smaller investment has been found. Simply put, it is to use a weaker language model, such as open source, and let it learn the output of a stronger language model, such as ChatGPT. It's like making an open source model a "shadow" of a proprietary model, mimicking its behavior.

We have conducted in-depth research on this method. We tried a series of models and made them mimic ChatGPT. These models come in a wide variety of sizes, sources and amounts of simulated data.

We then evaluated these models through crowdsourcing and NLP benchmarks.

At first, we were surprised by the results of the imitation model. They follow instructions better, like a ChatGPT clone.

However, when we used more rigorous testing methods, we found that these imitation models did not achieve the effect of ChatGPT on tasks where there is not enough imitation data.

We find that this gap is because the imitation model only imitates ChatGPT's appearance, but not its core. Like humans, we can imitate a person's movements and voice, but not his thoughts.

Therefore, we believe that relying solely on imitation is useless. There is a clear capability gap between open-source and proprietary language models, and we cannot just bridge this gap with massive imitation data or larger base models.

We believe that the best approach is to face the challenge and develop better base models, rather than taking shortcuts and only imitating proprietary systems.

7a8f52df8c80be1163852abc92c65a7d.png

It's time to let ChatGPT know what it means to be tireless.

So, I said:

Please shorten it further, shorten it to 3-4 sentences?

ChatGPT answers:

In our latest work, we try to improve the capabilities of open source language models by making them mimic proprietary models such as ChatGPT.

Preliminary results look good, and the imitation model is comparable to ChatGPT in some respects.

However, in-depth analysis found that these imitation models only imitated the surface and lacked core capabilities. Therefore, we believe that the most effective way is to face the challenge and fundamentally improve the capabilities of open source models, rather than simply imitating them.

c799771e4cc263699f7a564c2261b5eb.png

1e4187c697434c87aef421871b80cdd5.png

Artificial intelligence is cool, but I don't think I'm cool at all. If I write articles like this in the future, ask myself, where is my value?

Well, it's time to show the real technology (show my worth).

After chatting with two top AI scientists in Silicon Valley in the United States and a factory in China, I learned about the situation.

There are also large-scale model R&D teams in my country who use the method of imitation to create large-scale models, which is the "imitation" mentioned in this academic paper.

The specific method includes using the open-source GPT as the basic model, that is, using this as the basic framework, and then frantically asking ChatGPT questions, and then inputting the output data into the domestic large model.

On the difficult road of climbing "humanity's highest intelligent software product", will it work?

Large models are in the midst of a storm at the research front, with academic viewpoints battling fiercely.

The University of California, Berkeley is the forefront of artificial intelligence research. Two different research teams at the same university hold diametrically opposed views.

As indicated above, the team of the paper "The False Promise of Imitating Proprietary LLMs" does not endorse this approach.

Another team, the research team behind the well-known open source large model vicuna (Vicuna), agrees with this approach.

433ff6a7daebb4386666a04b755db71c.png

Although the "fire of war" of this academic debate has not yet ignited domestically, the differences between the two sides have become very clear.

"Imitation" is opportunistic?

No, this is an academic debate and one should not jump to conclusions.

It will take time to resolve doubts.

-Finish-

23cb83a35c275c7548fe35e254533174.jpeg

read more

AI framework series:

1. The group of people who engage in deep learning frameworks are either lunatics or liars (1)

2. The group of people who engage in AI frameworks 丨 Liaoyuanhuo, Jia Yangqing (2)

3. Those who engage in AI frameworks (3): the fanatical AlphaFold and the silent Chinese scientists

4. The group of people who engage in AI framework (4): the prequel of AI framework, the past of big data system

Note: (3) and (4) have not been published yet, and will meet you in the form of book publishing.

comic series

1.  Interpretation of the Silicon Valley Venture Capital A16Z "Top 50" data company list

2.  AI algorithm is a brother, isn't AI operation and maintenance a brother?

3.  How did the big data's social arrogance come about?

4.  AI for Science, is it "science or not"?

5.  If you want to help mathematicians, how old is AI? 

6.  The person who called Wang Xinling turned out to be the magical smart lake warehouse

7.  It turns out that the knowledge map is a cash cow for "finding relationships"?

8.  Why can graph computing be able to positively push the wool of the black industry?

9.  AutoML: Saving up money to buy a "Shan Xia Robot"?

10.  AutoML : Your favorite hot pot base is automatically purchased by robots

11. Reinforcement learning: Artificial intelligence plays chess, take a step, how many steps can you see?

12.  Time-series database: good risk, almost did not squeeze into the high-end industrial manufacturing

13.  Active learning: artificial intelligence was actually PUA?

14.  Cloud Computing Serverless: An arrow piercing the clouds, thousands of troops will meet each other

15.  Data center network : data arrives on the battlefield in 5 nanoseconds

16. Being late is not terrible, what is terrible is that no one else is late ,  the data center network "volume" AI:

17.  Is it joy or sorrow? AI actually helped us finish the Office work

AI large model and ChatGPT series:

18. ChatGPT fire, how to set up an AIGC company, and then make money?

19.  ChatGPT: Never bully liberal arts students

20.  How does ChatGPT learn by analogy? 

21.  Exclusive丨From the resignation of the great gods Alex Smola and Li Mu to the successful financing of AWS startups, look back at the evolution of the "underlying weapon" in the era of ChatGPT large-scale models

22.  Exclusive 丨 Former Meituan co-founder Wang Huiwen is "acquiring" the domestic AI framework OneFlow, and wants to add a new general from light years away

23.  Is it only a fictional story that the ChatGPT large model is used for criminal investigation?

24.  Game of Thrones of the Large Model "Economy on the Cloud"

25.  Deep chat丨Fourth Paradigm Chen Yuqiang: How to use AI large models to open up the trillion-scale traditional software market?

26. CloudWalk's large-scale model: what is the relationship between the large model and the AI ​​platform? Why build an industry model?

27. In-depth chat with He Xiaodong: A "departure" nine years ago: laying the foundation for multi-modality and competing for large-scale models


6fe215d2fe11e3d965e137f250c0b80c.jpeg

d41bc6dafbd9899d3fcd2c6e1475a08f.png

Finally, let me introduce myself as the editor-in-chief.

I'm Tan Jing, author of science and technology and popular science topics.

To discover stories in the times,

I am chasing the gods of technology and blocking technology companies.

Occasionally write novels and draw comics.

Life is short, don't take shortcuts.

Originality is not easy, thank you for forwarding

If you still want to read my articles, just pay attention to "Dear Data"

Guess you like

Origin blog.csdn.net/weixin_39640818/article/details/130939383