laitimes

Within a few hours of its release, Microsoft deleted the GPT-4 open-source model in seconds!

author:New Zhiyuan

Editor: Editorial Department

Some time ago, Microsoft announced and open-sourced the latest generation of large model WizardLM-2, which claims to be comparable in performance to GPT-4. However, before it was launched for a day, the model weights and announcements were all deleted, and the reason was ......

Last week, Microsoft parachuted in WizardLM-2, an open-source model that can be called GPT-4.

Unexpectedly, a few hours after it was released, it was immediately deleted.

Some netizens suddenly found that WizardLM's model weights and announcement posts were all deleted, and they were no longer in Microsoft's collection.

Within a few hours of its release, Microsoft deleted the GPT-4 open-source model in seconds!

The GitHub project homepage has become 404.

Within a few hours of its release, Microsoft deleted the GPT-4 open-source model in seconds!

Project Address: https://wizardlm.github.io/

The weights of the model, including the HF, are also all gone.....

Within a few hours of its release, Microsoft deleted the GPT-4 open-source model in seconds!

全网满脸疑惑,WizardLM怎么没了?

Within a few hours of its release, Microsoft deleted the GPT-4 open-source model in seconds!

However, the reason why Microsoft did this was because the team forgot to "test" the model.

Within a few hours of its release, Microsoft deleted the GPT-4 open-source model in seconds!

Subsequently, the Microsoft team came forward to apologize and explain that it had been a while since WizardLM was released a few months ago, so we were not very familiar with the new release process now.

We accidentally missed one thing that was required in the model release process: poison testing
Within a few hours of its release, Microsoft deleted the GPT-4 open-source model in seconds!

Microsoft WizardLM is the second generation

In June last year, the original WizardLM, which was fine-tuned based on LlaMA, was released, attracting a lot of attention from the open source community.

Within a few hours of its release, Microsoft deleted the GPT-4 open-source model in seconds!

Address: https://arxiv.org/pdf/2304.12244.pdf

Subsequently, a code version of WizardCoder was born, a model based on Code Llama and fine-tuned using Evol-Instruct.

The test results showed that WizardCoder's pass@1 on HumanEval reached a staggering 73.2%, surpassing the original GPT-4.

Within a few hours of its release, Microsoft deleted the GPT-4 open-source model in seconds!

Fast forward to April 15th, and Microsoft developers officially announced the new generation of WizardLM, this time fine-tuned from Mixtral 8x22B.

It contains three parametric versions, which are 8x22B, 70B, and 7B.

Within a few hours of its release, Microsoft deleted the GPT-4 open-source model in seconds!

Most notably, in the MT-Bench benchmark, the new model achieved a leading edge.

Within a few hours of its release, Microsoft deleted the GPT-4 open-source model in seconds!

Specifically, the performance of the WizardLM 8x22B model in the maximum-parameter version is almost close to that of GPT-4 and Claude 3.

At the same parametric scale, the 70B version ranks first.

The 7B version is the fastest, even achieving performance comparable to the leading model with 10 times the parameter scale.

Within a few hours of its release, Microsoft deleted the GPT-4 open-source model in seconds!

The secret behind WizardLM 2's outstanding performance is the revolutionary training methodology Evol-Instruct developed by Microsoft.

Evol-Instruct leverages large language models to iteratively rewrite the initial instruction set into increasingly complex variants. These evolutionary instruction data are then used to fine-tune the base model to significantly improve its ability to handle complex tasks.

The other is RLEIF, a reinforcement learning framework that also played an important role in the development of WizardLM 2.

In WizardLM 2 training, the AI Align AI (AAA) approach is also employed, which allows multiple leading large models to guide and improve with each other.

The AAA framework consists of two main components, namely "co-teaching" and "self-learning".

During this phase of co-teaching, WizardLM and a variety of licensed, open-source, and proprietary advanced models conduct mock chats, quality judges, suggestions for improvements, and close skill gaps.

Within a few hours of its release, Microsoft deleted the GPT-4 open-source model in seconds!

By communicating with each other and providing feedback, models can learn from their peers and refine their capabilities.

For self-learning, WizardLM can generate new evolutionary training data for supervised learning and preference data for reinforcement learning through active self-learning.

This self-learning mechanism allows the model to continuously improve performance by learning from the data and feedback it generates on its own.

In addition, the WizardLM 2 model was trained using the generated synthetic data.

In the researchers' view, the training data for large models is drying up day by day, and it is believed that the carefully created data of the AI and the model supervised by the AI step by step will be the only way to a more powerful AI.

As a result, they created a synthetic training system that is entirely AI-driven to improve WizardLM-2.

Within a few hours of its release, Microsoft deleted the GPT-4 open-source model in seconds!

Netizens who are quick have already downloaded the weights

However, many people had already downloaded model weights before the library was deleted.

Several users also tested on some additional benchmarks before the model was removed.

Within a few hours of its release, Microsoft deleted the GPT-4 open-source model in seconds!

Fortunately, the netizens who tested it were impressed by the 7B model and said that it would be the preferred model for performing local assistant tasks by themselves.

Within a few hours of its release, Microsoft deleted the GPT-4 open-source model in seconds!

It was also poisoned and found to have a score of 98.33 for the WizardLM-8x22B, compared to 89.46 for the base Mixtral 8x22B and 92.93 for the Mixtral 8x7B-Indict.

The higher the score, the better, which means that the WizardLM-8x22B is still very strong.

Within a few hours of its release, Microsoft deleted the GPT-4 open-source model in seconds!

If there is no poisoning test, it is absolutely impossible to send the model out.

Large models are hallucination-prone and well-known.

If WizardLM 2 outputs "toxic, biased, or incorrect" content in its response, it is not friendly to large models.

In particular, these mistakes have attracted the attention of the whole network, and Microsoft itself will be subject to criticism, and even be investigated by the authorities.

Within a few hours of its release, Microsoft deleted the GPT-4 open-source model in seconds!

Some netizens wondered that you can update the indicator through the "poison test". Why would I want to delete the entire repository and weights?

According to the Microsoft authors, this is the only way to do this according to the latest internal regulations.

Within a few hours of its release, Microsoft deleted the GPT-4 open-source model in seconds!

Others say that we want a model that has not been lobotomy.

Within a few hours of its release, Microsoft deleted the GPT-4 open-source model in seconds!

However, developers will need to be patient, and the Microsoft team has promised that it will go live again after the testing is complete.

Read on