laitimes

开源模型即将追上闭源模型?Meta发布Llama 3叫板GPT-4

author:175500; yse
开源模型即将追上闭源模型?Meta发布Llama 3叫板GPT-4

On April 18th, the AI circle welcomed another blockbuster news, and Meta debuted with Llama 3, which is known as "the most powerful open-source model ever". Meta has open-sourced Llama 3 8B and 70B models of different sizes for external developers to use for free, and in the coming months, Meta will launch a series of new models with multi-modal, multi-language conversations, longer context windows, etc. Among them, the large version of Llama 3 will have more than 400 billion parameters and is expected to "compete" with Claude 3.

开源模型即将追上闭源模型?Meta发布Llama 3叫板GPT-4

At the same time, Meta CEO Zuckerberg announced that based on the latest Llama 3 model, the Meta AI assistant has now covered all applications such as Instagram, WhatsApp, Facebook, etc., and has opened a separate website, as well as an image generator that can generate pictures based on natural language prompt words.

开源模型即将追上闭源模型?Meta发布Llama 3叫板GPT-4

The emergence of Llama 3 is directly benchmarked against OpenAI's GPT-4, which is completely different from OpenAI, which is "not Open".

According to people familiar with the matter, researchers have not yet begun fine-tuning Llama 3 and have not yet decided whether Llama 3 will be a multimodal model. It is reported that the official version of Llama 3 will be officially launched in July this year. Yann LeCun, chief scientist of Meta AI and winner of the Turing Award, while "waving the flag" for the release of Llama 3, announced that more versions will be launched in the coming months, saying that Llama 3 8B and Llama 3 70B are currently the best performing open source models of the same volume. The llama 3 8B outperforms the llama 2 70B in some test sets.

开源模型即将追上闭源模型?Meta发布Llama 3叫板GPT-4

Even Musk appeared in the comment section, with a concise "Not bad" expressing his recognition and expectation for Llama 3.

开源模型即将追上闭源模型?Meta发布Llama 3叫板GPT-4

Jim Fan, a senior scientist at NVIDIA, believes that the launch of Llama 3 has departed from the progress of technology, and it is a symbol of the difference between the open source model and the top closed-source model. From the benchmark shared by Jim Fan, it can be seen that the strength of the Llama 3 400B is almost comparable to Claude's "super cup" and the new version of GPT-4 Turbo, which will be a "watershed", and it is believed that it will unleash huge research potential, promote the development of the entire ecosystem, and the open source community may be able to use GPT-4 level models.

开源模型即将追上闭源模型?Meta发布Llama 3叫板GPT-4

The day of the announcement coincided with the birthday of Stanford University professor and top AI expert Ng Enda, Ng said bluntly that the release of Llama 3 is the best gift he has ever received in his life, thank you Meta!

开源模型即将追上闭源模型?Meta发布Llama 3叫板GPT-4

Andrej Karpathy, one of the founding members of OpenAI and former AI director of Tesla, also praised Llama 3. As one of the pioneers in the field of large language models, Karpathy believes that the performance of Llama3 is close to that of GPT-4: Llama3 is a model released by Meta that looks very powerful. Stick to the basic principles, spend a lot of high-quality time working on reliable systems and data, and explore the limits of long-term training models. I'm also very excited about the 400B model, which could be the first open-source model at the GPT-4 level. I think a lot of people will ask for a longer context length.

I'd like to have models with smaller parameters than 8B, ideally around 0.1B to 1B, for educational work, (unit) testing, embedded applications, etc.

开源模型即将追上闭源模型?Meta发布Llama 3叫板GPT-4

According to Cameron R. Wolfe, director of AI at Rebuy and a Ph.D. in deep learning, Llama 3 proves that the key to training a good large language model is data quality. He analyzed Llama 3's data efforts in detail, including: 1) 15 trillion tokens of pre-trained data: 7 times more than Llama 2 and more than DBRX's 12 trillion;

2) More code data: The pre-training process contains more code data, which improves the inference ability of the model.

3) More efficient tokenizer: Having a larger vocabulary (128K tokens) improves the efficiency and performance of the model.

开源模型即将追上闭源模型?Meta发布Llama 3叫板GPT-4

After the release of Llama 3, Xiaozha told the media, "Our goal is not to compete with open source models, but to surpass everyone and build the most advanced artificial intelligence." In the future, the Meta team will publish a technical report on Llama 3, revealing more details about the model. The debate over open source versus closed source is far from over, and GPT-4.5/5, which is secretly poised to take off, may arrive this summer, and the battle for large models in the AI field is still going on.

来 源 | 华尔街见闻官方

OpenAI robot Figure 01's reaction speed is close to that of a human

AI再迎重大飞跃!"解码"一夜刷屏的Sora

Challenging GPT-4, Google CEO Pichai explains the most powerful model Gemini in detail

IBM's new AI chip, the world's strongest?Ministry of Industry and Information Technology: The scale of China's AI core industry reaches 500 billion yuanBeijing issued the "Implementation Plan for Artificial Intelligence Computing Power Vouchers (2023-2025)Gartner released the 2023 China Data Analysis and Artificial Intelligence Technology Maturity Curve Seven Departments: Interim Measures for the Management of Generative AI Services

开源模型即将追上闭源模型?Meta发布Llama 3叫板GPT-4

☞ Business Cooperation: ☏ Please call 010-82306118 / ✐ or to [email protected]

开源模型即将追上闭源模型?Meta发布Llama 3叫板GPT-4

Read on