laitimes

The secret of the AI circle: a big copy of the world's models

author:Wall Street Sights

Plagiarism has become an open secret in the AI world.

According to Monday's article by The Information, many startups' AI chatbots are likely to have been developed using data from OpenAI and other companies. These bots are comparable to GPT-4 in some tasks, but charge a fraction of the charge.

The startup did not disclose the technology used by OpenAI during its development. However, The Information reports that OpenAI CEO Sam Altman told startup founders last summer that it was acceptable for startups to use OpenAI's technology in this way.

While Altman's response has come as a relief to some startups, the practice has essentially hurt OpenAI's growth, and Altman could change his mind at any time.

The secret of the AI circle: a big copy of the world's models
The secret of the AI circle: a big copy of the world's models

In startups, plagiarism has become the norm

Startups copy OpenAI by opening up a membership to GPT-4 and then asking it a series of questions, such as "What's wrong with this line of code?" and they use these questions and answers to train their own competing models.

There are many startups that have adopted this strategy.

Daniel Han, co-founder of Unsloth AI, estimates that about half of his clients take data from GPT-4 or Anthropic's Claude model and use it to improve their own models. Many companies also get this kind of data from ShareGPT, a website where developers share answers generated using OpenAI models.

Models from smaller developers are often based on popular open-source models that are freely available from Meta Platforms or Mistral AI, but by incorporating answers from OpenAI models, the quality of the output of these models can be significantly improved. Han said that some developers are using a service called OpenPipe to automate this process.

"This is what will happen in a new ecosystem where there are no clear rules in place," said Matt Murphy, managing director of Menlo Ventures, which invested in OpenAI's rival Anthropic. Murphy says:

If everyone is using the same data, how can you be better than everyone else?

It's unclear to what extent OpenAI, Google, Anthropic, and other big developers will allow startup rivals to use their data to catch up.

Radical Ventures合伙人Rob Toews表示:

The quality and source of training data for AI models is becoming one of the most important hot issues. No one knows exactly how things are going to play out, but any AI startup that doesn't have a thorough and strategic consideration of [the data source] is falling behind.

Developers who have secretly relied on other AI services while developing their models may face an awkward position if they are exposed.

For example, Paris-based Mistral created its own AI using Meta's open-source AI model, Llama 2, but didn't disclose it until it was accidentally leaked, causing some developers to be unhappy. Mistral has raised hundreds of millions of dollars.

The secret of the AI circle: a big copy of the world's models

The same goes for large companies?

In fact, startups use OpenAI's data to train models in the same way that AI giants like OpenAI do.

OpenAI's chief technology officer, Mira Murati, showed hesitation and confusion last month in response to questions about whether the company was using data from Google's YouTube, as well as Meta Platforms' Facebook and Instagram, to train Sora, which generates AI videos.

It wouldn't be surprising if OpenAI actually used this data.

According to a recent report by the New York Times, OpenAI has created a speech recognition tool called Whisper for transcribing YouTube videos as a way to improve GPT-4. Previously, some media reported that OpenAI secretly used YouTube data to train its early AI models.

Just earlier this month, YouTube CEO Neal Mohan also said he disapproved of OpenAI's use of YouTube videos to develop a Wensheng video model like Sora.

This behavior also led to OpenAI incurring an infringement lawsuit. The New York Times sued OpenAI and its biggest supporter, Microsoft, in December, alleging that they illegally copied the newspaper's news articles while training the model. OpenAI's chatbot "can generate Times content word for word," the lawsuit alleges.

In response, OpenAI argued that it had worked to establish partnerships with news publishers and that its training practices fell within the scope of "fair use" permitted by the U.S. copyright doctrine.

Still, both OpenAI and Google have struck multimillion-dollar licensing deals with publishers like Axel Springer and larger deals with major sites like Reddit.

Even the tech giants can't resist the lure of shortcuts.

The Information reported that Google had transcribed YouTube videos, Meta hired contractors to summarize copyrighted books, and Adobe used Midjourney's AI to generate photos, all to train its own AI models. A Google engineer has resigned amid concerns about the company's use of OpenAI's ChatGPT data.

Sharon Zhou, CEO of startup Lamini, said the rapid pace of AI development and fierce competition have forced developers to turn to controversial sources of training data, such as copyrighted content or LLMs.

Zhou says:

In this space, investors need to see very fast progress.

⭐ Star Wall Street news, good content is not missed ⭐

This article does not constitute personal investment advice, does not represent the views of the platform, the market is risky, investment needs to be cautious, please make independent judgment and decision-making.

Read on