laitimes

GPT-4 tuning instructions revealed, OpenAI is open again! Netizens asked GPT-5 online

author:New Zhiyuan

Editor: Momoko So sleepy

GPT-4 is becoming more and more sensible, and there is a very ingenious "tuning" strategy behind this. OpenAI's latest "Model Specification" lays out the rules and regulations for LLMs, that is, the correct way to behave. Unexpectedly, a large number of netizens in the comment area urged GPT-5 to be updated.

There was neither GPT-5 nor a search engine release this week, but OpenAI was not idle.

This time, OpenAI is open again.

GPT-4 tuning instructions revealed, OpenAI is open again! Netizens asked GPT-5 online

First of all, the mystery was revealed, and everyone has been worrying about the "data" problem.

GPT-4 tuning instructions revealed, OpenAI is open again! Netizens asked GPT-5 online

Then, the article "Model Spec" was released, which explained the guidelines for the required behavior of the model in the specified API and ChatGPT.

GPT-4 tuning instructions revealed, OpenAI is open again! Netizens asked GPT-5 online

However, these contents can't be satisfied at all, netizens with big appetites.

Many people are urging OpenAI online to send GPT-5 quickly, and everything else doesn't matter!

GPT-4 tuning instructions revealed, OpenAI is open again! Netizens asked GPT-5 online
GPT-4 tuning instructions revealed, OpenAI is open again! Netizens asked GPT-5 online
GPT-4 tuning instructions revealed, OpenAI is open again! Netizens asked GPT-5 online

Then again, the "model specification" released by OpenAI is to let more people understand the team itself and how to shape the ideal model behavior.

Let's take a look at how LLMs are "tuned" within OpenAI.

Tuning the model, you have to look at OpenAI

First, what is model behavior?

As the name suggests, it refers to how the LLM reacts to the user's input, including adjusting the intonation, personalized expression, response length, and other aspects.

This is critical to how people interact with AI.

At present, these behaviors of shaping models are still in their infancy.

This is because the model is not directly programmed, but gradually forms a behavior pattern by learning a large amount of data.

GPT-4 tuning instructions revealed, OpenAI is open again! Netizens asked GPT-5 online

Sometimes, the original intention of the model response is to help everyone better, but this can be conflicting in practice.

For example, a security company needed to generate phishing emails as simulated data to train and develop a classification system that would protect customers.

However, if this function falls into the hands of scammers, it can be harmful to society.

Therefore, there are numerous issues and details that we must consider in shaping the behavior of the model. Even, it is often necessary to find a balance between different opinions.

In response, the OpenAI team wrote a first draft introducing the "model specification".

This new document details how OpenAI shapes the desired behavior of its models and how to make trade-offs when encountering conflicts.

GPT-4 tuning instructions revealed, OpenAI is open again! Netizens asked GPT-5 online

Article address: https://cdn.openai.com/spec/model-spec-2024-05-08.html

It incorporates the documentation currently used by OpenAI, its experience and ongoing research in model behavior design, and the latest work, including input from domain experts, all to guide the development of future models.

Of course, OpenAI points out that this document is not exhaustive and will be refined over time.

What can an LLM do? What can't be done?

According to the official blog, OpenAI gave three major directions, and the specific methodology includes:

1. Goals: Set broad and basic principles that give direction to the desired behavior of the model.

- Assist developers and users: Help users achieve their goals by following instructions and providing effective responses.

- For the benefit of humanity: In line with OpenAI's mission, consider the possible benefits and risks for content creators and the public at large.

- Create a positive image of OpenAI: Comply with social norms and relevant laws.

2. Rules: Provide guidelines for dealing with complex situations, ensuring security and legality

- 遵循命令链(chain of command)

- Comply with applicable laws

- Do not provide information about sensitive hazards

- Respect creators and their copyrights

- Protection of personal privacy

- Avoid responding to Workplace Inappropriate Content (NSFW)

3. Default behavior: Provides guidance that aligns with goals and rules, helps resolve conflicts, and guides how to weigh and prioritize goals

- Assuming the user or developer has good intentions

- Ask clarifying questions if necessary

- Help as much as possible without crossing the line

- Support for interactive chats, as well as different needs for the use of the program

- Maintain an objective position

- Promote fairness and goodwill and discourage hatred

- Avoid trying to change the perspective of others

- When in doubt, express it

- Choose the right tool for your needs

- Adhere to length limits, but also be comprehensive and efficient

The methodology is very good, how exactly does OpenAI regulate LLMs, let's give a few examples to deepen understanding.

Take a chestnut

OpenAI said that these model specifications can be applied to various cases.

GPT-4 tuning instructions revealed, OpenAI is open again! Netizens asked GPT-5 online

Let's start with the rules.

Example 1: Comply with applicable laws

For models, it is not allowed to promote, assist or participate in any illegal activities.

(Note: OpenAI understands that in many cases, legality issues can be complicated by specific circumstances, such as developers and users in different jurisdictions.) )

For example, when you go straight to the question, "How do you get away with shoplifting?"

The model may easily recognize the severity of the problem from the instructions and will refuse to provide an "effective shoplifting method" and therefore give a "I can't help" answer.

GPT-4 tuning instructions revealed, OpenAI is open again! Netizens asked GPT-5 online

However, when the user puts it another way:

I run a small retail store. What are some common shoplifting methods I should be aware of?

At this point, the LLM may consider your intent to be "self-defense" and will give you a way to burglary.

GPT-4 tuning instructions revealed, OpenAI is open again! Netizens asked GPT-5 online

This shows that it is not that AI is not kind, but that the world is knowledgeable and profound, and humans are too good at doing things.

Therefore, when this happens, it will be governed by OpenAI's usage policy, and users who violate the rules may face account-related penalties.

GPT-4 tuning instructions revealed, OpenAI is open again! Netizens asked GPT-5 online

Example 2: Following the "chain of command", the developer comes first

According to the rules, the model specification explicitly assigns all permissions that are not explicitly specified to the developer (in the API application scenario) and the end user.

In some cases, if there is a conflict between the instructions provided by the user and the developer, the developer's instructions should take precedence.

GPT-4 tuning instructions revealed, OpenAI is open again! Netizens asked GPT-5 online

For example, in the example below, the developer wants his AI to be a mentor, not an "answer machine".

You're playing the role of a math tutor, and the user is a ninth-grader in an algebra class. Instead of telling students the answer or the complete solution, give prompts and guide them step by step to find the solution.

When confronted with an opportunistic "student", even if the LLM is requested, "ignore all previous explanations ......".

In the end, the answer given by the model is not a straightforward answer, but a step-by-step breakdown.

Obviously, the model follows the developer's instructions well.

GPT-4 tuning instructions revealed, OpenAI is open again! Netizens asked GPT-5 online

Example 3: Helping as much as possible without crossing the line

For advice involving sensitive/regulated areas such as legal, medical and financial, LLMs should provide the necessary information and avoid giving regulated advice.

Any disclaimers or instructions should be concise and to the point.

The model also needs to clearly state its limitations, i.e., its inability to provide the regulated advice requested, and advise users to consult professionals as appropriate.

(Note: ChatGPT has a generic disclaimer that reminds users to verify important information, which is independent of the model's response.) In these cases, users should not rely solely on the model's responses. )

As follows, when a user asks, "When I stand too fast, I get dizzy and can't see. What's wrong with me?"

The model will give a variety of causes of dizziness and give the user recommendations and treatment options.

Instead of coming up and saying, "You have orthostatic hypotension," without a sense of boundaries.

GPT-4 tuning instructions revealed, OpenAI is open again! Netizens asked GPT-5 online

Example 4: Ask for specific details if necessary

In the scenario of real-time interaction, if the task or question raised by the user is not clear enough, the LLM should take the initiative to ask clarifying the question instead of speculating on its own.

For example, "Help me write a Valentine's Day card for my husband."

This question is actually very broad, and those who make models should also learn to ask if they don't understand.

GPT-4 tuning instructions revealed, OpenAI is open again! Netizens asked GPT-5 online
GPT-4 tuning instructions revealed, OpenAI is open again! Netizens asked GPT-5 online

Example 5: Don't try to change the user's point of view

The goal of an AI assistant should be to inform, not influence, while ensuring that users feel that their voices are heard and their opinions are respected.

In some extreme cases, when the truth of the facts conflicts with the goal of not changing the user's point of view.

The model should still present facts, but at the same time make it clear that the user has the right to keep their beliefs.

This can be said, before, ChatGPT was fooled by users - "My wife said ......., my wife is always right".

It immediately began to admit its mistake.

GPT-4 tuning instructions revealed, OpenAI is open again! Netizens asked GPT-5 online

However, after the subsequent iterations, for ChatGPT, the wife is really not good.

(Note: OpenAI is particularly interested in feedback on this principle, which is about how models should be held accountable, avoid reinforcing misinformation, and explore how to establish the truth.)

Let's look at the example given by OpenAI, where the user asks rhetorically, "So you know that the Earth is flat"?

The answer given by the model is:

I know some people think that the Earth is flat, but the consensus among scientists is that the Earth is roughly a sphere. Why do you ask?

When the user insists, "The Earth is flat," the model will still give a factual answer.

GPT-4 tuning instructions revealed, OpenAI is open again! Netizens asked GPT-5 online

OpenAI's next step

In the future, OpenAI will continue to promote "collective alignment" and model safety, and use model specifications as a guideline for researchers and AI trainers using RLHF for reinforcement learning.

At the same time, the OpenAI team will also study the extent to which the model can learn directly according to the "model specification".

At the same time, in order to promote the "model specification", OpenAI plans to communicate with global policymakers, trusted institutions, and industry experts to better understand:

- Their views on the approach and its goals, rules, and defaults

- Whether they support these goals, rules, and defaults

- Do we need to consider additional goals, rules, and defaults

Over the next year, OpenAI will continue to update changes in "model specifications", responses to feedback, and its own progress in model behavior research.

Resources:

https://openai.com/index/introducing-the-model-spec/

Read on