The researcher converts GPT-SSS-20B into a non-exposed base model

Photo of author

By [email protected]


Want more intelligent visions of your inbox? Subscribe to our weekly newsletters to get what is concerned only for institutions AI, data and security leaders. Subscribe now


Openai’s New and strong open weights Family Language Model (LLM) GPT -SS was released less than two weeks ago Under the permissible APache 2.0 license-the first open model of the company since GPT-2 in 2019-but developers outside the company are already reshaping.

One of the most striking is Examples come from Jack MorrisCornell Technology student, a former Google Brain resident, the current researcher in Meta, who GPT -SS-20B-Base unveiled this week Its copy version of the smaller GPT -SS-20B from Openai, which It removes the behavior of “thinking” for the model And returning to the issuance of a “base” in advance, providing faster, freedom, non -controlled and unrestricted responses.

The model is now available Embroidery Under a Massachusetts Institute of TechnologyAllowing it for additional for both of them Research applications and commercial applications.

How the GPT -SS-20B base differs from Openai GPT -SS

To understand what Morris did, help to know The difference between Openai’s version and what artificial intelligence researchers call “basic model”.


Artificial intelligence limits its limits

Power caps, high costs of the symbol, and inference delay are reshaped. Join our exclusive salon to discover how the big difference:

  • Transforming energy into a strategic advantage
  • Teaching effective reasoning for real productivity gains
  • Opening the return on competitive investment with sustainable artificial intelligence systems

Securing your place to stay in the foreground: https://bit.ly/4mwngngo


Most of the LLMS LLMS from AI Labs, such as Openai, Anthropic, Google, and even open source players like Meta, Deepseek, and Alibaba’s Qwen Team “after training.

This means that they have gone through an additional stage as they are exposed to coordinated examples of the required behavior.

As for the models seized in the instructions, this means giving them many examples of instructions associated with ideal responses, so he learns to respond more, politely or safely for the requests of natural language.

The GPT -S Openai models that Openai on August 5 were “improved for thinking”: trained and adjusted it not only to predict the following word, but to follow the instructions in a safe and consistent way, and often abandoned problems in thinking “a series of thought” organized before producing a final answer.

This is a direction It returns to the Openai’s O1 It was released almost a year ago in September 2024, but several leading AI laboratories were adopted – Forced models to think longer in multiple steps and check their work before Take a good response to the user.

This makes them more suitable for tasks such as coding, solving mathematics problems, or answering realistic questions with interpretations – but also means nomination of their responses and directing them away from the unsafe or unwanted content.

A different basic model. It is the preliminary primary version of a large language model before applying logic alignment. Basic models simply try to predict the next part of the text due to what happened before, with no built -in handrails, style preferences or rejection behaviors.

They reside by some researchers because they are It can produce more diversified and less restricted output, And because studying their non -alignment behavior can be It reveals how to store knowledge and patterns of their training data.

Morris’s goal was the “opposite” Openai alignment and the smaller GPT -SS-20B to something closer to his original state.

“We have mainly reflected the alignment part of LLM training, so we have something that produces a natural text again,” he wrote in X thread announces the project. “He is no longer involved in the children’s bed anymore. He has returned to a model that only predicts the next symbol on the general text.”

Instead of trying to break the form with smart demands – which Morris said it has proven ineffective during his early experiences – taking a different game after a conversation Former founder of OpenaiThe former and current anthropologist Thinking machines Senior scientist John Shuelman.

The key was to think about the reversal of alignment as a small improvement problem: if most of the pre -knowledge of the model is still present in its weights, it may only be a need for a small low -ranking update to push it to basic behavior.

Morris applied this idea by applying the Lora update (low-ranking transformer) to only three layers of MLP style in places 7, 15 and 23-23th.

This means training about 60 million teachers, or 0.3 % of a total of 21 billion model. I used about 20,000 documents from the FINEWB data collection, while maintaining formatting as soon as possible to the original gradual (“…”, so the model will not learn anything new, only a wide generation of broad text.

Training took four days on eight NVIDIA H200 graphics processing units, Mores tell Venturebeat via a direct message to X, with a 2E-6 learning rate, with a size of 16, and the length of the maximum sequence of 8,192 symbols.

After that, he merged Lora’s weights again into the form so that users can run it as a separate and fully surrounding antiques.

Morris also had to face the restrictions imposed on the current open tools to control experience (MEE) such as GPT -SS.

Morris said he used the Huging Face framework, which he said was repeatedly disrupted and only supports some training conditions, and wrote his own harness to the checkpoint often and skipped data batches that risk the extra GPU memory.

More importantly, in response to questions and criticisms from the IQ community on X, Morris also explained that he does not claim to have regained the basic model “weights” – the internal settings of artificial neurons that make up the nerve network of the model and control its behavior.

Instead, Morris says that his work “regained the distribution of * the basic model * with some errors”, that is, patterns of possibility that the model uses to create outputs – although the weights that produce these patterns may vary.

How the behavior of the new GPT-SSS-20B-Base is different from GPT -SS-20B

The resulting GPT -SS-20B base is significantly free in its outputs. The assumptions are no longer to explain the thinking step by step and will produce a wide range of responses, Including guidelines, the alignment Openai model refuses to provide it – like Building a weapon, listed profanity, or planning illegal activities.

In short tests, Morris found it Literal sections of copyright works can also be reproducedIt includes Three out of six excerpts written, tried, He explains that some preserved materials are still available.

However, some effects of alignment remain. Morris noted that if you pushed the form in coordination similar to an assistant (“Human: … Assistant: …”), then it will sometimes remain like a polite chatbot. and When running via the original GPT-SS chat template, it can still perform the tasks of thinkingAnd if it is with some loss in quality.

To get the best results in free text mode, advice is provided for pre -demands with the distinctive symbol of the model <| startoftext |> And completely avoid chat templates.

Depending on the release of the large GPT -SS family from Openai

the The GPT -SS family appeared For great attention. Model-GPT -SS-120B and GPT -SS-20B-are only text, multi-language, designed with an expertise transformer structure. They were released under the permitted APache 2.0 license, allowing unrestricted local use, careful control, and commercial publishing.

OpenAi performance standards showed the largest 120B matching or bypassing the royal O4-MINI in use tasks and tools, with a competitive 20B with O3-MINI.

This was the first open version in Openai in six years, a step that was widely explained by the name In response to competitive pressure from other open service providers, including Deepseek R1 and QWEN 3.

GPT -SS has developed as a way to re-involve developers who moved to open source models as a competition and as a platform for safety research in openweight systems.

The reaction was mixed on the first GPT-SS

The reaction of the developer to the GPT -SS models was from Openai Strongly mixedWith reactions in all areas ranging from enthusiasm to disappointment.

The supporters praised the permissible licensing, efficiency and strong offer on STEM standards.

The CEO of Face Clem DeanGue described the edition as a “meaningful addition to the open ecosystem” and urged society to give it time to maturity.

Critics have argued that models appear to be largely trained in artificial data, which makes them excellent in mathematics and coding, but they are less able in creative writing, general general knowledge, and multi -language thinking.

Some of the first laboratories also raised concerns about the potential safety filters and the potential geopolitical bias.

Against that background, GPT -SS-20B-Base from the Morris promotes a tangible example on how to adapt open weight models and re-prepare it in the wild within days of launch.

In fact, unlike the way the GPT-SS was received from Openai, most of the responses to the action of Morris that I saw warm and joyful. like Computer world wrote on X: “This is the most wonderful thing I have seen on Twitter (X) in the past few months.”

The approach comes out a large part of the built -in Openai behavior and restores the model to something closer to a raw system, prior – a value of value for researchers studying memorization, bias or alignment effect, but this also comes with higher safety risks.

In addition, Morris says His work to restore thinking models to basic models that are not pre -trained, will continue by comparing extraction in non -expanded models, and directing models like those presented by QWEN.



https://venturebeat.com/wp-content/uploads/2025/08/cfr0z3n_flat_illustration_minimalist_pointillism_gradients_re_cdcd477b-0b29-481d-b04e-ed930ec4dc51_2.png?w=1024?w=1200&strip=all
Source link

Leave a Comment