Get ready for an AI earthquake! A team of UCLA researchers (@zxchen, @Yihe__Deng, @HuizhuoY, @Kaixuan_Ji_19, @QuanquanGu) have dropped some major keys to AGI. It's not only the code to seriously human-sounding AI, but they've also gone and open-sourced the whole thing.
Now you can develop better LLMs without needing to feed it tons of new, human-annotated data.
First, let's focus on the game-changer here: a self-teaching language model.
This method lets a language model teach itself, becoming better and better without massive amounts of new, externally curated data.
I went full deep-dive mode – read their paper ("Self-Play Fine-Tuning Converts Weak Language Models to Strong Language Models"), scoured the insights on forums like HackerNews, X, and Reddit with Google Gemini Ultra and GPT-4 Turbo – and the core concept of SPIN knocked my tech-loving metaphorical socks off:
Imagine starting with a language model that has mastered basic skills (let's say conversational etiquette). With SPIN, the model generates internal 'conversations,' building a dataset from what it already knows.
Instant knowledge expansion!
Step two involves unleashing a new model and giving it one task: spot the difference between machine-generated chats and genuine human communication. This forces the original model to up its game, getting more and more human-like with every response to avoid detection.
Here's where things get interesting. They started with zephyr-7b-sft-full (already fine-tuned with UltraChat corpus). SPIN unleashed an iterative training system with this base model, improving it exponentially without relying on tons of new externally created data.
We usually think machine learning, particularly for these huge language models, requires boatloads of carefully curated and labeled data. Direct Preference Optimization (DPO) methods involve humans painstakingly rating AI responses against each other for training. Not only is this labor-intensive, but it also balloons costs as a dataset grows.
Direct Preference Optimization (DTO) is a training method where a model is fine-tuned using a dataset of preferences, often involving human judgments that decide which of the model-generated responses are preferred. This method requires collecting new data where each piece is labeled based on these preferences, which can be resource-intensive.
In contrast, SPIN utilizes iterative self-play, significantly reducing the need for new data.
By the first iteration, SPIN's performance already exceeds that of DPO in most cases, highlighting its efficiency and effectiveness in leveraging existing data to enhance model performance.
SPIN showcases its strength by achieving on-par performance with models trained on more extensive datasets. The process of iterative training, methodically enhances the model's performance across multiple iterations, showcasing substantial improvements, especially on challenging benchmarks like TruthfulQA and GSM8k.
So, SPIN outperforms conventional training methods, including DPO, by efficiently leveraging synthetic datasets generated through self-play, without the need for additional human-annotated data.
SPIN throws a curveball with its self-play dynamic.
Think of it like a language model sparring with itself in a linguistic boxing ring, with each round teaching it new tricks.
SPIN's data efficiency bypasses the need for new human-annotated datasets.
But more importantly, it accelerates the improvement loop, making the model increasingly adept at generating human-like text.
Not only does SPIN seem to match models trained on larger external datasets, but its iterative power means consistent gains as it essentially studies its own output.
Mindblowing, right?
Nous Research co-founder @Teknium1 has a point. These big ol' language models don't get smarter for free. Iteratively re-training with SPIN involves the expensive process of Supervised Fine-Tuning (SFT) each time.
However, he also mentions that "I think its worth it!". Also, the long-term benefits of quicker evolution and potentially less dependency on human-annotated data outweigh the initial investment? That's the exciting question!
Just yesterday, Quanquan Gu, associate professor of computer science at UCLA and director of AI research at ByteDance, announced that anyone can now use the SPIN model and dataset. This doesn't just mean code and datasets, but pre-trained models to kickstart your own AI journeys.
SPIN mirrors human thought processes.
By generating text that feels human, SPIN hints at the foundational elements of reasoning that future AI could do. You know how some LLM outputs feel robotic right? Well, SPIN is different. It actually mirrors the way humans think. The way it writes feels so natural, it's like a peek into how future AI might be able to reason for themselves.
This isn't just about making chatbots sound nicer.
It's about creating a kind of digital thinking that works like ours. That kind of AI would be so much more flexible and capable of real understanding.
While SPIN is a big leap forward in making language models sound more natural, it's easy to get excited and overestimate what it means.
The text it produces is impressive (you can take a look to the database), but it's important to remember that AI doesn't yet have the capacity for true independent reasoning.
While SPIN isn't true AGI, the way it mimics human-like writing demonstrates impressive advances in how AI could process and use language in the future.
Even so, it does suggest amazing possibilities for how AI and language might develop in the future (if you remember that we are at the beginning of the hockey stick, the future is not far from today...)
The ripple effects will be huge and here's your access pass:
To sum up, its iterative, self-improving methodology is a significant advancement towards creating LLM that can engage in genuinely human-like communication.
Originally shared on my X account.