It's effectively a layer of (well needed) sugar on top of their existing fine-tuning mechanism.
The challenge with fine-tuning is collecting a representative dataset to tune against. The tooling they added makes it easy for you to persist your prompts and responses within the OpenAI platform, and then later select those persisted pairs (that were created with e.g. GPT-4o) and use them to fine-tune a cheaper model (like GPT-4o mini) - such that the more expensive model is effectively "teaching" the cheaper model what to do.
You could do this before, but it was a LOT of work. The new "distillation" features make it easier.
Am I missing something?
Remember though, OpenAI and pretty much all of the other providers have things in their terms that roughly say you're not allowed to use their models to train models from other providers - so fine-tuning a model against synthetic data created using OpenAI is likely a terms violation anyway.
Here's an (now slightly outdated) example of OpenAI GPT-4 => OpenAI GPT-3.5: https://datadreamer.dev/docs/latest/pages/get_started/quick_...
But you can also do GPT-4 to any model on HuggingFace. Or something like Llama-70B to Llama-1B.
For some tasks, this kind of distillation works extremely well given even a few hundred examples of the larger model performing the task.
I'm confused why you are mentioning 3.5 here. The weights aren't public, so you aren't actually running any derivative of GPT-3.5
Or am I mistaken. Can you clarify?
They don't limit it only to open-source models. And you can finetune 3.5 Turbo on OpenAI API.
Sure you could use a different providers, but you're going to be stuck with an incredibly fragmented ops stack. My experience with Google has been shockingly bad and Anthropic has got a good amount of catching up to do. No one else is remotely competitive. Honestly would love to see something from Meta long term.
This does raise the bar for any future startups though. If your plan was to distill GPT4 outputs and lease the weights to me through a REST API, I probably won’t be interested.
Currently, OAI does all the following:
- offers flagship models
- offers lite models
- offers easy finetuning of their models
- offers structured output and guaranteed JSON output.
- offers parallel tool/function calling which remains unmatched.
- has low API costs.
- offers a nice UI for their models
- offers Mac, iOS, Android, Windows app clients.
- offers image generation capabilities INTEGRATED with their language models.
- offers two-tier subscription plans for ordinary/pro (team) users.
- offers custom GPTs which can be used by ordinary people to create GPT experiences tailored to specific tasks (no need to build a website on your own).
- allows users to easily share chats!! (it took Anthropic a long time to have this feature, and even now it's not as good as OpenAI's solution).
- offers prompt caching and task scheduling to further save costs.
- offers unrivaled voice-to-text models at different sizes (Whisper).
- offers text-to-voice models that feel much more natural than the competition.
- has outstanding documentation.
- sets the standard for API (all other companies have to follow their conventions, such as `messages`, `.choices[0].message.content`, etc.)
- has the most capable team to, idk, build AGI/ASI...
It may be (close to) the best, but Whisper is nowhere near good enough.
So, if you're got a great idea for how to massively improve on that (I don't), there's a business opportunity there.