14
Hot take: I was sure fine-tuning a model from scratch was the only way to get good results for my project.
For a small text sorting tool I'm making, I spent two weeks trying to fine-tune a base GPT model with my own data. It was slow and the results were messy. Then a friend said to just use prompt engineering with GPT-4's API instead. I switched last Friday, wrote a better system prompt with clear examples, and got it working perfectly in about three hours. The API call was way simpler and cheaper. Has anyone else moved from custom training to better prompting and been shocked by the difference?
3 comments
Log in to join the discussion
Log In3 Comments
torres.pat22d ago
Wow, but doesn't that just depend on what you're trying to do?
6
the_jesse22d ago
Yeah, "fine-tuning from scratch" is almost never the right first step.
5
matthewh2818h ago
Seriously, why would you start from zero? I tried that once for a customer service bot. Spent weeks and a ton of money on compute just to get it to where a good base model already was. Starting with a strong pre-trained model and then tweaking it is the only way that makes sense.
6