Less Data, Same LLM Performance? UGA Says Yes
A paper from UGA shows you can match SOTA LLM performance with 2000 targeted samples instead of 300000. The secret? Looking inside model’s feature space. A recent paper called Less is Enough from University of Georgia, UC San Diego matched SOTA performance on instruction-following using 2,000 synthetic samples instead of 300,000. That’s 150x less data … Read more