Copied


Llama-3 Fine-Tuning Achieves 90% of GPT-4's Performance at Lower Cost

Luisa Crawford   Jul 14, 2024 02:46 0 Min Read


The success of Llama-3 has been remarkable, showcasing that open-source models are closing the gap with their closed-source counterparts, according to together.ai. By leveraging proprietary data, customers have been able to fine-tune smaller open-source software (OSS) models like Llama-3 to achieve higher accuracy than top-tier closed-source models.

Fine-Tuning Process

Together AI's platform allows users to fine-tune Llama-3-8B on proprietary data, creating custom models that outperform larger OSS alternatives like Llama-3-70B and are comparable to leading closed-source models like GPT-4, all at a fraction of the cost. A detailed guide demonstrates how a fine-tuned Llama-3 8B model improved from 47% accuracy to 65%, surpassing Llama-3-70B's 64% and nearing GPT-4's 71% accuracy.

The fine-tuning process involves several steps, including dataset transformation, uploading and verifying datasets, starting a fine-tuning job, and running evaluations to compare the results. The initial step requires downloading the Math Instruct dataset from HuggingFace, cleaning it up, and transforming it into a JSONL file format suitable for Together's platform.

Dataset Transformation

The transformation process involves loading the original JSON data, defining the Llama-3 prompt format, and converting the data into the correct format. This formatted dataset is then validated using Together's SDK before being uploaded for fine-tuning.

Uploading and Fine-Tuning

Once the dataset is prepared, it is uploaded to Together AI via the Python SDK. The fine-tuning job is then created using the Llama-3-8B base model, specifying the dataset, number of epochs, and other parameters. Users can monitor the fine-tuning job through Together AI's dashboard.

Evaluation and Results

After fine-tuning, the model's performance is evaluated using 1000 math problems. The fine-tuned Llama-3-8B model's accuracy is compared to the base Llama-3-8B, Llama-3-70B, and GPT-4. The fine-tuned model achieved a 65.2% accuracy, outperforming the base model's 47.2% and Llama-3-70B's 64.2%, and coming close to GPT-4's 71.4% accuracy.

The results indicate that the fine-tuned Llama-3-8B model outperformed the base model by nearly 20%, surpassed the top OSS model Llama-3-70B, and achieved over 90% of GPT-4's accuracy. Additionally, the fine-tuned model is faster, 50 times cheaper than GPT-4, and offers full ownership of the model and weights.

Conclusion

This fine-tuning approach demonstrates that small open-source models like Llama-3-8B can be customized to perform specific tasks with high accuracy, speed, and cost-efficiency. Users can leverage their proprietary data to fine-tune a model and either host it on Together AI or run it independently, maintaining full control and ownership.

The Llama-3-8B model trained on math problems outperformed leading OSS models and approached GPT-4's performance, with a total fine-tuning cost of less than $100 on Together AI.


Read More
Bitcoin (BTC) has held the top spot in the cryptocurrency world since its creation in 2009. It remains the largest and most recognized digital asset by market capitalization.
Institutional interest in crypto surges; regulatory clarity and tokenization reshape the landscape.
AI and blockchain converge, enabling decentralized data ownership and real-time integration for better predictions.
Crypto for Everyone: Crypto must focus on real-world utility and user experience to gain mainstream acceptance and rebuild trust.
Online casinos have experienced rapid growth during the last decade as they have had to overcome security issues all while working to establish transparency.
Blockchain technology transformed digital transactions, with crypto apps playing a crucial role in this transformation.
Grayscale is expanding its ETF lineup with two new Bitcoin income funds designed to generate monthly payouts.