Skip to content

Commit a74ac09

Browse files
committed
Changed README
2 parents 47b9b46 + 3bd68be commit a74ac09

File tree

12 files changed

+4743
-1570
lines changed

12 files changed

+4743
-1570
lines changed

README.md

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -57,6 +57,8 @@ We are looking for contributors! Please send an email to [[email protected]
5757

5858
Simplifine is licensed under the GNU General Public License Version 3. See the LICENSE file for more details.
5959

60+
## 📚 Documentation
61+
MAJOR OVERHAUL OF DOCUMENTATION IN THE WORKS (Done by 11th Aug 2024). In the meantime, please use this [notebook](https://github.com/simplifine-llm/Simplifine/blob/main/examples/cloud_quickstart.ipynb) here to learn how to use the model.
6062

6163
## 💬 Support
6264

README.md.orig

Lines changed: 135 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,135 @@
1+
# 🌟 Simplifine 🌟
2+
3+
<<<<<<< HEAD
4+
## Super-Easy, Open-Source Cloud-Based LLM Finetuning
5+
6+
**Try here –** [![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/simplifine-llm/Simplifine/blob/main/examples/cloud_quickstart.ipynb)
7+
8+
### **Get a FREE API Key for FINETUNING [HERE](https://app.simplifine.com/#/signup)**
9+
=======
10+
Simplifine lets you invoke LLM finetuning with just one line of code using any Hugging Face dataset or model.
11+
> The easiest, fully open-source LLM finetuning library!
12+
13+
14+
**Get free Simplifine Cloud Credits to finetune [here](https://www.simplifine.com/api-key-interest)**
15+
16+
## Roadmap
17+
- **COMPREHENSIVE UPDATE of DOCUMENTATIONS on INCOMING (By Aug 9th, 2024) to match new config files.
18+
19+
## 🔄 Updates
20+
**v0.0.8 (2024-08-08)**
21+
- **Bug Fixes:** Code clean up and trainer fixes.
22+
- **New Feature:** Ability to define more complex configuration files for the trainer.
23+
- **Examples:** -New examples on training cloud and training a fake news detector.
24+
>>>>>>> 3bd68be2d8b6173abf701764e497240bff577925
25+
26+
27+
Simplifine streamlines LLM finetuning on any dataset or model with one simple command, handling all infrastructure, job management, cloud model storage, and inference.
28+
29+
## Features
30+
- **🚀 Easy Cloud-Based LLM Finetuning:** Fine-tune any LLM with just one command.
31+
32+
- **☁️ Seamless Cloud Integration:** Automatically manage the downloading, storing, and running of models directly from the cloud.
33+
34+
- **🤖 Built-in AI Assistance:** Get help with hyperparameter selection, synthetic dataset generation, and data quality checks.
35+
36+
- **🔄 On-Device to Cloud Switching:** Add a simple decorator to transition from local to cloud-based training.
37+
38+
- **⚡ Auto-Optimization:** Automatically optimizes model and data parallelization Unsloth (*coming soon!*), Deepspeed ✅ and FDSP ✅
39+
40+
- **📊 Custom Evaluation Support:** Use the built-in LLM for evaluations functions or import your own custom evaluation metrics.
41+
42+
- **💼 Community Support:** Asking any support questions on the Simplifine Community Discord.
43+
44+
- **🏅 Trusted by Leading Institutions:** Research labs at the University of Oxford rely on Simplifine for their LLM finetuning needs.
45+
46+
---
47+
48+
## 🏁 Quickstart
49+
50+
Get started here > [![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/simplifine-llm/Simplifine/blob/main/examples/cloud_quickstart.ipynb)
51+
52+
53+
## 📚 Documentation
54+
55+
Find our full documentation at [docs.simplifine.com](http://docs.simplifine.com).
56+
57+
## 📦 Installation
58+
59+
Installing from PyPI
60+
```bash
61+
pip install simplifine-alpha
62+
```
63+
64+
You can also directly install from github using the following command:
65+
```bash
66+
pip install git+https://github.com/simplifine-llm/Simplifine.git
67+
```
68+
69+
## 🤝 Contributing
70+
71+
We are looking for contributors! Please send an email to [[email protected]](mailto:[email protected]) to get onboarded! We welcome all types of contributions.
72+
73+
## 📄 License
74+
75+
Simplifine is licensed under the GNU General Public License Version 3. See the LICENSE file for more details.
76+
77+
<<<<<<< HEAD
78+
=======
79+
## 📚 Documentation
80+
MAJOR OVERHAUL OF DOCUMENTATION IN THE WORKS (Done by 11th Aug 2024). In the meantime, please use this [notebook](https://github.com/simplifine-llm/Simplifine/blob/main/examples/cloud_quickstart.ipynb) here to learn how to use the model.
81+
>>>>>>> 3bd68be2d8b6173abf701764e497240bff577925
82+
83+
## 💬 Support
84+
85+
If you have any suggestions for new features you'd like to see implemented, please raise an issue—we will work hard to make it happen ASAP! For any other questions, feel free to contact us at [[email protected]](mailto:[email protected]).
86+
87+
88+
89+
## 🔄 Updates
90+
91+
#### **v0.0.8**
92+
- **🐛 Bug Fixes:** Streamlined code and resolved trainer-related issues for smoother operation.
93+
- **✨ New Feature:** Introduced support for defining more complex configuration files, enhancing the flexibility of the trainer.
94+
- **📚 Documentation:** Added new examples, including tutorials on cloud-based training and creating a fake news detector.
95+
- **🔗 Updated Documentation:** Check out the latest docs at [docs.simplifine.com](https://docs.simplifine.com).
96+
97+
#### **v0.0.71**
98+
- **🐛 Bug Fixes:** Fixed issues that caused loading failures on certain configurations, ensuring broader compatibility.
99+
- **✨ New Feature:** Enabled direct installation from Git and added support for Hugging Face API Tokens, allowing access to restricted models.
100+
- **📚 Documentation:** Refreshed examples to reflect the latest features.
101+
102+
103+
104+
## ⛮ General Compute Considerations
105+
106+
We currently support both DistributedDataParallel (DDP) and ZeRO from DeepSpeed.
107+
108+
**TL;DR**:
109+
- **DDP** is useful when a model can fit in GPU memory (this includes gradients and activation states).
110+
- **ZeRO** is useful when a model requires sharding across multiple GPUs.
111+
112+
**Longer Version**:
113+
114+
- **DDP**: Distributed Data Parallel (DDP) creates a replica of the model on each processor (GPU). For example, imagine 8 GPUs, each being fed a single data point—this would make a batch size of 8. The model replicas are then updated on each device. DDP speeds up training by parallelizing the data-feeding process. However, DDP **fails** if the replica cannot fit in GPU memory. Remember, the memory not only hosts parameters but also gradients and optimizer states.
115+
116+
- **ZeRO**: ZeRO is a powerful optimization developed by DeepSpeed and comes in different stages (1, 2, and 3). Each stage shards different parts of the training process (parameters, gradients, and activation states). This is really useful if a model cannot fit in GPU memory. ZeRO also supports offloading to the CPU, making even more room for training larger models.
117+
118+
### Example Scenarios and Appropriate Optimization Methods:
119+
1. **LLaMA-3-8b model with 16-bit precision**: Use ZeRO Stage 3 on 8 A100s.
120+
2. **LLaMA-3-8b model with LoRA adapters**: Usually fine with DDP on A100s.
121+
3. **GPT-2 with 16-bit precision**: Use DDP.
122+
123+
## 🪲 FAQs
124+
125+
**Issue: RuntimeError: Error building extension 'cpu_adam' python dev**
126+
127+
This error occurs when `python-dev` is not installed, and ZeRO is using offload. To resolve this, try:
128+
129+
```bash
130+
# Try sudo apt-get install python3-dev if the following fails.
131+
apt-get install python-dev # for Python 2.x installs
132+
apt-get install python3-dev # for Python 3.x installs
133+
```
134+
135+
See this [link](https://stackoverflow.com/questions/21530577/fatal-error-python-h-no-such-file-or-directory)

examples/.DS_Store

-6 KB
Binary file not shown.

examples/SFT_finetune/readme.md

Lines changed: 0 additions & 19 deletions
This file was deleted.

examples/SFT_finetune/sft_ft.py

Lines changed: 0 additions & 15 deletions
This file was deleted.

examples/SFT_finetune_cloud/readme.md

Lines changed: 0 additions & 5 deletions
This file was deleted.

examples/SFT_finetune_cloud/sft_ft_cloud.py

Lines changed: 0 additions & 19 deletions
This file was deleted.

0 commit comments

Comments
 (0)