-
Notifications
You must be signed in to change notification settings - Fork 1
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Run reference code for mixtral-8x7b
#48
Comments
1. Download a repository with reference code
Directory with the reference code:
2. Copy mperf.conf to mixtral-8x7b
3. Set python3 version to 3.9
|
4. Install python packages
If you want to use CPU:
We don't have this version
Now we have only these
in https://download.pytorch.org/whl/nightly/torch/
|
If run on CPU:
|
If run on GPU:
For running experiments we also need pandas:
|
5. Install loadgen
|
6. Get model (checkpoint)
Results:
|
7. Download dataset
We don't need calibration dataset for accuracy/performance running. |
8. Run performance
|
The reason of this issue is missing of tokenizer files in downloaded model checkpoint:
These files are located in https://huggingface.co/mistralai/Mixtral-8x7B-Instruct-v0.1/tree/main |
So login to huggingface (https://huggingface.co/mistralai/Mixtral-8x7B-Instruct-v0.1/tree/main) and download these files to Windows.
|
Run Performance:
it was stopped because full experiment took a lot of time. |
AccuracySetting of
|
But we have an issue when we run
The reason of this issue is full dataset using but results for short run( |
If I commented code for Open Orca and MBXP samples we can have:
|
Solution:Create dataset with 15 samples:
Accuracy
6.93 mins per 1 sample. |
Accuracy on GPU( all GPUs on
|
For checking how it runs on GPU we should use
|
According to README.md run reference code for
mixtral-8x7b
.The text was updated successfully, but these errors were encountered: