File tree 1 file changed +4
-2
lines changed
docker/transformers-quantization-latest-gpu
1 file changed +4
-2
lines changed Original file line number Diff line number Diff line change @@ -36,8 +36,10 @@ RUN python3 -m pip install --no-cache-dir einops
36
36
# Add bitsandbytes for mixed int8 testing
37
37
RUN python3 -m pip install --no-cache-dir bitsandbytes
38
38
39
- # Add auto-gptq for gtpq quantization testing
40
- RUN python3 -m pip install --no-cache-dir auto-gptq --extra-index-url https://huggingface.github.io/autogptq-index/whl/cu118/
39
+ # Add auto-gptq for gtpq quantization testing, installed from source for pytorch==2.5.1 compatibility
40
+ # TORCH_CUDA_ARCH_LIST="7.5+PTX" is added to make the package compile for Tesla T4 gpus available for the CI.
41
+ RUN pip install gekko
42
+ RUN git clone https://github.com/PanQiWei/AutoGPTQ.git && cd AutoGPTQ && TORCH_CUDA_ARCH_LIST="7.5+PTX" python3 setup.py install
41
43
42
44
# Add optimum for gptq quantization testing
43
45
RUN python3 -m pip install --no-cache-dir git+https://github.com/huggingface/optimum@main#egg=optimum
You can’t perform that action at this time.
0 commit comments