Skip to content

Commit f491096

Browse files
authored
Fix docker CI : install autogptq from source (#35000)
* Fixed Docker * Test ci * Finally * add comment
1 parent 01ad80f commit f491096

File tree

1 file changed

+4
-2
lines changed
  • docker/transformers-quantization-latest-gpu

1 file changed

+4
-2
lines changed

docker/transformers-quantization-latest-gpu/Dockerfile

+4-2
Original file line numberDiff line numberDiff line change
@@ -36,8 +36,10 @@ RUN python3 -m pip install --no-cache-dir einops
3636
# Add bitsandbytes for mixed int8 testing
3737
RUN python3 -m pip install --no-cache-dir bitsandbytes
3838

39-
# Add auto-gptq for gtpq quantization testing
40-
RUN python3 -m pip install --no-cache-dir auto-gptq --extra-index-url https://huggingface.github.io/autogptq-index/whl/cu118/
39+
# Add auto-gptq for gtpq quantization testing, installed from source for pytorch==2.5.1 compatibility
40+
# TORCH_CUDA_ARCH_LIST="7.5+PTX" is added to make the package compile for Tesla T4 gpus available for the CI.
41+
RUN pip install gekko
42+
RUN git clone https://github.com/PanQiWei/AutoGPTQ.git && cd AutoGPTQ && TORCH_CUDA_ARCH_LIST="7.5+PTX" python3 setup.py install
4143

4244
# Add optimum for gptq quantization testing
4345
RUN python3 -m pip install --no-cache-dir git+https://github.com/huggingface/optimum@main#egg=optimum

0 commit comments

Comments
 (0)