diff --git a/optional_requirements.txt b/optional_requirements.txt new file mode 100644 index 00000000..0e03b238 --- /dev/null +++ b/optional_requirements.txt @@ -0,0 +1 @@ +gptqmodel diff --git a/requirements.txt b/requirements.txt index 4f022b43..026ec48b 100644 --- a/requirements.txt +++ b/requirements.txt @@ -17,8 +17,6 @@ sentencepiece==0.2.0 torchvision==0.21.0 xformers==0.0.29.post3 -# InternLM-XComposer2 -gptqmodel==1.9.0 numpy==2.2.3 # WD Tagger diff --git a/taggui/auto_captioning/models_list.py b/taggui/auto_captioning/models_list.py index e5c3d15a..5bfbfe37 100644 --- a/taggui/auto_captioning/models_list.py +++ b/taggui/auto_captioning/models_list.py @@ -11,14 +11,13 @@ from auto_captioning.models.moondream import Moondream1, Moondream2 from auto_captioning.models.phi_3_vision import Phi3Vision from auto_captioning.models.wd_tagger import WdTagger -from auto_captioning.models.xcomposer2 import Xcomposer2, Xcomposer2_4khd +try: + from auto_captioning.models.xcomposer2 import Xcomposer2, Xcomposer2_4khd +except: + print('GPTQModel probably failed to install') MODELS = [ 'fancyfeast/llama-joycaption-alpha-two-hf-llava', - 'internlm/internlm-xcomposer2-vl-7b-4bit', - 'internlm/internlm-xcomposer2-vl-7b', - 'internlm/internlm-xcomposer2-vl-1_8b', - 'internlm/internlm-xcomposer2-4khd-7b', 'THUDM/cogvlm-chat-hf', 'THUDM/cogagent-vqa-hf', 'THUDM/cogvlm2-llama3-chat-19B-int4', @@ -60,7 +59,12 @@ 'Salesforce/blip2-flan-t5-xxl', 'microsoft/kosmos-2-patch14-224' ] - +if Xcomposer: + MODELS.extend([ + 'internlm/internlm-xcomposer2-vl-7b-4bit', + 'internlm/internlm-xcomposer2-vl-7b', + 'internlm/internlm-xcomposer2-vl-1_8b', + 'internlm/internlm-xcomposer2-4khd-7b',]) def get_model_class(model_id: str) -> type[AutoCaptioningModel]: lowercase_model_id = model_id.lower()