text
stringlengths
0
759
class_ref, pretrained_model_name_or_path, code_revision=code_revision, **hub_kwargs, **kwargs
)
File "/tmp/.cache/uv/environments-v2/f64622993fb28786/lib/python3.13/site-packages/transformers/dynamic_module_utils.py", line 570, in get_class_from_dynamic_module
final_module = get_cached_module_file(
repo_id,
...<8 lines>...
repo_type=repo_type,
)
File "/tmp/.cache/uv/environments-v2/f64622993fb28786/lib/python3.13/site-packages/transformers/dynamic_module_utils.py", line 393, in get_cached_module_file
modules_needed = check_imports(resolved_module_file)
File "/tmp/.cache/uv/environments-v2/f64622993fb28786/lib/python3.13/site-packages/transformers/dynamic_module_utils.py", line 225, in check_imports
raise ImportError(
...<2 lines>...
)
ImportError: This modeling file requires the following packages that were not found in your environment: einops. Run `pip install einops`
Traceback (most recent call last):
File "/tmp/LiquidAI_LFM2-VL-1.6B_0X7Okei.py", line 13, in <module>
pipe = pipeline("image-text-to-text", model="LiquidAI/LFM2-VL-1.6B", trust_remote_code=True)
File "/tmp/.cache/uv/environments-v2/6e3e46311085f7ee/lib/python3.13/site-packages/transformers/pipelines/__init__.py", line 1198, in pipeline
raise e
File "/tmp/.cache/uv/environments-v2/6e3e46311085f7ee/lib/python3.13/site-packages/transformers/pipelines/__init__.py", line 1189, in pipeline
processor = AutoProcessor.from_pretrained(processor, _from_pipeline=task, **hub_kwargs, **model_kwargs)
File "/tmp/.cache/uv/environments-v2/6e3e46311085f7ee/lib/python3.13/site-packages/transformers/models/auto/processing_auto.py", line 382, in from_pretrained
processor_class = get_class_from_dynamic_module(
processor_auto_map, pretrained_model_name_or_path, **kwargs
)
File "/tmp/.cache/uv/environments-v2/6e3e46311085f7ee/lib/python3.13/site-packages/transformers/dynamic_module_utils.py", line 569, in get_class_from_dynamic_module
final_module = get_cached_module_file(
repo_id,
...<8 lines>...
repo_type=repo_type,
)
File "/tmp/.cache/uv/environments-v2/6e3e46311085f7ee/lib/python3.13/site-packages/transformers/dynamic_module_utils.py", line 392, in get_cached_module_file
modules_needed = check_imports(resolved_module_file)
File "/tmp/.cache/uv/environments-v2/6e3e46311085f7ee/lib/python3.13/site-packages/transformers/dynamic_module_utils.py", line 224, in check_imports
raise ImportError(
...<2 lines>...
)
ImportError: This modeling file requires the following packages that were not found in your environment: PIL. Run `pip install PIL`
Everything was good in LiquidAI_LFM2-VL-1.6B_1.txt
Traceback (most recent call last):
File "/tmp/LiquidAI_LFM2-VL-450M_07uPCVK.py", line 13, in <module>
pipe = pipeline("image-text-to-text", model="LiquidAI/LFM2-VL-450M", trust_remote_code=True)
File "/tmp/.cache/uv/environments-v2/7e38b95086156af9/lib/python3.13/site-packages/transformers/pipelines/__init__.py", line 1178, in pipeline
raise e
File "/tmp/.cache/uv/environments-v2/7e38b95086156af9/lib/python3.13/site-packages/transformers/pipelines/__init__.py", line 1169, in pipeline
processor = AutoProcessor.from_pretrained(processor, _from_pipeline=task, **hub_kwargs, **model_kwargs)
File "/tmp/.cache/uv/environments-v2/7e38b95086156af9/lib/python3.13/site-packages/transformers/models/auto/processing_auto.py", line 376, in from_pretrained
processor_class = get_class_from_dynamic_module(
processor_auto_map, pretrained_model_name_or_path, **kwargs
)
File "/tmp/.cache/uv/environments-v2/7e38b95086156af9/lib/python3.13/site-packages/transformers/dynamic_module_utils.py", line 570, in get_class_from_dynamic_module
final_module = get_cached_module_file(
repo_id,
...<8 lines>...
repo_type=repo_type,
)
File "/tmp/.cache/uv/environments-v2/7e38b95086156af9/lib/python3.13/site-packages/transformers/dynamic_module_utils.py", line 393, in get_cached_module_file
modules_needed = check_imports(resolved_module_file)
File "/tmp/.cache/uv/environments-v2/7e38b95086156af9/lib/python3.13/site-packages/transformers/dynamic_module_utils.py", line 225, in check_imports
raise ImportError(
...<2 lines>...
)
ImportError: This modeling file requires the following packages that were not found in your environment: PIL. Run `pip install PIL`
Everything was good in LiquidAI_LFM2-VL-450M_1.txt
No suitable GPU found for MiniMaxAI/MiniMax-M1-80k | 1104.39 GB VRAM requirement
No suitable GPU found for MiniMaxAI/MiniMax-M1-80k | 1104.39 GB VRAM requirement
No suitable GPU found for MiniMaxAI/MiniMax-M1-80k | 1104.39 GB VRAM requirement
No suitable GPU found for MiniMaxAI/MiniMax-M1-80k | 1104.39 GB VRAM requirement
No suitable GPU found for OpenGVLab/InternVL3_5-241B-A28B | 582.84 GB VRAM requirement
No suitable GPU found for OpenGVLab/InternVL3_5-241B-A28B | 582.84 GB VRAM requirement
No suitable GPU found for OpenGVLab/InternVL3_5-30B-A3B | 74.70 GB VRAM requirement
No suitable GPU found for OpenGVLab/InternVL3_5-30B-A3B | 74.70 GB VRAM requirement
No suitable GPU found for OpenGVLab/InternVL3_5-38B | 92.96 GB VRAM requirement
No suitable GPU found for OpenGVLab/InternVL3_5-38B | 92.96 GB VRAM requirement
Traceback (most recent call last):
File "/tmp/OpenGVLab_InternVL3_5-8B_06uMxsa.py", line 13, in <module>
pipe = pipeline("image-text-to-text", model="OpenGVLab/InternVL3_5-8B", trust_remote_code=True)
File "/tmp/.cache/uv/environments-v2/92a0710a497a595f/lib/python3.13/site-packages/transformers/pipelines/__init__.py", line 1028, in pipeline
framework, model = infer_framework_load_model(
~~~~~~~~~~~~~~~~~~~~~~~~~~^
adapter_path if adapter_path is not None else model,
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
...<5 lines>...
**model_kwargs,
^^^^^^^^^^^^^^^
)
^
File "/tmp/.cache/uv/environments-v2/92a0710a497a595f/lib/python3.13/site-packages/transformers/pipelines/base.py", line 333, in infer_framework_load_model
raise ValueError(
f"Could not load model {model} with any of the following classes: {class_tuple}. See the original errors:\n\n{error}\n"
)
ValueError: Could not load model OpenGVLab/InternVL3_5-8B with any of the following classes: (<class 'transformers.models.auto.modeling_auto.AutoModelForImageTextToText'>,). See the original errors:
while loading with AutoModelForImageTextToText, an error is thrown:
Traceback (most recent call last):
File "/tmp/.cache/uv/environments-v2/92a0710a497a595f/lib/python3.13/site-packages/transformers/pipelines/base.py", line 293, in infer_framework_load_model
model = model_class.from_pretrained(model, **kwargs)
File "/tmp/.cache/uv/environments-v2/92a0710a497a595f/lib/python3.13/site-packages/transformers/models/auto/auto_factory.py", line 607, in from_pretrained
raise ValueError(