text
stringlengths 0
759
|
---|
class_ref, pretrained_model_name_or_path, code_revision=code_revision, **hub_kwargs, **kwargs
|
)
|
File "/tmp/.cache/uv/environments-v2/f64622993fb28786/lib/python3.13/site-packages/transformers/dynamic_module_utils.py", line 570, in get_class_from_dynamic_module
|
final_module = get_cached_module_file(
|
repo_id,
|
...<8 lines>...
|
repo_type=repo_type,
|
)
|
File "/tmp/.cache/uv/environments-v2/f64622993fb28786/lib/python3.13/site-packages/transformers/dynamic_module_utils.py", line 393, in get_cached_module_file
|
modules_needed = check_imports(resolved_module_file)
|
File "/tmp/.cache/uv/environments-v2/f64622993fb28786/lib/python3.13/site-packages/transformers/dynamic_module_utils.py", line 225, in check_imports
|
raise ImportError(
|
...<2 lines>...
|
)
|
ImportError: This modeling file requires the following packages that were not found in your environment: einops. Run `pip install einops`
|
Traceback (most recent call last):
|
File "/tmp/LiquidAI_LFM2-VL-1.6B_0X7Okei.py", line 13, in <module>
|
pipe = pipeline("image-text-to-text", model="LiquidAI/LFM2-VL-1.6B", trust_remote_code=True)
|
File "/tmp/.cache/uv/environments-v2/6e3e46311085f7ee/lib/python3.13/site-packages/transformers/pipelines/__init__.py", line 1198, in pipeline
|
raise e
|
File "/tmp/.cache/uv/environments-v2/6e3e46311085f7ee/lib/python3.13/site-packages/transformers/pipelines/__init__.py", line 1189, in pipeline
|
processor = AutoProcessor.from_pretrained(processor, _from_pipeline=task, **hub_kwargs, **model_kwargs)
|
File "/tmp/.cache/uv/environments-v2/6e3e46311085f7ee/lib/python3.13/site-packages/transformers/models/auto/processing_auto.py", line 382, in from_pretrained
|
processor_class = get_class_from_dynamic_module(
|
processor_auto_map, pretrained_model_name_or_path, **kwargs
|
)
|
File "/tmp/.cache/uv/environments-v2/6e3e46311085f7ee/lib/python3.13/site-packages/transformers/dynamic_module_utils.py", line 569, in get_class_from_dynamic_module
|
final_module = get_cached_module_file(
|
repo_id,
|
...<8 lines>...
|
repo_type=repo_type,
|
)
|
File "/tmp/.cache/uv/environments-v2/6e3e46311085f7ee/lib/python3.13/site-packages/transformers/dynamic_module_utils.py", line 392, in get_cached_module_file
|
modules_needed = check_imports(resolved_module_file)
|
File "/tmp/.cache/uv/environments-v2/6e3e46311085f7ee/lib/python3.13/site-packages/transformers/dynamic_module_utils.py", line 224, in check_imports
|
raise ImportError(
|
...<2 lines>...
|
)
|
ImportError: This modeling file requires the following packages that were not found in your environment: PIL. Run `pip install PIL`
|
Everything was good in LiquidAI_LFM2-VL-1.6B_1.txt
|
Traceback (most recent call last):
|
File "/tmp/LiquidAI_LFM2-VL-450M_07uPCVK.py", line 13, in <module>
|
pipe = pipeline("image-text-to-text", model="LiquidAI/LFM2-VL-450M", trust_remote_code=True)
|
File "/tmp/.cache/uv/environments-v2/7e38b95086156af9/lib/python3.13/site-packages/transformers/pipelines/__init__.py", line 1178, in pipeline
|
raise e
|
File "/tmp/.cache/uv/environments-v2/7e38b95086156af9/lib/python3.13/site-packages/transformers/pipelines/__init__.py", line 1169, in pipeline
|
processor = AutoProcessor.from_pretrained(processor, _from_pipeline=task, **hub_kwargs, **model_kwargs)
|
File "/tmp/.cache/uv/environments-v2/7e38b95086156af9/lib/python3.13/site-packages/transformers/models/auto/processing_auto.py", line 376, in from_pretrained
|
processor_class = get_class_from_dynamic_module(
|
processor_auto_map, pretrained_model_name_or_path, **kwargs
|
)
|
File "/tmp/.cache/uv/environments-v2/7e38b95086156af9/lib/python3.13/site-packages/transformers/dynamic_module_utils.py", line 570, in get_class_from_dynamic_module
|
final_module = get_cached_module_file(
|
repo_id,
|
...<8 lines>...
|
repo_type=repo_type,
|
)
|
File "/tmp/.cache/uv/environments-v2/7e38b95086156af9/lib/python3.13/site-packages/transformers/dynamic_module_utils.py", line 393, in get_cached_module_file
|
modules_needed = check_imports(resolved_module_file)
|
File "/tmp/.cache/uv/environments-v2/7e38b95086156af9/lib/python3.13/site-packages/transformers/dynamic_module_utils.py", line 225, in check_imports
|
raise ImportError(
|
...<2 lines>...
|
)
|
ImportError: This modeling file requires the following packages that were not found in your environment: PIL. Run `pip install PIL`
|
Everything was good in LiquidAI_LFM2-VL-450M_1.txt
|
No suitable GPU found for MiniMaxAI/MiniMax-M1-80k | 1104.39 GB VRAM requirement
|
No suitable GPU found for MiniMaxAI/MiniMax-M1-80k | 1104.39 GB VRAM requirement
|
No suitable GPU found for MiniMaxAI/MiniMax-M1-80k | 1104.39 GB VRAM requirement
|
No suitable GPU found for MiniMaxAI/MiniMax-M1-80k | 1104.39 GB VRAM requirement
|
No suitable GPU found for OpenGVLab/InternVL3_5-241B-A28B | 582.84 GB VRAM requirement
|
No suitable GPU found for OpenGVLab/InternVL3_5-241B-A28B | 582.84 GB VRAM requirement
|
No suitable GPU found for OpenGVLab/InternVL3_5-30B-A3B | 74.70 GB VRAM requirement
|
No suitable GPU found for OpenGVLab/InternVL3_5-30B-A3B | 74.70 GB VRAM requirement
|
No suitable GPU found for OpenGVLab/InternVL3_5-38B | 92.96 GB VRAM requirement
|
No suitable GPU found for OpenGVLab/InternVL3_5-38B | 92.96 GB VRAM requirement
|
Traceback (most recent call last):
|
File "/tmp/OpenGVLab_InternVL3_5-8B_06uMxsa.py", line 13, in <module>
|
pipe = pipeline("image-text-to-text", model="OpenGVLab/InternVL3_5-8B", trust_remote_code=True)
|
File "/tmp/.cache/uv/environments-v2/92a0710a497a595f/lib/python3.13/site-packages/transformers/pipelines/__init__.py", line 1028, in pipeline
|
framework, model = infer_framework_load_model(
|
~~~~~~~~~~~~~~~~~~~~~~~~~~^
|
adapter_path if adapter_path is not None else model,
|
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
...<5 lines>...
|
**model_kwargs,
|
^^^^^^^^^^^^^^^
|
)
|
^
|
File "/tmp/.cache/uv/environments-v2/92a0710a497a595f/lib/python3.13/site-packages/transformers/pipelines/base.py", line 333, in infer_framework_load_model
|
raise ValueError(
|
f"Could not load model {model} with any of the following classes: {class_tuple}. See the original errors:\n\n{error}\n"
|
)
|
ValueError: Could not load model OpenGVLab/InternVL3_5-8B with any of the following classes: (<class 'transformers.models.auto.modeling_auto.AutoModelForImageTextToText'>,). See the original errors:
|
while loading with AutoModelForImageTextToText, an error is thrown:
|
Traceback (most recent call last):
|
File "/tmp/.cache/uv/environments-v2/92a0710a497a595f/lib/python3.13/site-packages/transformers/pipelines/base.py", line 293, in infer_framework_load_model
|
model = model_class.from_pretrained(model, **kwargs)
|
File "/tmp/.cache/uv/environments-v2/92a0710a497a595f/lib/python3.13/site-packages/transformers/models/auto/auto_factory.py", line 607, in from_pretrained
|
raise ValueError(
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.