RuntimeError: shape '[32, 8]' is invalid for input of size 0
#13
by
SpaceCowboy850
- opened
I haven't "built from source" as instructed, but the official releases of both Transformers and Auto_GPTQ are now at/beyond the dev branch suggested. Is it reasonable to assume that the latest versions have what is needed to run this?
I'm running this: GPTQ_3bit_128g_actorderTrue
On a 4090, and ultimately get this
return forward_call(*args, **kwargs)
File "C:\Anaconda3\envs\huggingface\lib\site-packages\transformers\models\mixtral\modeling_mixtral.py", line 802, in forward
router_logits = self.gate(hidden_states)
File "C:\Anaconda3\envs\huggingface\lib\site-packages\torch\nn\modules\module.py", line 1518, in _wrapped_call_impl
return self._call_impl(*args, **kwargs)
File "C:\Anaconda3\envs\huggingface\lib\site-packages\torch\nn\modules\module.py", line 1527, in _call_impl
return forward_call(*args, **kwargs)
File "C:\Anaconda3\envs\huggingface\lib\site-packages\auto_gptq\nn_modules\qlinear\qlinear_cuda.py", line 245, in forward
zeros = zeros.reshape(self.scales.shape)
RuntimeError: shape '[32, 8]' is invalid for input of size 0
Making sure there are no conflicts, here are the versions from a python command line:
>>> import transformers
>>> transformers.__version__
'4.37.0'
>>> import auto_gptq
>>> auto_gptq.__version__
'0.6.0'
Posting this here as a link for anyone that finds it. I still haven't solved my problem, but the primary discussion seems to be in this thread
https://huggingface.co/TheBloke/Mixtral-8x7B-v0.1-GPTQ/discussions/5
SpaceCowboy850
changed discussion status to
closed