可亲
fix(pad zero) pad intermediate_size to 29696 to make sure quantized model can use 8 tensor-parallel in vllm
d1eab90
-
1.52 kB
-
6.96 kB
-
18.5 kB
-
392 Bytes
-
1.05 kB
-
1.39 kB
-
247 Bytes
-
1.67 MB
-
3.97 GB
LFS
-
3.93 GB
LFS
-
3.86 GB
LFS
-
3.93 GB
LFS
-
3.93 GB
LFS
-
3.86 GB
LFS
-
3.93 GB
LFS
-
3.93 GB
LFS
-
3.86 GB
LFS
-
3.93 GB
LFS
-
3.93 GB
LFS
-
3.86 GB
LFS
-
3.93 GB
LFS
-
3.93 GB
LFS
-
3.86 GB
LFS
-
3.93 GB
LFS
-
3.93 GB
LFS
-
3.86 GB
LFS
-
3.93 GB
LFS
-
1.73 GB
LFS
-
2.49 GB
LFS
-
244 kB
-
594 Bytes
-
207 Bytes
-
613 Bytes
-
7.03 MB
-
4.3 kB
-
2.78 MB