Hello,
I'm trying to reproduce the following two model files from the v1-backup:
qwen1.5-1.8b-chat-rot_q4_0.mllm
qwen1.5-1.8b-chat-rot_qnn.mllm
I strictly followed the instructions step by step in:
However, the generated qwen1.5-1.8b-chat-rot_q4_0.mllm file is 1.0 GB in size, which does not match the file size of the official version hosted on ModelScope:
🔗 mllmTeam/Qwen1.5-1.8B-Chat on ModelScope
- My generated file:
qwen1.5-1.8b-chat-rot_q4_0.mllm (1.0 GB)
- Official ModelScope file:
qwen1.5-1.8b-chat-rot_q4_0.mllm (3.17GB)
Whether there are additional quantization or export steps I might have missed? How did you obtain the model files?
Thank you for your help!