Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Quantization Configuration in VLM Model #301

Open
aptsunny opened this issue Jan 16, 2025 · 3 comments
Open

Quantization Configuration in VLM Model #301

aptsunny opened this issue Jan 16, 2025 · 3 comments

Comments

@aptsunny
Copy link

I've been going through the latest commit with ID 69d7dc0 and noticed that there seems to be an issue with the quantization configuration for the VLM model. The original configuration file is configs/quantization/methods/Awq/awq_w_only_vlm.yml, which is shown below:

calib:
    name: custom_mm
    download: False
    path: calib data path
    apply_chat_template: True
    add_answer: True # Default is False. If set it to True, calib data will add answers.
    n_samples: 8
    bs: -1
    seq_len: 512
    padding: True
    seed: *seed

I've found that this configuration for the calibration dataset can lead to a mismatch in the model dimensions for the LLM part. This issue has been observed in both InternVL2 and Qwen2VL models. I've attached the bug log files for your reference.
I believe resolving this would greatly improve the performance and stability of the VLM model. If you need any further information or clarification for bug verification, please feel free to reply to this issue on GitHub. Thank you so much for your time and effort.
Qwen2VL_awq.log

InternVL2_awq.log

@XiaotaoChen
Copy link

Hi, i met another bug when quantizing internVL2 ref the config: configs/quantization/methods/Awq/awq_w_only_vlm.yml, the detail info is here #303. have you met it?

@aptsunny
Copy link
Author

Hi, i met another bug when quantizing internVL2 ref the config: configs/quantization/methods/Awq/awq_w_only_vlm.yml, the detail info is here #303. have you met it?

I recommend checking the download method for the MME dataset carefully

@XiaotaoChen
Copy link

Hi, i met another bug when quantizing internVL2 ref the config: configs/quantization/methods/Awq/awq_w_only_vlm.yml, the detail info is here #303. have you met it?

I recommend checking the download method for the MME dataset carefully

thanks. it's solved

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants