Skip to content

Confusion about finetune_cls.sh #175

@TUDelftHao

Description

@TUDelftHao

Thank you for this amazing project!

While I am going to conduct the QwenVL classification task, I have some questions about the parameters used in finetune_cls.sh:

  1. What is the global batch size used for the default learning rate?
  2. Why is the head_lr different from learning_rate? Is this the empirical set?
  3. Why are you setting the llm frozen while the vision tower and merger module trainable, which is in contrast to the setup of sft the mllm (usually we only finetune the LLM part).

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions