Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

about finetune train #90

Closed
ZuyongWu opened this issue Jun 7, 2024 · 2 comments
Closed

about finetune train #90

ZuyongWu opened this issue Jun 7, 2024 · 2 comments

Comments

@ZuyongWu
Copy link

ZuyongWu commented Jun 7, 2024

I only have 4x4090 cards,under this circumstance, can i finetune a MLLM?
How should I do to train, thanks a lot.

@RussRobin
Copy link
Collaborator

RussRobin commented Jul 6, 2024

Thank you for your interest in Bunny.
24G per device is enough to pretrain and finetune Bunny. However, the actual GPU memory consumption depends on your base model, image resolution and data.

For finetuning, setting per-device-batch-size to 2 or 4 may be good to you. In order to use the default learning rate in finetune_lora.sh, we recommend keeping global batch size 128. Global batch size = num of GPU * batch size per GPU * accumulation step. In your case, num of GPU is 4. All these parameters can be set in finetune_lora.sh. Similarly, set batch size that fits for you in pretraining, of full parameter tuning.

Feel free to further comment on this issue if you meet any problems in using Bunny.

Regards
Russell

@RussRobin
Copy link
Collaborator

I'll close this issue since no further discussion is raised yet. Please reopen it if you still have concerns.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants