Skip to content

Why does optimization_strategy differ between training and prediction for Fine-Tuning Qwen2.5-VL? #174

Answered by SkalskiP
dcfabian asked this question in Q&A
Discussion options

You must be logged in to vote

Hi @dcfabian 👋🏻 Good question!

When training, we use an optimization strategy like LoRA or QLoRA to help the model learn efficiently. QLoRA applies techniques like quantization and low-rank adaptation to reduce memory usage and speed up training without losing performance. This makes the training process more efficient and resource-friendly.

Once training is complete, however, the model has already learned the necessary patterns and relationships. For prediction (or inference), we simply load the trained checkpoint and run the model as-is—there’s no need to use the training-specific optimizations. That’s why the optimization strategy is set to NONE during prediction.

In short, QLoRA is us…

Replies: 1 comment 1 reply

Comment options

You must be logged in to vote
1 reply
@dcfabian
Comment options

Answer selected by dcfabian
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Category
Q&A
Labels
None yet
2 participants