Skip to content
This repository has been archived by the owner on Jul 1, 2024. It is now read-only.

Discriminative learning rates for FineTuningTask #289

Open
vreis opened this issue Dec 5, 2019 · 0 comments
Open

Discriminative learning rates for FineTuningTask #289

vreis opened this issue Dec 5, 2019 · 0 comments
Labels
enhancement New feature or request

Comments

@vreis
Copy link
Contributor

vreis commented Dec 5, 2019

🚀 Feature

Right now we only support fine tuning by freezing the trunk weights, or training all weights together. Discriminative learning rates means we can apply different learning rates for different parts of the model, which usually leads to better performance.

Motivation

https://arxiv.org/pdf/1801.06146.pdf introduced discriminative fine-tuning in NLP. Since then it's been found to be useful in computer vision as well.

Pitch

This could be implemented in either FineTuningTask or ClassyModel. I'd rather keep ClassyModel as simple as possible and move this type of logic to the task level.

Alternatives

N/A

Additional context

N/A

@vreis vreis added the enhancement New feature or request label Dec 5, 2019
Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Labels
enhancement New feature or request
Projects
None yet
Development

No branches or pull requests

1 participant