Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Why most people fine-tune full model instead of just adapter? Should we educate people more on that topic? #518

Open
cyberluke opened this issue Oct 10, 2024 · 0 comments

Comments

@cyberluke
Copy link

Why most people fine-tune full model instead of just adapter? Should we educate people more on that topic?

You could then use parallel adapter forward pass as implemented by the ‘Parallel’ block and it is already supported by HunggingFace:

Video education about full fine tuning vs adapter
https://www.youtube.com/watch?v=s2BF_gC0X1o

You could compose several adapters in parallel for each language with fine-tuning and compose it on top of base model saving memory, space. This is more modular approach.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant