Add inputs_embeds as alternative for input_ids #158
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
Adding to the model.forward inputs_embeds (same name as HF models use) to give the model token embedding directly rather than token ids
The main use case is training soft prompts without adding N tokens to tokenizer that never get outputted(and then jumping through hoops to train them only without touching existing tokens)
(ETA: from more recent example: techniques like neft are much easier to do when it's possible to pass embedding directly)