You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Jina.ai support a token limit of 8192 for generating the embeddings. For late chunking if my context is more than 8192, then what are the best strategies to implement late chunking?
The text was updated successfully, but these errors were encountered:
I think if you have very long documents, not all of the context might be necessary. So if you can split the text into chapters or longer sections, there might be enough context for the embedding model to interpret all of the tokens correctly. Otherwise you can also pass a bit more text before and after the first chunk yu are interested. Maybe als adding summaries before the text chunks could further improve it, but I haven't tried something like this.
Now we have implemented a strategy that uses overlapping macro chunks to solve this problem. Just set --long-late-chunking-embed-size to the maximum context length of the model that you are using and it will automatically use this strategy.
help='Token length of the embeddings that come before/after soft boundaries (i.e. overlapping embeddings). Above zero, overlap is used between neighbouring embeddings.',
Jina.ai support a token limit of 8192 for generating the embeddings. For late chunking if my context is more than 8192, then what are the best strategies to implement late chunking?
The text was updated successfully, but these errors were encountered: