Adding a max chunk size with SemanticChunker #18014
killian-mahe
announced in
Ideas
Replies: 3 comments
-
I've created a pull requet for this feature : #18019 |
Beta Was this translation helpful? Give feedback.
0 replies
-
Please add this feature because otherwise it is very difficult to use SemanticChunker. |
Beta Was this translation helpful? Give feedback.
0 replies
-
I'm not sure I understood the suggestion, but I created a PR for it #24342. |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
Checked
Feature request
Adding a way to limit the size of the chunks created by the
SemanticChunker
.Motivation
The models have limits on the number of input tokens and cannot embed chunks that are too large.
SemanticChunker
has no limit on chunk size, which can lead to errors when chunking or embedding large chunks.Proposal (If applicable)
Add a
max_chunk_size
paramter in theSemanticChunker
instantiation.Beta Was this translation helpful? Give feedback.
All reactions