Replies: 1 comment 3 replies
-
@dluc Maybe you or someone else can guide on this? |
Beta Was this translation helpful? Give feedback.
3 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
Hi everyone,
I'm working with Microsoft Kernel Memory's ImportDocumentAsync() and running into an issue where only the last chunk is being stored in Azure AI Search when I pass multiple chunks with the same documentId.
Setup:
I'm performing chunking before calling ImportDocumentAsync().
I'm not using the built-in chunking (TextPartitioningOptions).
Each chunk has a unique PageNumber in the metadata.
However, only one chunk appears in AI Search (likely overwriting previous ones).
Issue:
Does ImportDocumentAsync() expect all chunks of a document to be passed together?
If I pass multiple chunks sequentially with the same documentId, does it overwrite earlier chunks?
Should I be adding a unique identifier per chunk (like ChunkID) to ensure all chunks are indexed?
What I’ve Tried:
Verified logging: Each chunk is passed separately, but only one remains in AI Search.
Tested adding PageNumber as metadata, but that didn’t prevent overwrites.
Considering modifying documentId per chunk, but that might break document-level search.
Would appreciate insights from anyone who has worked with Microsoft Kernel Memory and Azure AI Search indexing!
I can provide the code snippet if that would be helpful.
Thanks!
Beta Was this translation helpful? Give feedback.
All reactions