Skip to content

Lumen-Labs/cpp-chunker

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

4 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Semantic Text Chunker

Advanced semantic text chunking library that preserves meaning and context

Overview

This project provides a C++ library (with Python bindings via pybind11) for fast semantically chunking large bodies of text. It aims to split text into coherent, context-preserving segments using semantic similarity, discourse markers, and section detection, fast.

Features

  • Semantic chunking of text based on meaning and context
  • Adjustable chunk size and coherence thresholds
  • Extraction of chunk details: coherence scores, dominant topics, section types
  • Python bindings for easy integration
  • Memory-safe C++ implementation with comprehensive error handling

Installation

Prerequisites

  • C++17 compiler (GCC 7+, Clang 5+, or MSVC 2017+)
  • pybind11 (version 2.10.0+)
  • CMake (version 3.16+)
  • Python 3.7+

Installing pybind11

Option 1: Using pip (Recommended)

pip install pybind11

Option 2: Using conda

conda install pybind11

Option 3: From source

git clone https://github.com/pybind/pybind11.git
cd pybind11
pip install .

Option 4: System package manager

# Ubuntu/Debian
sudo apt-get install pybind11-dev

# macOS with Homebrew
brew install pybind11

# Arch Linux
sudo pacman -S pybind11

Build Instructions

Method 1: Using CMake (Recommended)

mkdir build
cd build
cmake ..
make

Method 2: Using pip (Automatic build)

pip install .

Method 3: Development installation

pip install -e .

Usage

Basic Usage

Simple Text Chunking

import chunker_cpp

text = """This is a long text for testing the semantic chunker. It contains multiple sentences, some of which are quite lengthy and elaborate, while others are short. The purpose of this text is to simulate a realistic document that might be processed by the chunker.

In addition to regular sentences, this text includes various structures such as lists:
- First item in the list.
- Second item, which is a bit longer and more descriptive.
- Third item.

There are also paragraphs separated by blank lines.

Here is a new paragraph. It discusses a different topic and is intended to test how the chunker handles paragraph boundaries. Sometimes, paragraphs can be very long, spanning several lines and containing a lot of information. Other times, they are short.

Finally, this text ends with a concluding sentence to ensure the chunker can handle the end of input gracefully."""

# Basic chunking with default parameters
chunks = chunker_cpp.chunk_text_semantically(text)
print("Number of chunks:", len(chunks))
for i, chunk in enumerate(chunks):
    print(f"Chunk {i+1}: {chunk[:100]}...")

Advanced Usage with Custom Parameters

# Custom chunking parameters
chunks = chunker_cpp.chunk_text_semantically(
    text,
    max_chunk_size=1000,      # Maximum characters per chunk
    min_chunk_size=200,       # Minimum characters per chunk
    min_coherence_threshold=0.5  # Higher threshold for more coherent chunks
)

Getting Detailed Chunk Information

# Get detailed information about each chunk
chunk_details = chunker_cpp.get_chunk_details(text)

for i, detail in enumerate(chunk_details):
    print(f"Chunk {i+1}:")
    print(f"  Text: {detail['text'][:100]}...")
    print(f"  Coherence Score: {detail['coherence_score']:.3f}")
    print(f"  Dominant Topics: {detail['dominant_topics']}")
    print(f"  Sentence Count: {detail['sentence_count']}")
    print(f"  Section Type: {detail['primary_section_type']}")
    print()

Using the Class Interface

# Create a chunker instance
chunker = chunker_cpp.SemanticTextChunker()

# Use the instance methods
chunks = chunker.chunk_text_semantically(text, max_chunk_size=1500)
details = chunker.get_chunk_details(text, min_coherence_threshold=0.4)

Error Handling and Memory Safety

The library includes comprehensive error handling and memory safety features:

import chunker_cpp

# Safe handling of edge cases
try:
    # Empty string
    result = chunker_cpp.chunk_text_semantically("")
    print("Empty string handled:", result)

    # Very large text
    large_text = "This is a test sentence. " * 10000
    result = chunker_cpp.chunk_text_semantically(large_text, 1000, 200)
    print("Large text processed successfully")

    # Special characters
    special_text = "Text with special chars: \x00\x01\x02\n\r\t" + "A" * 1000
    result = chunker_cpp.chunk_text_semantically(special_text)
    print("Special characters handled safely")

except Exception as e:
    print(f"Error: {e}")

Performance Considerations

import time

# Benchmarking chunking performance
text = "Your large text here..." * 100  # Create a large text

start_time = time.time()
chunks = chunker_cpp.chunk_text_semantically(text)
end_time = time.time()

print(f"Processed {len(text)} characters in {end_time - start_time:.3f} seconds")
print(f"Generated {len(chunks)} chunks")
print(f"Average chunk size: {sum(len(chunk) for chunk in chunks) / len(chunks):.0f} characters")

API Reference

Functions

chunk_text_semantically(text, max_chunk_size=2000, min_chunk_size=500, min_coherence_threshold=0.3)

Chunk text semantically while preserving meaning and context.

Parameters:

  • text (str): The input text to be chunked
  • max_chunk_size (int): Maximum size for each chunk (default: 2000)
  • min_chunk_size (int): Minimum size for each chunk (default: 500)
  • min_coherence_threshold (float): Minimum coherence threshold (default: 0.3)

Returns:

  • List[str]: List of text chunks

get_chunk_details(text, max_chunk_size=2000, min_chunk_size=500, min_coherence_threshold=0.3)

Get detailed information about each chunk including coherence scores and topics.

Parameters:

  • text (str): The input text to be chunked
  • max_chunk_size (int): Maximum size for each chunk (default: 2000)
  • min_chunk_size (int): Minimum size for each chunk (default: 500)
  • min_coherence_threshold (float): Minimum coherence threshold (default: 0.3)

Returns:

  • List[Dict[str, Any]]: List of dictionaries containing chunk details

Class

SemanticTextChunker

Advanced semantic text chunking class that preserves meaning and context.

Methods:

  • chunk_text_semantically(text, max_chunk_size=2000, min_chunk_size=500, min_coherence_threshold=0.3)
  • get_chunk_details(text, max_chunk_size=2000, min_chunk_size=500, min_coherence_threshold=0.3)

Development

Building from Source

# Clone the repository
git clone [<repository-url>](https://github.com/Lumen-Labs/cpp-chunker.git)
cd cpp_chunker

# Install dependencies
pip install pybind11

# Build using CMake
mkdir build && cd build
cmake ..
make

# Or build using pip
pip install .

Running Tests

# Run basic functionality test
python test_cunk.py

# Run memory safety tests
python test_memory_safety.py

License

This project is licensed under the MIT License - see the LICENSE file for details.

About

Implementation of a fast semantic chunker in C++, installable in python 3.7+ projects.

Resources

Stars

Watchers

Forks

Packages

No packages published