Split large text documents into smaller, manageable chunks using different chunking strategies optimized for RAG (Retrieval-Augmented Generation) workflows. This endpoint supports multiple chunking algorithms including token-based, sentence-based, recursive, semantic, and specialized strategies.
Bearer authentication header of the form Bearer <token>, where <token> is your auth token.
Splits text based on token count. Best for ensuring chunks fit within LLM context windows and maintaining consistent chunk sizes for embedding models.
The text content to be chunked
token Whether to include metadata for each chunk
Return format: chunks (with metadata) or texts (plain strings)
chunks, texts Maximum tokens per chunk
Number of tokens to overlap between chunks
x >= 0Text successfully chunked