
At its core, data indexing is the process of transforming raw data into a format that's optimized for retrieval. Unlike an arbitrary application that may generate new source-of-truth data, indexing pipelines process existing data in various ways while maintaining traceability back to the original source. This intrinsic nature - being a derivative rather than a source of truth - creates unique challenges and requirements.
A well-designed indexing pipeline should possess several key traits:
People should be able to build a new indexing pipeline without mastering techniques such as database manipulation/access, streaming processing, parallelization, fault recovery, etc. In addition, transformation components (a.k.a. operations) should be easily composable and reusable across different pipelines.
The pipeline should be easy to understand, modify, and debug. Complex transformation logic should be manageable without becoming a maintenance burden.
On the other hand, the indexing pipeline is a stateful system, so besides the transformation logic, it's also important to expose the clear status of the pipeline states, e.g., statistics of the number of data entries, their freshness, and how a specific piece of derived data tracks back to the original source.
Data transformation (with the necessary tracking of relationships between data) should be done efficiently without excessive computational or storage costs. Moreover, existing computations should be reused whenever possible. For example, 1% of document change, or a chunking strategy change that only affects 1% of chunks, shouldn't need to entail rerunning the expensive embedding model over the entire dataset.
For many applications, the source of truth for indexing is consistently updated, so it's important to make sure the indexing pipeline is also updated accordingly in a timely manner.
The ability to process only new or changed data rather than reprocessing everything is crucial for both cost efficiency and indexing freshness. This becomes especially important as your data grows.
To make incremental update work, we need to carefully track the state of the pipeline, to decide which portion of the data needs to be reprocessed, and make sure states derived from old versions are fully deleted or replaced. It's challenging to make things right while considering various complexities, like fan-in/fan-out in transformations, out-of-order processing, recovery after early termination, etc.
Many implementations focus on the initial setup but neglect how the pipeline will evolve. When requirements change or new processing steps need to be added, the system should adapt without requiring a complete rebuild.
Traditional pipeline implementations often struggle with changes to the processing steps. Adding or modifying steps typically requires reprocessing all data, which can be extremely expensive and involves a manual process.
Many systems require deterministic processing logic - meaning the same input should always produce the same output. This becomes problematic when:
CocoIndex approaches indexing pipelines with a fundamentally different mental model - similar to how React revolutionized UI development compared to vanilla JavaScript. Instead of focusing on the mechanics of data processing, users can concentrate on their business logic and desired state.
Just as React changed how developers think about UI updates by introducing the concept of declarative rendering, CocoIndex changes how we think about data indexing. Instead of writing imperative processing logic, users declare their desired transformations and let CocoIndex handle the complexities of efficient execution.
This shift allows developers to focus on what their data should look like rather than the mechanics of how to get it there. The result is more maintainable, efficient, and reliable indexing pipelines that can evolve with your application's needs.
A well-designed indexing pipeline is crucial for production RAG applications, but building one that's maintainable, efficient, and evolvable is challenging. CocoIndex provides a framework that handles these complexities while allowing developers to focus on their core business logic.
By learning from the challenges faced by traditional approaches, we've created a system that makes robust data indexing accessible to everyone building RAG applications.
It would mean a lot to us if you could support CocoIndex on Github (https://github.com/cocoindex-io/cocoindex) with a star if you like our work. Thank you so much with a warm coconut hug 🥥🤗.