Tiders is an open-source framework that simplifies getting data out of blockchains and into your favorite tools. Whether you are building a DeFi dashboard, tracking NFT transfers, or running complex analytics, Tiders handles the heavy lifting of fetching, cleaning, transforming and storing blockchain data.
Tiders is modular. A Tiders pipeline is built from four components:
| Component | Description |
|---|---|
Provider |
Data source (HyperSync, SQD, or RPC) |
Query |
What data to fetch (block range, transaction, logs, filters, field selection) |
Steps |
Transformations to apply (decode, cast, encode, custom) |
Writer |
Output destination |
Most indexers lock you into a specific platform or database. Tiders is built to be modular, meaning you can swap parts in and out without breaking your setup:
- Swap Providers: Don't like your current data source? Switch between HyperSync, SQD, or a standard RPC node by changing one line of code.
- Plug-and-Play data transformations: Need to decode smart contract events or change data types? Use our built-in Rust-powered steps or write your own custom logic.
- Write Anywhere: Send your data to a local DuckDB file for prototyping, or a production-grade ClickHouse or PostgreSQL instance when you're ready to scale.
- Modular Reusable Pipelines: Protocols often reuse the same data structures. You don't need write modules from scratch every time. Since Tiders pipelines are regular Python objects, you can build functions around them, reuse across pipelines, or set input parameters to customize as needed.
See getting started section of the docs.
| Mode | How | When to use |
|---|---|---|
| Python SDK | Write a Python script, import tiders |
Full control, custom logic, complex pipelines |
| CLI (No-Code) | Write a YAML config, run tiders start |
Quick setup, no Python required, standard pipelines |
Both modes share the same pipeline engine.
You can also use tiders codegen to generate a Python script from a YAML config — a quick way to move from no-code to full Python control.
For both CLI and SDK
pip install tiders- Continuous Ingestion: Keep your datasets live and fresh. Tiders can poll the chain head to ensure your data is always up to date.
- Switch Providers: Move between HyperSync, SQD, or standard RPC nodes with a single config change.
- No Vendor Lock-in: Use the best data providers in the industry without being tied to their specific platforms or database formats.
- Custom Logic: Easily extend and customize your pipeline code in Python for complete flexibility.
- Advanced Analytics: Seamlessly works with industry-standard tools like Polars, Pandas, Data fusion and PyArrow as the data is fetched.
- Multiple Outputs: Send the same data to a local file and a production database simultaneously.
- Rust-Powered Speed: Core tasks like decoding and transforming data are handled in Rust, giving you massive performance without needing to learn a low-level language.
- Parallel Execution: Tiders doesn't wait around. While it's writing the last batch of data to your database, it’s already fetching and processing the next one in the background.
Connect to the best data sources in the industry without vendor lock-in. Tiders decouples the provider from the destination, giving you a consistent way to fetch data.
| Provider | Ethereum (EVM) | Solana (SVM) |
|---|---|---|
| HyperSync | ✅ | ❌ |
| SQD | ✅ | ✅ |
| RPC | ✅ | ❌ |
Tiders can support new providers. If your project has custom APIs to fetch blockchain data, especially ones that support server-side filtering, you can create a client for it, similar to the Tiders RPC client. Get in touch with us.
Leverage the tools you already know. Tiders automatically convert data batch-by-batch into your engine's native format, allowing for seamless, custom transformations on every incoming increment immediately before it is written.
| Engine | Data format in your function | Best for |
|---|---|---|
| Polars | Dict[str, pl.DataFrame] |
Fast columnar operations, expressive API |
| Pandas | Dict[str, pd.DataFrame] |
Familiar API, complex row-level operations |
| DataFusion | Dict[str, datafusion.DataFrame] |
SQL-based transformations, lazy evaluation |
| PyArrow | Dict[str, pa.Table] |
Zero-copy, direct Arrow manipulation |
Whether local or a production-grade data lake, Tiders handles the schema mapping and batch-loading to your destination of choice.
| Destination | Type | Description |
|---|---|---|
| DuckDB | Database | Embedded analytical database, great for local exploration and prototyping |
| ClickHouse | Database | Column-oriented database optimized for real-time analytical queries |
| PostgreSQL | Database | General-purpose relational database with broad ecosystem support |
| Apache Iceberg | Table Format | Open table format for large-scale analytics on data lakes |
| Delta Lake | Table Format | Storage layer with ACID transactions for data lakes |
| Parquet | File | Columnar file format, efficient for analytical workloads |
| CSV | File | Plain-text format, widely compatible and easy to inspect |
| Example | Chain | Provider | Decoding | Writer |
|---|---|---|---|---|
| rETH Transfer (no code) | Ethereum (EVM) | HyperSync | EVM event decode | Parquet |
| Jupiter Swaps | Solana (SVM) | SQD | SVM instruction decode | DuckDB |
| Uniswap V3 | Ethereum (EVM) | HyperSync / SQD / RPC | EVM event decode (factory + children) | DuckDB / Parquet / Delta Lake / ClickHouse / Iceberg |
- rETH Transfer — Simplest starting point. Uses a YAML config with no Python code to index a single event from a single contract.
- Jupiter Swaps — Uses the Python SDK on Solana. Shows instruction decoding and custom Polars steps.
- Uniswap V3 — Demonstrates the factory + children two-stage indexing pattern, chaining two pipelines where the first discovers contracts and the second indexes their events.
Browse all examples in the examples/ directory.
Python code uses the standard logging module of python, so it can be configured according to python docs.
Set RUST_LOG environment variable according to env_logger docs in order to see logs from rust modules.
To run an example with trace level logging for rust modules:
RUST_LOG=trace uv run examples/path/to/my/example
Tiders is composed of some repositories. 3 owned ones.
This repo uses uv for development. Clone all three projects side by side:
git clone https://github.com/yulesa/tiders.git
git clone https://github.com/yulesa/tiders-core.git
git clone https://github.com/yulesa/tiders-rpc-client.gitConfigure tiders to use your local tiders-core Python package by adding to pyproject.toml:
[tool.uv.sources]
tiders-core = { path = "../tiders-core/python", editable = true }Then sync the environment:
cd tiders
uv syncFor full instructions including building tiders-core and tiders-rpc-client from source, see the Development Setup docs.
Core libraries we use for ingesting/decoding/validating/transforming blockchain data are implemented in tiders-core repo.
Tiders is a fork of Cherry and cherry-core, a blockchain data pipeline framework built by the SteelCake team. Cherry laid the architectural foundation that Tiders builds upon, and we're grateful for their work and the open-source spirit that made this continuation possible.
Licensed under either of
- Apache License, Version 2.0 (LICENSE-APACHE or http://www.apache.org/licenses/LICENSE-2.0)
- MIT license (LICENSE-MIT or http://opensource.org/licenses/MIT)
at your option.
Unless you explicitly state otherwise, any contribution intentionally submitted for inclusion in the work by you, as defined in the Apache-2.0 license, shall be dual licensed as above, without any additional terms or conditions.


