Skip to main content
SonarX Complete Historical Streaming offers comprehensive blockchain records, spanning from genesis to near real-time, that are fully indexed and easily accessible via SQL. Designed for institutions, enterprises, and other sophisticated data consumers, our historical datasets encompass over 120 chains and feature rigorous data quality controls to ensure completeness, consistency, and accuracy. These datasets are available via Snowflake, BigQuery, Databricks, and File drops in multiple formats (CSV, Parquet, Iceberg, etc.) and into all clouds (S3, GCS, ABS, etc.).

Industry-Leading Chain Coverage

We have indexed and organized all of the data for every chain, so you don’t have to. Schemas are normalized to provide an efficient querying experience across chains and tables. Tables include blocks, logs, decoded logs, traces, state diffs, priced transfers, and other related data. Furthermore, our tables are partitioned for optimal scanning and querying across these large datasets.

Technical Specs

Quality:

Highest quality (P90) re-org aware Historical Data

Freshness:

Minutes from tip (depending on each specific chain, this is the minimum amount of time necessary to achieve P90 re-org aware quality for the chain).

Delivery Method:

Available via APIs (batch delivery only), Snowflake, Databricks, BigQuery, Parquet / CSV/ Iceberg files, and replication on AWS, GCP, or Azure.

Coverage:

Complete historical streaming data across 120+ chains and all protocols, and counting.

Data History:

From the genesis block to the latest finalized block across all 120+ chains that we cover.

Benefits of SonarX’s Full Historical Stream

Comprehensive Historical Coverage

Access the full blockchain history, from genesis to tip, for in-depth and comprehensive analysis.

Trusted Quality Controls

Automated checks ensure industry-leading data consistency and reliability across large datasets.

Flexible Delivery Options

Access data in various formats and across major cloud providers, tailored to your infrastructure.

Seamless Integration

Effortlessly query massive historical datasets using SQL with no need for additional storage.

24/7 Automated Data Quality Controls

The highest quality level in the Industry (P90), fully re-org aware. SonarX’s Data Quality process is designed to perform rigorous checks, addressing various aspects to guarantee the reliability and completeness of the data we provide. These are examples (not an exhaustive list) of the checks we automatically perform:
1.      Reorg Handling: Correctly processes blockchain reorganizations.
2.      Dupes Check: Ensures no duplicate transactions or blocks. 3.      Gaps Check: Ensures there are no missing data sequences. 4.      Check Sums: Verifies data integrity during processing. 5.      Table Flow Validation: Ensures consistency between raw and processed data. 6.      Null Data Handling: Identifies and corrects missing data fields. 7.      Match Checks: Verifies consistency between block and transaction hashes.
For more details on SonarX’s best-in-class quality process, visit our Data Quality page.

Access to sample data, Data Catalog, Schemas, and Data Dictionaries

Because our curation and enhancement processes are proprietary, we don’t share public samples or our complete Data Catalog. However, if you’d like to explore our datasets in more detail, you can easily request access to SonarX’s Console through the form below — no obligations, no sales pressure. Request Console Access Form Pn