Industry-Leading Chain Coverage
We have indexed and organized all of the data for every chain, so you don’t have to. Schemas are normalized to provide an efficient querying experience across chains and tables. Tables include blocks, logs, decoded logs, traces, state diffs, priced transfers, and other related data. Furthermore, our tables are partitioned for optimal scanning and querying across these large datasets.Technical Specs
Quality:
Highest quality (P90) re-org aware Historical Data
Freshness:
Minutes from tip (depending on each specific chain, this is the minimum amount of time necessary to achieve P90 re-org aware quality for the chain).
Delivery Method:
Available via APIs (batch delivery only), Snowflake, Databricks, BigQuery, Parquet / CSV/ Iceberg files, and replication on AWS, GCP, or Azure.
Coverage:
Complete historical streaming data across 120+ chains and all protocols, and counting.
Data History:
From the genesis block to the latest finalized block across all 120+ chains that we cover.
Benefits of SonarX’s Full Historical Stream
Comprehensive Historical Coverage
Access the full blockchain history, from genesis to tip, for in-depth and comprehensive analysis.
Trusted Quality Controls
Automated checks ensure industry-leading data consistency and reliability across large datasets.
Flexible Delivery Options
Access data in various formats and across major cloud providers, tailored to your infrastructure.
Seamless Integration
Effortlessly query massive historical datasets using SQL with no need for additional storage.
24/7 Automated Data Quality Controls
The highest quality level in the Industry (P90), fully re-org aware. SonarX’s Data Quality process is designed to perform rigorous checks, addressing various aspects to guarantee the reliability and completeness of the data we provide. These are examples (not an exhaustive list) of the checks we automatically perform:1. Reorg Handling: Correctly processes blockchain reorganizations. 2. Dupes Check: Ensures no duplicate transactions or blocks. 3. Gaps Check: Ensures there are no missing data sequences. 4. Check Sums: Verifies data integrity during processing. 5. Table Flow Validation: Ensures consistency between raw and processed data. 6. Null Data Handling: Identifies and corrects missing data fields. 7. Match Checks: Verifies consistency between block and transaction hashes.
Access to sample data, Data Catalog, Schemas, and Data Dictionaries
Because our curation and enhancement processes are proprietary, we don’t share public samples or our complete Data Catalog. However, if you’d like to explore our datasets in more detail, you can easily request access to SonarX’s Console through the form below — no obligations, no sales pressure.