Skip to main content
Schema: sui.price Table: ez_asset_metadata Type: Base Table A convenience table holding prioritized asset metadata and other relevant details pertaining to each token_address and native asset. This data set is highly curated and contains metadata for one unique asset per blockchain.

Columns

Column NameData TypeDescription
TOKEN_ADDRESSTEXTThe specific address representing the asset on a specific platform. This will be NULL if referring to a native asset.
ASSET_IDTEXTThe unique identifier representing the asset.
SYMBOLTEXTThe symbol of asset.
NAMETEXTThe name of asset.
DECIMALSNUMBERThe number of decimals for the asset. May be NULL.
BLOCKCHAINTEXTThe Blockchain, Network, or Platform for this asset.
IS_NATIVEBOOLEANA flag indicating assets native to the respective blockchain.
IS_DEPRECATEDBOOLEANA flag indicating if the asset is deprecated or no longer supported by the provider.
TOKEN_IS_VERIFIEDBOOLEANBoolean flag indicating whether the token or price record is verified by Flipside’s crosschain curation process. Verified tokens are prioritized for analytics and are considered reliable for most use cases. Unverified tokens may be incomplete, deprecated, or experimental.
EZ_ASSET_METADATA_IDTEXTUnique identifier for the token metadata record, linking metadata to on-chain token types. Used for metadata management, registry operations, and analytics joins. Example: ‘tokenmeta_123’.
INSERTED_TIMESTAMPTIMESTAMP_NTZTimestamp when the record was inserted into the analytics database. System-generated by the ETL pipeline, typically in TIMESTAMP_NTZ format. Used for data lineage, ETL monitoring, and freshness checks. In Sui analytics, this field is essential for tracking data ingestion latency, validating pipeline health, and supporting incremental model builds. Example: ‘2024-06-01 12:34:56.789’.
MODIFIED_TIMESTAMPTIMESTAMP_NTZTimestamp when the record was last modified in the analytics database. System-generated for change tracking, data versioning, and consistency verification. In Sui, this supports incremental processing, late-arriving data correction, and auditability of analytics workflows. Used to monitor data staleness and trigger downstream updates. Example: ‘2024-06-01 12:34:56.789’.