Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

refactor(indexer): Derive dynamic field info #5073

Merged
merged 4 commits into from
Feb 5, 2025
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
208 changes: 116 additions & 92 deletions crates/iota-indexer/src/indexer_reader.rs
Original file line number Diff line number Diff line change
Expand Up @@ -2,10 +2,7 @@
// Modifications Copyright (c) 2024 IOTA Stiftung
// SPDX-License-Identifier: Apache-2.0

use std::{
collections::HashMap,
sync::{Arc, Mutex},
};
use std::sync::{Arc, Mutex};

use anyhow::{Result, anyhow};
use cached::{Cached, SizedCache};
Expand All @@ -15,20 +12,22 @@ use diesel::{
r2d2::ConnectionManager, sql_types::Bool,
};
use fastcrypto::encoding::{Encoding, Hex};
use iota_json::MoveTypeLayout;
use iota_json_rpc_types::{
AddressMetrics, Balance, CheckpointId, Coin as IotaCoin, DisplayFieldsResponse, EpochInfo,
EventFilter, IotaCoinMetadata, IotaEvent, IotaObjectDataFilter, IotaTransactionBlockEffects,
IotaTransactionBlockEffectsAPI, IotaTransactionBlockResponse, MoveCallMetrics,
MoveFunctionName, NetworkMetrics, TransactionFilter,
EventFilter, IotaCoinMetadata, IotaEvent, IotaMoveValue, IotaObjectDataFilter,
IotaTransactionBlockEffects, IotaTransactionBlockEffectsAPI, IotaTransactionBlockResponse,
MoveCallMetrics, MoveFunctionName, NetworkMetrics, TransactionFilter,
};
use iota_package_resolver::{Package, PackageStore, PackageStoreWithLruCache, Resolver};
use iota_types::{
TypeTag,
balance::Supply,
base_types::{IotaAddress, ObjectID, ObjectRef, SequenceNumber, VersionNumber},
base_types::{IotaAddress, ObjectID, VersionNumber},
coin::{CoinMetadata, TreasuryCap},
committee::EpochId,
digests::{ObjectDigest, TransactionDigest},
dynamic_field::{DynamicFieldInfo, DynamicFieldName},
digests::TransactionDigest,
dynamic_field::{DynamicFieldInfo, DynamicFieldName, DynamicFieldType},
effects::TransactionEvents,
event::EventID,
iota_system_state::{
Expand All @@ -38,7 +37,7 @@ use iota_types::{
is_system_package,
object::{Object, ObjectRead},
};
use itertools::{Itertools, any};
use itertools::Itertools;
use move_core_types::{annotated_value::MoveStructLayout, language_storage::StructTag};
use tap::TapFallible;

Expand All @@ -53,7 +52,7 @@ use crate::{
events::StoredEvent,
move_call_metrics::QueriedMoveCallMetrics,
network_metrics::StoredNetworkMetrics,
objects::{CoinBalance, ObjectRefColumn, StoredObject},
objects::{CoinBalance, StoredObject},
transactions::{
StoredTransaction, StoredTransactionEvents, stored_events_to_events,
tx_events_to_iota_tx_events,
Expand Down Expand Up @@ -1290,61 +1289,37 @@ impl IndexerReader {
cursor: Option<ObjectID>,
limit: usize,
) -> Result<Vec<DynamicFieldInfo>, IndexerError> {
let objects = self
let stored_objects = self
.spawn_blocking(move |this| {
this.get_dynamic_fields_raw(parent_object_id, cursor, limit)
})
.await?;

if any(objects.iter(), |o| o.df_object_id.is_none()) {
return Err(IndexerError::PersistentStorageDataCorruption(format!(
"Dynamic field has empty df_object_id column for parent object {}",
parent_object_id
)));
}

// for Dynamic field objects, df_object_id != object_id, we need another look up
// to get the version and digests.
// TODO: simply store df_object_version and df_object_digest as well?
let dfo_ids = objects
.iter()
.filter_map(|o| {
// Unwrap safe: checked nullity above
if o.df_object_id.as_ref().unwrap() == &o.object_id {
None
} else {
Some(o.df_object_id.clone().unwrap())
}
})
.collect::<Vec<_>>();

let object_refs = self
.spawn_blocking(move |this| this.get_object_refs(dfo_ids))
.await?;
let mut df_futures = vec![];
for object in objects {
let package_resolver_clone = self.package_resolver.clone();
df_futures.push(tokio::task::spawn(
object.try_into_expectant_dynamic_field_info(package_resolver_clone),
));
let indexer_reader_arc = Arc::new(self.clone());
for stored_object in stored_objects {
let indexer_reader_arc_clone = Arc::clone(&indexer_reader_arc);
df_futures.push(tokio::task::spawn(async move {
indexer_reader_arc_clone
.try_create_dynamic_field_info(stored_object)
.await
}));
}
let mut dynamic_fields = futures::future::join_all(df_futures)
let df_infos = futures::future::try_join_all(df_futures)
.await
.into_iter()
.collect::<Result<Vec<_>, _>>()
.tap_err(|e| tracing::error!("Error joining DF futures: {:?}", e))?
.into_iter()
.collect::<Result<Vec<_>, _>>()
.tap_err(|e| tracing::error!("Error calling DF try_into function: {:?}", e))?;

for df in dynamic_fields.iter_mut() {
if let Some(obj_ref) = object_refs.get(&df.object_id) {
df.version = obj_ref.1;
df.digest = obj_ref.2;
}
}

Ok(dynamic_fields)
.tap_err(|e| {
tracing::error!(
"Error calling DF try_create_dynamic_field_info function: {:?}",
e
)
})?
.into_iter()
.flatten()
.collect::<Vec<_>>();
Ok(df_infos)
}

pub async fn get_dynamic_fields_raw_in_blocking_task(
Expand Down Expand Up @@ -1381,6 +1356,91 @@ impl IndexerReader {
Ok(objects)
}

async fn try_create_dynamic_field_info(
&self,
stored_object: StoredObject,
) -> Result<Option<DynamicFieldInfo>, IndexerError> {
if stored_object.df_kind.is_none() {
return Ok(None);
}

let object: Object = stored_object.try_into()?;
let Some(move_object) = object.data.try_as_move().cloned() else {
return Err(IndexerError::ResolveMoveStruct(
"Object is not a MoveObject".to_string(),
));
};
let struct_tag: StructTag = move_object.type_().clone().into();
let move_type_layout = self
.package_resolver
.type_layout(TypeTag::Struct(Box::new(struct_tag.clone())))
.await
.map_err(|e| {
IndexerError::ResolveMoveStruct(format!(
"Failed to get type layout for type {}: {}",
struct_tag, e
))
})?;
let MoveTypeLayout::Struct(move_struct_layout) = move_type_layout else {
return Err(IndexerError::ResolveMoveStruct(
"MoveTypeLayout is not Struct".to_string(),
));
};

let move_struct = move_object.to_move_struct(&move_struct_layout)?;
let (move_value, type_, object_id) =
DynamicFieldInfo::parse_move_object(&move_struct).tap_err(|e| tracing::warn!("{e}"))?;
let name_type = move_object.type_().try_extract_field_name(&type_)?;
let bcs_name = bcs::to_bytes(&move_value.clone().undecorate()).map_err(|e| {
IndexerError::Serde(format!(
"Failed to serialize dynamic field name {:?}: {e}",
move_value
))
})?;
let name = DynamicFieldName {
type_: name_type,
value: IotaMoveValue::from(move_value).to_json_value(),
};

Ok(Some(match type_ {
DynamicFieldType::DynamicObject => {
let object = self.get_object_in_blocking_task(object_id).await?.ok_or(
IndexerError::Uncategorized(anyhow::anyhow!(
"Failed to find object_id {:?} when trying to create dynamic field info",
object_id
)),
)?;

let version = object.version();
let digest = object.digest();
let object_type = object
.data
.type_()
.expect("Data represents a Move object and therefore should have type")
.clone();
DynamicFieldInfo {
name,
bcs_name,
type_,
object_type: object_type.to_canonical_string(/* with_prefix */ true),
object_id,
version,
digest,
}
}
DynamicFieldType::DynamicField => DynamicFieldInfo {
name,
bcs_name,
type_,
object_type: move_object.into_type().into_type_params()[1]
.to_canonical_string(/* with_prefix */ true),
object_id: object.id(),
version: object.version(),
digest: object.digest(),
},
}))
}

pub async fn bcs_name_from_dynamic_field_name(
&self,
name: &DynamicFieldName,
Expand All @@ -1400,42 +1460,6 @@ impl IndexerReader {
Ok(name_bcs_value)
}

fn get_object_refs(
&self,
object_ids: Vec<Vec<u8>>,
) -> IndexerResult<HashMap<ObjectID, ObjectRef>> {
run_query!(&self.pool, |conn| {
let query = objects::dsl::objects
.select((
objects::dsl::object_id,
objects::dsl::object_version,
objects::dsl::object_digest,
))
.filter(objects::dsl::object_id.eq_any(object_ids))
.into_boxed();
query.load::<ObjectRefColumn>(conn)
})?
.into_iter()
.map(|object_ref: ObjectRefColumn| {
let object_id = ObjectID::from_bytes(object_ref.object_id.clone()).map_err(|_e| {
IndexerError::PersistentStorageDataCorruption(format!(
"Can't convert {:?} to ObjectID",
object_ref.object_id
))
})?;
let seq = SequenceNumber::from_u64(object_ref.object_version as u64);
let object_digest = ObjectDigest::try_from(object_ref.object_digest.as_slice())
.map_err(|e| {
IndexerError::PersistentStorageDataCorruption(format!(
"object {:?} has incompatible object digest. Error: {e}",
object_ref.object_digest
))
})?;
Ok((object_id, (object_id, seq, object_digest)))
})
.collect::<IndexerResult<HashMap<_, _>>>()
}

pub async fn get_display_object_by_type(
&self,
object_type: &move_core_types::language_storage::StructTag,
Expand Down
Loading
Loading