mod builtin_item_migration;
use std::collections::{BTreeMap, BTreeSet};
use std::sync::Arc;
use std::time::{Duration, Instant};
use futures::future::{BoxFuture, FutureExt};
use itertools::{Either, Itertools};
use mz_adapter_types::compaction::CompactionWindow;
use mz_adapter_types::dyncfgs::{ENABLE_CONTINUAL_TASK_BUILTINS, ENABLE_EXPRESSION_CACHE};
use mz_catalog::builtin::{
Builtin, BuiltinTable, Fingerprint, BUILTINS, BUILTIN_CLUSTERS, BUILTIN_CLUSTER_REPLICAS,
BUILTIN_PREFIXES, BUILTIN_ROLES, MZ_STORAGE_USAGE_BY_SHARD_DESCRIPTION,
RUNTIME_ALTERABLE_FINGERPRINT_SENTINEL,
};
use mz_catalog::config::{ClusterReplicaSizeMap, StateConfig};
use mz_catalog::durable::objects::{
SystemObjectDescription, SystemObjectMapping, SystemObjectUniqueIdentifier,
};
use mz_catalog::durable::{
ClusterVariant, ClusterVariantManaged, Transaction, SYSTEM_CLUSTER_ID_ALLOC_KEY,
};
use mz_catalog::expr_cache::{
ExpressionCacheConfig, ExpressionCacheHandle, GlobalExpressions, LocalExpressions,
};
use mz_catalog::memory::error::{Error, ErrorKind};
use mz_catalog::memory::objects::{
BootstrapStateUpdateKind, CatalogEntry, CatalogItem, CommentsMap, DefaultPrivileges,
StateUpdate,
};
use mz_catalog::SYSTEM_CONN_ID;
use mz_compute_client::logging::LogVariant;
use mz_controller::clusters::{ReplicaAllocation, ReplicaLogging};
use mz_controller_types::ClusterId;
use mz_ore::cast::usize_to_u64;
use mz_ore::collections::{CollectionExt, HashSet};
use mz_ore::now::to_datetime;
use mz_ore::{instrument, soft_assert_no_log};
use mz_repr::adt::mz_acl_item::PrivilegeMap;
use mz_repr::namespaces::is_unstable_schema;
use mz_repr::role_id::RoleId;
use mz_repr::{CatalogItemId, Diff, GlobalId, RelationVersion, Timestamp};
use mz_sql::catalog::{
BuiltinsConfig, CatalogError as SqlCatalogError, CatalogItem as SqlCatalogItem,
CatalogItemType, RoleMembership, RoleVars,
};
use mz_sql::func::OP_IMPLS;
use mz_sql::names::SchemaId;
use mz_sql::rbac;
use mz_sql::session::user::{MZ_SYSTEM_ROLE_ID, SYSTEM_USER};
use mz_sql::session::vars::{SessionVars, SystemVars, VarError, VarInput};
use mz_sql_parser::ast::display::AstDisplay;
use mz_storage_client::controller::StorageController;
use timely::Container;
use tracing::{error, info, warn, Instrument};
use uuid::Uuid;
use crate::catalog::open::builtin_item_migration::{
migrate_builtin_items, BuiltinItemMigrationResult,
};
use crate::catalog::state::LocalExpressionCache;
use crate::catalog::{
is_reserved_name, migrate, BuiltinTableUpdate, Catalog, CatalogPlans, CatalogState, Config,
};
use crate::AdapterError;
#[derive(Debug)]
pub struct BuiltinMigrationMetadata {
pub previous_storage_collection_ids: BTreeSet<CatalogItemId>,
pub migrated_system_object_mappings: BTreeMap<CatalogItemId, SystemObjectMapping>,
pub introspection_source_index_updates:
BTreeMap<ClusterId, Vec<(LogVariant, String, CatalogItemId, GlobalId, u32)>>,
pub user_item_drop_ops: Vec<CatalogItemId>,
pub user_item_create_ops: Vec<CreateOp>,
}
#[derive(Debug)]
pub struct CreateOp {
id: CatalogItemId,
oid: u32,
global_id: GlobalId,
schema_id: SchemaId,
name: String,
owner_id: RoleId,
privileges: PrivilegeMap,
item_rebuilder: CatalogItemRebuilder,
}
impl BuiltinMigrationMetadata {
fn new() -> BuiltinMigrationMetadata {
BuiltinMigrationMetadata {
previous_storage_collection_ids: BTreeSet::new(),
migrated_system_object_mappings: BTreeMap::new(),
introspection_source_index_updates: BTreeMap::new(),
user_item_drop_ops: Vec::new(),
user_item_create_ops: Vec::new(),
}
}
}
#[derive(Debug)]
pub enum CatalogItemRebuilder {
SystemSource(CatalogItem),
Object {
sql: String,
is_retained_metrics_object: bool,
custom_logical_compaction_window: Option<CompactionWindow>,
},
}
impl CatalogItemRebuilder {
fn new(
entry: &CatalogEntry,
id: CatalogItemId,
ancestor_ids: &BTreeMap<CatalogItemId, CatalogItemId>,
) -> Self {
if id.is_system()
&& (entry.is_table() || entry.is_introspection_source() || entry.is_source())
{
Self::SystemSource(entry.item().clone())
} else {
let create_sql = entry.create_sql().to_string();
let mut create_stmt = mz_sql::parse::parse(&create_sql)
.expect("invalid create sql persisted to catalog")
.into_element()
.ast;
mz_sql::ast::transform::create_stmt_replace_ids(&mut create_stmt, ancestor_ids);
Self::Object {
sql: create_stmt.to_ast_string_stable(),
is_retained_metrics_object: entry.item().is_retained_metrics_object(),
custom_logical_compaction_window: entry.item().custom_logical_compaction_window(),
}
}
}
fn build(
self,
global_id: GlobalId,
state: &CatalogState,
versions: &BTreeMap<RelationVersion, GlobalId>,
) -> CatalogItem {
match self {
Self::SystemSource(item) => item,
Self::Object {
sql,
is_retained_metrics_object,
custom_logical_compaction_window,
} => state
.parse_item(
global_id,
&sql,
versions,
None,
is_retained_metrics_object,
custom_logical_compaction_window,
&mut LocalExpressionCache::Closed,
)
.unwrap_or_else(|error| panic!("invalid persisted create sql ({error:?}): {sql}")),
}
}
}
pub struct InitializeStateResult {
pub state: CatalogState,
pub storage_collections_to_drop: BTreeSet<CatalogItemId>,
pub migrated_storage_collections_0dt: BTreeSet<CatalogItemId>,
pub new_builtin_collections: BTreeSet<GlobalId>,
pub builtin_table_updates: Vec<BuiltinTableUpdate>,
pub last_seen_version: String,
pub expr_cache_handle: Option<ExpressionCacheHandle>,
pub cached_global_exprs: BTreeMap<GlobalId, GlobalExpressions>,
pub uncached_local_exprs: BTreeMap<GlobalId, LocalExpressions>,
}
pub struct OpenCatalogResult {
pub catalog: Catalog,
pub storage_collections_to_drop: BTreeSet<CatalogItemId>,
pub migrated_storage_collections_0dt: BTreeSet<CatalogItemId>,
pub new_builtin_collections: BTreeSet<GlobalId>,
pub builtin_table_updates: Vec<BuiltinTableUpdate>,
pub cached_global_exprs: BTreeMap<GlobalId, GlobalExpressions>,
pub uncached_local_exprs: BTreeMap<GlobalId, LocalExpressions>,
}
impl Catalog {
pub async fn initialize_state<'a>(
config: StateConfig,
storage: &'a mut Box<dyn mz_catalog::durable::DurableCatalogState>,
) -> Result<InitializeStateResult, AdapterError> {
for builtin_role in BUILTIN_ROLES {
assert!(
is_reserved_name(builtin_role.name),
"builtin role {builtin_role:?} must start with one of the following prefixes {}",
BUILTIN_PREFIXES.join(", ")
);
}
for builtin_cluster in BUILTIN_CLUSTERS {
assert!(
is_reserved_name(builtin_cluster.name),
"builtin cluster {builtin_cluster:?} must start with one of the following prefixes {}",
BUILTIN_PREFIXES.join(", ")
);
}
let mut state = CatalogState {
database_by_name: BTreeMap::new(),
database_by_id: BTreeMap::new(),
entry_by_id: BTreeMap::new(),
entry_by_global_id: BTreeMap::new(),
ambient_schemas_by_name: BTreeMap::new(),
ambient_schemas_by_id: BTreeMap::new(),
clusters_by_name: BTreeMap::new(),
clusters_by_id: BTreeMap::new(),
roles_by_name: BTreeMap::new(),
roles_by_id: BTreeMap::new(),
network_policies_by_id: BTreeMap::new(),
network_policies_by_name: BTreeMap::new(),
system_configuration: {
let mut s =
SystemVars::new(config.active_connection_count).set_unsafe(config.unsafe_mode);
if config.all_features {
s.enable_all_feature_flags_by_default();
}
s
},
default_privileges: DefaultPrivileges::default(),
system_privileges: PrivilegeMap::default(),
comments: CommentsMap::default(),
source_references: BTreeMap::new(),
storage_metadata: Default::default(),
temporary_schemas: BTreeMap::new(),
config: mz_sql::catalog::CatalogConfig {
start_time: to_datetime((config.now)()),
start_instant: Instant::now(),
nonce: rand::random(),
environment_id: config.environment_id,
session_id: Uuid::new_v4(),
build_info: config.build_info,
timestamp_interval: Duration::from_secs(1),
now: config.now.clone(),
connection_context: config.connection_context,
builtins_cfg: BuiltinsConfig {
include_continual_tasks: get_dyncfg_val_from_defaults_and_remote(
&config.system_parameter_defaults,
config.remote_system_parameters.as_ref(),
&ENABLE_CONTINUAL_TASK_BUILTINS,
),
},
helm_chart_version: config.helm_chart_version,
},
cluster_replica_sizes: config.cluster_replica_sizes,
availability_zones: config.availability_zones,
egress_addresses: config.egress_addresses,
aws_principal_context: config.aws_principal_context,
aws_privatelink_availability_zones: config.aws_privatelink_availability_zones,
http_host_name: config.http_host_name,
};
let deploy_generation = storage.get_deployment_generation().await?;
let mut updates: Vec<_> = storage.sync_to_current_updates().await?;
assert!(!updates.is_empty(), "initial catalog snapshot is missing");
let mut txn = storage.transaction().await?;
let (migrated_builtins, new_builtin_collections) = {
migrate::durable_migrate(&mut txn, config.boot_ts)?;
if let Some(remote_system_parameters) = config.remote_system_parameters {
for (name, value) in remote_system_parameters {
txn.upsert_system_config(&name, value)?;
}
txn.set_system_config_synced_once()?;
}
let (migrated_builtins, new_builtin_collections) =
add_new_remove_old_builtin_items_migration(&state.config().builtins_cfg, &mut txn)?;
let cluster_sizes = BuiltinBootstrapClusterSizes {
system_cluster: config.builtin_system_cluster_replica_size,
catalog_server_cluster: config.builtin_catalog_server_cluster_replica_size,
probe_cluster: config.builtin_probe_cluster_replica_size,
support_cluster: config.builtin_support_cluster_replica_size,
analytics_cluster: config.builtin_analytics_cluster_replica_size,
};
add_new_builtin_clusters_migration(
&mut txn,
&cluster_sizes,
&state.cluster_replica_sizes,
)?;
add_new_remove_old_builtin_introspection_source_migration(&mut txn)?;
add_new_builtin_cluster_replicas_migration(
&mut txn,
&cluster_sizes,
&state.cluster_replica_sizes,
)?;
add_new_builtin_roles_migration(&mut txn)?;
remove_invalid_config_param_role_defaults_migration(&mut txn)?;
(migrated_builtins, new_builtin_collections)
};
remove_pending_cluster_replicas_migration(&mut txn)?;
let op_updates = txn.get_and_commit_op_updates();
updates.extend(op_updates);
{
for (name, value) in config.system_parameter_defaults {
match state.set_system_configuration_default(&name, VarInput::Flat(&value)) {
Ok(_) => (),
Err(Error {
kind: ErrorKind::VarError(VarError::UnknownParameter(name)),
}) => {
warn!(%name, "cannot load unknown system parameter from catalog storage to set default parameter");
}
Err(e) => return Err(e.into()),
};
}
state.create_temporary_schema(&SYSTEM_CONN_ID, MZ_SYSTEM_ROLE_ID)?;
}
let mut builtin_table_updates = Vec::new();
let commit_ts = txn.commit_ts();
let mut updates = into_consolidatable_updates_startup(updates, commit_ts);
differential_dataflow::consolidation::consolidate_updates(&mut updates);
soft_assert_no_log!(
updates.iter().all(|(_, _, diff)| *diff == 1),
"consolidated updates should be positive during startup: {updates:?}"
);
let mut pre_item_updates = Vec::new();
let mut system_item_updates = Vec::new();
let mut item_updates = Vec::new();
let mut post_item_updates = Vec::new();
for (kind, ts, diff) in updates {
match kind {
BootstrapStateUpdateKind::Role(_)
| BootstrapStateUpdateKind::Database(_)
| BootstrapStateUpdateKind::Schema(_)
| BootstrapStateUpdateKind::DefaultPrivilege(_)
| BootstrapStateUpdateKind::SystemPrivilege(_)
| BootstrapStateUpdateKind::SystemConfiguration(_)
| BootstrapStateUpdateKind::Cluster(_)
| BootstrapStateUpdateKind::NetworkPolicy(_)
| BootstrapStateUpdateKind::ClusterReplica(_) => {
pre_item_updates.push(StateUpdate {
kind: kind.into(),
ts,
diff: diff.try_into().expect("valid diff"),
})
}
BootstrapStateUpdateKind::IntrospectionSourceIndex(_)
| BootstrapStateUpdateKind::SystemObjectMapping(_) => {
system_item_updates.push(StateUpdate {
kind: kind.into(),
ts,
diff: diff.try_into().expect("valid diff"),
})
}
BootstrapStateUpdateKind::Item(_) => item_updates.push(StateUpdate {
kind: kind.into(),
ts,
diff: diff.try_into().expect("valid diff"),
}),
BootstrapStateUpdateKind::Comment(_)
| BootstrapStateUpdateKind::AuditLog(_)
| BootstrapStateUpdateKind::StorageCollectionMetadata(_)
| BootstrapStateUpdateKind::SourceReferences(_)
| BootstrapStateUpdateKind::UnfinalizedShard(_) => {
post_item_updates.push(StateUpdate {
kind: kind.into(),
ts,
diff: diff.try_into().expect("valid diff"),
})
}
}
}
let builtin_table_update = state
.apply_updates_for_bootstrap(pre_item_updates, &mut LocalExpressionCache::Closed)
.await;
builtin_table_updates.extend(builtin_table_update);
let expr_cache_start = Instant::now();
info!("startup: coordinator init: catalog open: expr cache open beginning");
let expr_cache_enabled = ENABLE_EXPRESSION_CACHE.get(state.system_config().dyncfgs());
let (expr_cache_handle, cached_local_exprs, cached_global_exprs) = if expr_cache_enabled {
info!("using expression cache for startup");
let current_ids = txn
.get_items()
.flat_map(|item| {
let gid = item.global_id.clone();
let gids: Vec<_> = item.extra_versions.values().cloned().collect();
std::iter::once(gid).chain(gids.into_iter())
})
.chain(
txn.get_system_object_mappings()
.map(|som| som.unique_identifier.global_id),
)
.collect();
let dyncfgs = config.persist_client.dyncfgs().clone();
let expr_cache_config = ExpressionCacheConfig {
deploy_generation,
persist: config.persist_client,
organization_id: state.config.environment_id.organization_id(),
current_ids,
remove_prior_gens: !config.read_only,
compact_shard: config.read_only,
dyncfgs,
};
let (expr_cache_handle, cached_local_exprs, cached_global_exprs) =
ExpressionCacheHandle::spawn_expression_cache(expr_cache_config).await;
(
Some(expr_cache_handle),
cached_local_exprs,
cached_global_exprs,
)
} else {
(None, BTreeMap::new(), BTreeMap::new())
};
let mut local_expr_cache = LocalExpressionCache::new(cached_local_exprs);
info!(
"startup: coordinator init: catalog open: expr cache open complete in {:?}",
expr_cache_start.elapsed()
);
let builtin_table_update = state
.apply_updates_for_bootstrap(system_item_updates, &mut local_expr_cache)
.await;
builtin_table_updates.extend(builtin_table_update);
let last_seen_version = txn
.get_catalog_content_version()
.unwrap_or_else(|| "new".to_string());
let builtin_table_update = if !config.skip_migrations {
migrate::migrate(
&mut state,
&mut txn,
&mut local_expr_cache,
item_updates,
config.now,
config.boot_ts,
)
.await
.map_err(|e| {
Error::new(ErrorKind::FailedMigration {
last_seen_version: last_seen_version.clone(),
this_version: config.build_info.version,
cause: e.to_string(),
})
})?
} else {
state
.apply_updates_for_bootstrap(item_updates, &mut local_expr_cache)
.await
};
builtin_table_updates.extend(builtin_table_update);
let builtin_table_update = state
.apply_updates_for_bootstrap(post_item_updates, &mut local_expr_cache)
.await;
builtin_table_updates.extend(builtin_table_update);
let BuiltinItemMigrationResult {
builtin_table_updates: builtin_table_update,
storage_collections_to_drop,
migrated_storage_collections_0dt,
cleanup_action,
} = migrate_builtin_items(
&mut state,
&mut txn,
&mut local_expr_cache,
migrated_builtins,
config.builtin_item_migration_config,
)
.await?;
builtin_table_updates.extend(builtin_table_update);
let builtin_table_updates = state.resolve_builtin_table_updates(builtin_table_updates);
txn.commit().await?;
cleanup_action.await;
Ok(InitializeStateResult {
state,
storage_collections_to_drop,
migrated_storage_collections_0dt,
new_builtin_collections: new_builtin_collections.into_iter().collect(),
builtin_table_updates,
last_seen_version,
expr_cache_handle,
cached_global_exprs,
uncached_local_exprs: local_expr_cache.into_uncached_exprs(),
})
}
#[instrument(name = "catalog::open")]
pub fn open(config: Config<'_>) -> BoxFuture<'static, Result<OpenCatalogResult, AdapterError>> {
async move {
let mut storage = config.storage;
let InitializeStateResult {
state,
storage_collections_to_drop,
migrated_storage_collections_0dt,
new_builtin_collections,
mut builtin_table_updates,
last_seen_version: _,
expr_cache_handle,
cached_global_exprs,
uncached_local_exprs,
} =
Self::initialize_state(config.state, &mut storage)
.instrument(tracing::info_span!("catalog::initialize_state"))
.boxed()
.await?;
let catalog = Catalog {
state,
plans: CatalogPlans::default(),
expr_cache_handle,
transient_revision: 1,
storage: Arc::new(tokio::sync::Mutex::new(storage)),
};
for (op, func) in OP_IMPLS.iter() {
match func {
mz_sql::func::Func::Scalar(impls) => {
for imp in impls {
builtin_table_updates.push(catalog.state.resolve_builtin_table_update(
catalog.state.pack_op_update(op, imp.details(), 1),
));
}
}
_ => unreachable!("all operators must be scalar functions"),
}
}
for ip in &catalog.state.egress_addresses {
builtin_table_updates.push(
catalog
.state
.resolve_builtin_table_update(catalog.state.pack_egress_ip_update(ip)?),
);
}
Ok(OpenCatalogResult {
catalog,
storage_collections_to_drop,
migrated_storage_collections_0dt,
new_builtin_collections,
builtin_table_updates,
cached_global_exprs,
uncached_local_exprs,
})
}
.instrument(tracing::info_span!("catalog::open"))
.boxed()
}
async fn initialize_storage_controller_state(
&mut self,
storage_controller: &mut dyn StorageController<Timestamp = mz_repr::Timestamp>,
storage_collections_to_drop: BTreeSet<CatalogItemId>,
) -> Result<(), mz_catalog::durable::CatalogError> {
let collections = self
.entries()
.filter(|entry| entry.item().is_storage_collection())
.flat_map(|entry| entry.global_ids())
.collect();
let collections_to_drop = storage_collections_to_drop
.into_iter()
.flat_map(|item_id| self.get_entry(&item_id).global_ids())
.collect();
let mut state = self.state.clone();
let mut storage = self.storage().await;
let mut txn = storage.transaction().await?;
storage_controller
.initialize_state(&mut txn, collections, collections_to_drop)
.await
.map_err(mz_catalog::durable::DurableCatalogError::from)?;
let updates = txn.get_and_commit_op_updates();
let builtin_updates = state.apply_updates(updates)?;
assert_eq!(builtin_updates, Vec::new());
txn.commit().await?;
drop(storage);
self.state = state;
Ok(())
}
pub async fn initialize_controller(
&mut self,
config: mz_controller::ControllerConfig,
envd_epoch: core::num::NonZeroI64,
read_only: bool,
storage_collections_to_drop: BTreeSet<CatalogItemId>,
) -> Result<mz_controller::Controller<mz_repr::Timestamp>, mz_catalog::durable::CatalogError>
{
let controller_start = Instant::now();
info!("startup: controller init: beginning");
let mut controller = {
let mut storage = self.storage().await;
let mut tx = storage.transaction().await?;
mz_controller::prepare_initialization(&mut tx)
.map_err(mz_catalog::durable::DurableCatalogError::from)?;
let updates = tx.get_and_commit_op_updates();
assert!(
updates.is_empty(),
"initializing controller should not produce updates: {updates:?}"
);
tx.commit().await?;
let read_only_tx = storage.transaction().await?;
mz_controller::Controller::new(config, envd_epoch, read_only, &read_only_tx).await
};
self.initialize_storage_controller_state(
&mut *controller.storage,
storage_collections_to_drop,
)
.await?;
info!(
"startup: controller init: complete in {:?}",
controller_start.elapsed()
);
Ok(controller)
}
fn generate_builtin_migration_metadata(
state: &CatalogState,
txn: &mut Transaction<'_>,
migrated_ids: Vec<CatalogItemId>,
id_fingerprint_map: BTreeMap<CatalogItemId, String>,
) -> Result<BuiltinMigrationMetadata, Error> {
let mut visited_set = BTreeSet::new();
let mut sorted_entries = Vec::new();
for item_id in migrated_ids {
if !visited_set.contains(&item_id) {
let migrated_topological_sort =
Catalog::topological_sort(state, item_id, &mut visited_set);
sorted_entries.extend(migrated_topological_sort);
}
}
sorted_entries.reverse();
let mut migration_metadata = BuiltinMigrationMetadata::new();
let mut ancestor_ids = BTreeMap::new();
let mut migrated_log_ids = BTreeMap::new();
let log_name_map: BTreeMap<_, _> = BUILTINS::logs()
.map(|log| (log.variant.clone(), log.name))
.collect();
for entry in sorted_entries {
let id = entry.id();
let (new_item_id, new_global_id) = match id {
CatalogItemId::System(_) => txn.allocate_system_item_ids(1)?.into_element(),
CatalogItemId::User(_) => txn.allocate_user_item_ids(1)?.into_element(),
_ => unreachable!("can't migrate id: {id}"),
};
let name = state.resolve_full_name(entry.name(), None);
info!("migrating {name} from {id} to {new_item_id}");
if let Some(fingerprint) = id_fingerprint_map.get(&id) {
assert!(
id.is_system(),
"id_fingerprint_map should only contain builtin objects"
);
let schema_name = state
.get_schema(
&entry.name().qualifiers.database_spec,
&entry.name().qualifiers.schema_spec,
entry.conn_id().unwrap_or(&SYSTEM_CONN_ID),
)
.name
.schema
.as_str();
migration_metadata.migrated_system_object_mappings.insert(
id,
SystemObjectMapping {
description: SystemObjectDescription {
schema_name: schema_name.to_string(),
object_type: entry.item_type(),
object_name: entry.name().item.clone(),
},
unique_identifier: SystemObjectUniqueIdentifier {
catalog_id: new_item_id,
global_id: new_global_id,
fingerprint: fingerprint.clone(),
},
},
);
}
ancestor_ids.insert(id, new_item_id);
if entry.item().is_storage_collection() {
migration_metadata
.previous_storage_collection_ids
.insert(id);
}
match entry.item() {
CatalogItem::Log(log) => {
migrated_log_ids.insert(log.global_id(), log.variant.clone());
}
CatalogItem::Index(index) => {
if id.is_system() {
if let Some(variant) = migrated_log_ids.get(&index.on) {
migration_metadata
.introspection_source_index_updates
.entry(index.cluster_id)
.or_default()
.push((
variant.clone(),
log_name_map
.get(variant)
.expect("all variants have a name")
.to_string(),
new_item_id,
new_global_id,
entry.oid(),
));
}
}
}
CatalogItem::Table(_)
| CatalogItem::Source(_)
| CatalogItem::MaterializedView(_)
| CatalogItem::ContinualTask(_) => {
}
CatalogItem::Sink(_) => {
}
CatalogItem::View(_) => {
}
CatalogItem::Type(_)
| CatalogItem::Func(_)
| CatalogItem::Secret(_)
| CatalogItem::Connection(_) => unreachable!(
"impossible to migrate schema for builtin {}",
entry.item().typ()
),
}
if id.is_user() {
migration_metadata.user_item_drop_ops.push(id);
}
let name = entry.name().clone();
if id.is_user() {
let schema_id = name.qualifiers.schema_spec.clone().into();
let item_rebuilder = CatalogItemRebuilder::new(entry, new_item_id, &ancestor_ids);
migration_metadata.user_item_create_ops.push(CreateOp {
id: new_item_id,
oid: entry.oid(),
global_id: new_global_id,
schema_id,
name: name.item.clone(),
owner_id: entry.owner_id().clone(),
privileges: entry.privileges().clone(),
item_rebuilder,
});
}
}
migration_metadata.user_item_drop_ops.reverse();
Ok(migration_metadata)
}
fn topological_sort<'a, 'b>(
state: &'a CatalogState,
id: CatalogItemId,
visited_set: &'b mut BTreeSet<CatalogItemId>,
) -> Vec<&'a CatalogEntry> {
let mut sorted_entries = Vec::new();
visited_set.insert(id);
let entry = state.get_entry(&id);
for dependant in entry.used_by() {
if !visited_set.contains(dependant) {
let child_topological_sort =
Catalog::topological_sort(state, *dependant, visited_set);
sorted_entries.extend(child_topological_sort);
}
}
sorted_entries.push(entry);
sorted_entries
}
#[mz_ore::instrument]
async fn apply_builtin_migration(
state: &mut CatalogState,
txn: &mut Transaction<'_>,
migration_metadata: &mut BuiltinMigrationMetadata,
) -> Result<Vec<BuiltinTableUpdate<&'static BuiltinTable>>, Error> {
for id in &migration_metadata.user_item_drop_ops {
let entry = state.get_entry(id);
if entry.is_sink() {
let full_name = state.resolve_full_name(entry.name(), None);
error!(
"user sink {full_name} will be recreated as part of a builtin migration which \
can result in duplicate data being emitted. This is a known issue, \
https://github.com/MaterializeInc/database-issues/issues/5553. Please inform the \
customer that their sink may produce duplicate data."
)
}
}
let mut builtin_table_updates = Vec::new();
txn.remove_items(&migration_metadata.user_item_drop_ops.drain(..).collect())?;
txn.update_system_object_mappings(std::mem::take(
&mut migration_metadata.migrated_system_object_mappings,
))?;
txn.update_introspection_source_index_gids(
std::mem::take(&mut migration_metadata.introspection_source_index_updates)
.into_iter()
.map(|(cluster_id, updates)| {
(
cluster_id,
updates
.into_iter()
.map(|(_variant, name, item_id, index_id, oid)| {
(name, item_id, index_id, oid)
}),
)
}),
)?;
let updates = txn.get_and_commit_op_updates();
let builtin_table_update = state
.apply_updates_for_bootstrap(updates, &mut LocalExpressionCache::Closed)
.await;
builtin_table_updates.extend(builtin_table_update);
for CreateOp {
id,
oid,
global_id,
schema_id,
name,
owner_id,
privileges,
item_rebuilder,
} in migration_metadata.user_item_create_ops.drain(..)
{
let versions = BTreeMap::new();
let item = item_rebuilder.build(global_id, state, &versions);
let (create_sql, expect_gid, expect_versions) = item.to_serialized();
assert_eq!(
global_id, expect_gid,
"serializing a CatalogItem changed the GlobalId"
);
assert_eq!(
versions, expect_versions,
"serializing a CatalogItem changed the Versions"
);
txn.insert_item(
id,
oid,
global_id,
schema_id,
&name,
create_sql,
owner_id.clone(),
privileges.all_values_owned().collect(),
versions,
)?;
let updates = txn.get_and_commit_op_updates();
let builtin_table_update = state
.apply_updates_for_bootstrap(updates, &mut LocalExpressionCache::Closed)
.await;
builtin_table_updates.extend(builtin_table_update);
}
Ok(builtin_table_updates)
}
pub async fn expire(self) {
if let Some(storage) = Arc::into_inner(self.storage) {
let storage = storage.into_inner();
storage.expire().await;
}
}
}
impl CatalogState {
fn set_system_configuration_default(
&mut self,
name: &str,
value: VarInput,
) -> Result<(), Error> {
Ok(self.system_configuration.set_default(name, value)?)
}
}
fn add_new_remove_old_builtin_items_migration(
builtins_cfg: &BuiltinsConfig,
txn: &mut mz_catalog::durable::Transaction<'_>,
) -> Result<(Vec<CatalogItemId>, Vec<GlobalId>), mz_catalog::durable::CatalogError> {
let mut new_builtin_mappings = Vec::new();
let mut migrated_builtin_ids = Vec::new();
let mut builtin_descs = HashSet::new();
let mut builtins = Vec::new();
for builtin in BUILTINS::iter(builtins_cfg) {
let desc = SystemObjectDescription {
schema_name: builtin.schema().to_string(),
object_type: builtin.catalog_item_type(),
object_name: builtin.name().to_string(),
};
if !builtin_descs.insert(desc.clone()) {
panic!(
"duplicate builtin description: {:?}, {:?}",
SystemObjectDescription {
schema_name: builtin.schema().to_string(),
object_type: builtin.catalog_item_type(),
object_name: builtin.name().to_string(),
},
builtin
);
}
builtins.push((desc, builtin));
}
let mut system_object_mappings: BTreeMap<_, _> = txn
.get_system_object_mappings()
.map(|system_object_mapping| {
(
system_object_mapping.description.clone(),
system_object_mapping,
)
})
.collect();
let (existing_builtins, new_builtins): (Vec<_>, Vec<_>) =
builtins.into_iter().partition_map(|(desc, builtin)| {
let fingerprint = match builtin.runtime_alterable() {
false => builtin.fingerprint(),
true => RUNTIME_ALTERABLE_FINGERPRINT_SENTINEL.into(),
};
match system_object_mappings.remove(&desc) {
Some(system_object_mapping) => {
Either::Left((builtin, system_object_mapping, fingerprint))
}
None => Either::Right((builtin, fingerprint)),
}
});
let new_builtin_ids = txn.allocate_system_item_ids(usize_to_u64(new_builtins.len()))?;
let new_builtins = new_builtins.into_iter().zip(new_builtin_ids.clone());
for (builtin, system_object_mapping, fingerprint) in existing_builtins {
if system_object_mapping.unique_identifier.fingerprint != fingerprint {
assert_ne!(
*MZ_STORAGE_USAGE_BY_SHARD_DESCRIPTION, system_object_mapping.description,
"mz_storage_usage_by_shard cannot be migrated or else the table will be truncated"
);
assert_ne!(
builtin.catalog_item_type(),
CatalogItemType::Type,
"types cannot be migrated"
);
assert_ne!(
system_object_mapping.unique_identifier.fingerprint,
RUNTIME_ALTERABLE_FINGERPRINT_SENTINEL,
"clearing the runtime alterable flag on an existing object is not permitted",
);
assert!(
!builtin.runtime_alterable(),
"setting the runtime alterable flag on an existing object is not permitted"
);
migrated_builtin_ids.push(system_object_mapping.unique_identifier.catalog_id);
}
}
for ((builtin, fingerprint), (catalog_id, global_id)) in new_builtins {
new_builtin_mappings.push(SystemObjectMapping {
description: SystemObjectDescription {
schema_name: builtin.schema().to_string(),
object_type: builtin.catalog_item_type(),
object_name: builtin.name().to_string(),
},
unique_identifier: SystemObjectUniqueIdentifier {
catalog_id,
global_id,
fingerprint,
},
});
let handled_runtime_alterable = match builtin {
Builtin::Connection(c) if c.runtime_alterable => {
let mut acl_items = vec![rbac::owner_privilege(
mz_sql::catalog::ObjectType::Connection,
c.owner_id.clone(),
)];
acl_items.extend_from_slice(c.access);
let versions = BTreeMap::new();
txn.insert_item(
catalog_id,
c.oid,
global_id,
mz_catalog::durable::initialize::resolve_system_schema(c.schema).id,
c.name,
c.sql.into(),
*c.owner_id,
acl_items,
versions,
)?;
true
}
_ => false,
};
assert_eq!(
builtin.runtime_alterable(),
handled_runtime_alterable,
"runtime alterable object was not handled by migration",
);
}
txn.set_system_object_mappings(new_builtin_mappings)?;
let mut deleted_system_objects = BTreeSet::new();
let mut deleted_runtime_alterable_system_ids = BTreeSet::new();
for (_, mapping) in system_object_mappings {
deleted_system_objects.insert(mapping.description);
if mapping.unique_identifier.fingerprint == RUNTIME_ALTERABLE_FINGERPRINT_SENTINEL {
deleted_runtime_alterable_system_ids.insert(mapping.unique_identifier.catalog_id);
}
}
let delete_exceptions: HashSet<SystemObjectDescription> = [].into();
assert!(
deleted_system_objects
.iter()
.filter(|object| object.object_type != CatalogItemType::Index)
.all(
|deleted_object| is_unstable_schema(&deleted_object.schema_name)
|| delete_exceptions.contains(deleted_object)
),
"only objects in unstable schemas can be deleted, deleted objects: {:?}",
deleted_system_objects
);
txn.remove_items(&deleted_runtime_alterable_system_ids)?;
txn.remove_system_object_mappings(deleted_system_objects)?;
let new_builtin_collections = new_builtin_ids
.into_iter()
.map(|(_catalog_id, global_id)| global_id)
.collect();
Ok((migrated_builtin_ids, new_builtin_collections))
}
fn add_new_builtin_clusters_migration(
txn: &mut mz_catalog::durable::Transaction<'_>,
builtin_cluster_sizes: &BuiltinBootstrapClusterSizes,
cluster_sizes: &ClusterReplicaSizeMap,
) -> Result<(), mz_catalog::durable::CatalogError> {
let cluster_names: BTreeSet<_> = txn.get_clusters().map(|cluster| cluster.name).collect();
for builtin_cluster in BUILTIN_CLUSTERS {
if !cluster_names.contains(builtin_cluster.name) {
let cluster_size = builtin_cluster_sizes.get_size(builtin_cluster.name)?;
let cluster_allocation = cluster_sizes.get_allocation_by_name(&cluster_size)?;
let id = txn.get_and_increment_id(SYSTEM_CLUSTER_ID_ALLOC_KEY.to_string())?;
let id = ClusterId::System(id);
txn.insert_system_cluster(
id,
builtin_cluster.name,
vec![],
builtin_cluster.privileges.to_vec(),
builtin_cluster.owner_id.to_owned(),
mz_catalog::durable::ClusterConfig {
variant: mz_catalog::durable::ClusterVariant::Managed(ClusterVariantManaged {
size: cluster_size,
availability_zones: vec![],
replication_factor: builtin_cluster.replication_factor,
disk: cluster_allocation.is_cc,
logging: default_logging_config(),
optimizer_feature_overrides: Default::default(),
schedule: Default::default(),
}),
workload_class: None,
},
&HashSet::new(),
)?;
}
}
Ok(())
}
fn add_new_remove_old_builtin_introspection_source_migration(
txn: &mut mz_catalog::durable::Transaction<'_>,
) -> Result<(), AdapterError> {
let mut new_indexes = Vec::new();
let mut removed_indexes = BTreeSet::new();
for cluster in txn.get_clusters() {
let mut introspection_source_index_ids = txn.get_introspection_source_indexes(cluster.id);
let mut new_logs = Vec::new();
for log in BUILTINS::logs() {
if introspection_source_index_ids.remove(log.name).is_none() {
new_logs.push(log);
}
}
let new_ids = txn.allocate_system_item_ids(usize_to_u64(new_logs.len()))?;
let new_entries = new_logs
.into_iter()
.zip_eq(new_ids)
.map(|(log, (item_id, gid))| (log, item_id, gid));
for (log, item_id, gid) in new_entries {
new_indexes.push((cluster.id, log.name.to_string(), item_id, gid));
}
removed_indexes.extend(
introspection_source_index_ids
.into_keys()
.map(|name| (cluster.id, name)),
);
}
txn.insert_introspection_source_indexes(new_indexes, &HashSet::new())?;
txn.remove_introspection_source_indexes(removed_indexes)?;
Ok(())
}
fn add_new_builtin_roles_migration(
txn: &mut mz_catalog::durable::Transaction<'_>,
) -> Result<(), mz_catalog::durable::CatalogError> {
let role_names: BTreeSet<_> = txn.get_roles().map(|role| role.name).collect();
for builtin_role in BUILTIN_ROLES {
if !role_names.contains(builtin_role.name) {
txn.insert_builtin_role(
builtin_role.id,
builtin_role.name.to_string(),
builtin_role.attributes.clone(),
RoleMembership::new(),
RoleVars::default(),
builtin_role.oid,
)?;
}
}
Ok(())
}
fn add_new_builtin_cluster_replicas_migration(
txn: &mut Transaction<'_>,
builtin_cluster_sizes: &BuiltinBootstrapClusterSizes,
cluster_sizes: &ClusterReplicaSizeMap,
) -> Result<(), AdapterError> {
let cluster_lookup: BTreeMap<_, _> = txn
.get_clusters()
.map(|cluster| (cluster.name.clone(), cluster.clone()))
.collect();
let replicas: BTreeMap<_, _> =
txn.get_cluster_replicas()
.fold(BTreeMap::new(), |mut acc, replica| {
acc.entry(replica.cluster_id)
.or_insert_with(BTreeSet::new)
.insert(replica.name);
acc
});
for builtin_replica in BUILTIN_CLUSTER_REPLICAS {
let cluster = cluster_lookup
.get(builtin_replica.cluster_name)
.expect("builtin cluster replica references non-existent cluster");
let replica_names = replicas.get(&cluster.id);
if matches!(replica_names, None)
|| matches!(replica_names, Some(names) if !names.contains(builtin_replica.name))
{
let replica_size = match cluster.config.variant {
ClusterVariant::Managed(ClusterVariantManaged { ref size, .. }) => size.clone(),
ClusterVariant::Unmanaged => {
builtin_cluster_sizes.get_size(builtin_replica.cluster_name)?
}
};
let replica_allocation = cluster_sizes.get_allocation_by_name(&replica_size)?;
let config = builtin_cluster_replica_config(replica_size, replica_allocation);
txn.insert_cluster_replica(
cluster.id,
builtin_replica.name,
config,
MZ_SYSTEM_ROLE_ID,
)?;
}
}
Ok(())
}
fn remove_invalid_config_param_role_defaults_migration(
txn: &mut Transaction<'_>,
) -> Result<(), AdapterError> {
static BUILD_INFO: mz_build_info::BuildInfo = mz_build_info::build_info!();
let roles_to_migrate: BTreeMap<_, _> = txn
.get_roles()
.filter_map(|mut role| {
let session_vars = SessionVars::new_unchecked(&BUILD_INFO, SYSTEM_USER.clone(), None);
let mut invalid_roles_vars = BTreeMap::new();
for (name, value) in &role.vars.map {
let Ok(session_var) = session_vars.inspect(name) else {
invalid_roles_vars.insert(name.clone(), value.clone());
continue;
};
if session_var.check(value.borrow()).is_err() {
invalid_roles_vars.insert(name.clone(), value.clone());
}
}
if invalid_roles_vars.is_empty() {
return None;
}
tracing::warn!(?role, ?invalid_roles_vars, "removing invalid role vars");
for (name, _value) in invalid_roles_vars {
role.vars.map.remove(&name);
}
Some(role)
})
.map(|role| (role.id, role))
.collect();
txn.update_roles(roles_to_migrate)?;
Ok(())
}
fn remove_pending_cluster_replicas_migration(tx: &mut Transaction) -> Result<(), anyhow::Error> {
for replica in tx.get_cluster_replicas() {
if let mz_catalog::durable::ReplicaLocation::Managed { pending: true, .. } =
replica.config.location
{
tx.remove_cluster_replica(replica.replica_id)?;
}
}
Ok(())
}
pub(crate) fn builtin_cluster_replica_config(
replica_size: String,
replica_allocation: &ReplicaAllocation,
) -> mz_catalog::durable::ReplicaConfig {
mz_catalog::durable::ReplicaConfig {
location: mz_catalog::durable::ReplicaLocation::Managed {
availability_zone: None,
billed_as: None,
disk: replica_allocation.is_cc,
pending: false,
internal: false,
size: replica_size,
},
logging: default_logging_config(),
}
}
fn default_logging_config() -> ReplicaLogging {
ReplicaLogging {
log_logging: false,
interval: Some(Duration::from_secs(1)),
}
}
pub struct BuiltinBootstrapClusterSizes {
pub system_cluster: String,
pub catalog_server_cluster: String,
pub probe_cluster: String,
pub support_cluster: String,
pub analytics_cluster: String,
}
impl BuiltinBootstrapClusterSizes {
fn get_size(&self, cluster_name: &str) -> Result<String, mz_catalog::durable::CatalogError> {
if cluster_name == mz_catalog::builtin::MZ_SYSTEM_CLUSTER.name {
Ok(self.system_cluster.clone())
} else if cluster_name == mz_catalog::builtin::MZ_CATALOG_SERVER_CLUSTER.name {
Ok(self.catalog_server_cluster.clone())
} else if cluster_name == mz_catalog::builtin::MZ_PROBE_CLUSTER.name {
Ok(self.probe_cluster.clone())
} else if cluster_name == mz_catalog::builtin::MZ_SUPPORT_CLUSTER.name {
Ok(self.support_cluster.clone())
} else if cluster_name == mz_catalog::builtin::MZ_ANALYTICS_CLUSTER.name {
Ok(self.analytics_cluster.clone())
} else {
Err(mz_catalog::durable::CatalogError::Catalog(
SqlCatalogError::UnexpectedBuiltinCluster(cluster_name.to_owned()),
))
}
}
}
pub(crate) fn into_consolidatable_updates_startup(
updates: Vec<StateUpdate>,
ts: Timestamp,
) -> Vec<(BootstrapStateUpdateKind, Timestamp, Diff)> {
updates
.into_iter()
.map(|StateUpdate { kind, ts: _, diff }| {
let kind: BootstrapStateUpdateKind = kind
.try_into()
.unwrap_or_else(|e| panic!("temporary items do not exist during bootstrap: {e:?}"));
(kind, ts, Diff::from(diff))
})
.collect()
}
fn get_dyncfg_val_from_defaults_and_remote<T: mz_dyncfg::ConfigDefault>(
defaults: &BTreeMap<String, String>,
remote: Option<&BTreeMap<String, String>>,
cfg: &mz_dyncfg::Config<T>,
) -> T::ConfigType {
let mut val = T::into_config_type(cfg.default().clone());
let get_fn = |map: &BTreeMap<String, String>| {
let val = map.get(cfg.name())?;
match <T::ConfigType as mz_dyncfg::ConfigType>::parse(val) {
Ok(x) => Some(x),
Err(err) => {
tracing::warn!("could not parse {} value [{}]: {}", cfg.name(), val, err);
None
}
}
};
if let Some(x) = get_fn(defaults) {
val = x;
}
if let Some(x) = remote.and_then(get_fn) {
val = x;
}
val
}
#[cfg(test)]
mod builtin_migration_tests {
use std::collections::{BTreeMap, BTreeSet};
use itertools::Itertools;
use mz_catalog::memory::objects::{
CatalogItem, Index, MaterializedView, Table, TableDataSource,
};
use mz_catalog::SYSTEM_CONN_ID;
use mz_controller_types::ClusterId;
use mz_expr::MirRelationExpr;
use mz_ore::id_gen::Gen;
use mz_repr::{
CatalogItemId, GlobalId, RelationDesc, RelationType, RelationVersion, ScalarType,
};
use mz_sql::catalog::CatalogDatabase;
use mz_sql::names::{
DependencyIds, ItemQualifiers, QualifiedItemName, ResolvedDatabaseSpecifier, ResolvedIds,
};
use mz_sql::session::user::MZ_SYSTEM_ROLE_ID;
use mz_sql::DEFAULT_SCHEMA;
use mz_sql_parser::ast::Expr;
use crate::catalog::{Catalog, Op, OptimizedMirRelationExpr};
use crate::session::DEFAULT_DATABASE_NAME;
enum ItemNamespace {
System,
User,
}
enum SimplifiedItem {
Table,
MaterializedView { referenced_names: Vec<String> },
Index { on: String },
}
struct SimplifiedCatalogEntry {
name: String,
namespace: ItemNamespace,
item: SimplifiedItem,
}
impl SimplifiedCatalogEntry {
fn to_catalog_item(
self,
item_id_mapping: &BTreeMap<String, CatalogItemId>,
global_id_mapping: &BTreeMap<String, GlobalId>,
global_id_gen: &mut Gen<u64>,
) -> (String, ItemNamespace, CatalogItem) {
let item = match self.item {
SimplifiedItem::Table => CatalogItem::Table(Table {
create_sql: Some("CREATE TABLE materialize.public.t (a INT)".to_string()),
desc: RelationDesc::builder()
.with_column("a", ScalarType::Int32.nullable(true))
.with_key(vec![0])
.finish(),
collections: [(
RelationVersion::root(),
GlobalId::User(global_id_gen.allocate_id()),
)]
.into_iter()
.collect(),
conn_id: None,
resolved_ids: ResolvedIds::empty(),
custom_logical_compaction_window: None,
is_retained_metrics_object: false,
data_source: TableDataSource::TableWrites {
defaults: vec![Expr::null(); 1],
},
}),
SimplifiedItem::MaterializedView { referenced_names } => {
let table_list = referenced_names
.iter()
.map(|table| format!("materialize.public.{table}"))
.join(",");
let column_list = referenced_names
.iter()
.enumerate()
.map(|(idx, _)| format!("a{idx}"))
.join(",");
let resolved_ids =
convert_names_to_ids(referenced_names, item_id_mapping, global_id_mapping);
CatalogItem::MaterializedView(MaterializedView {
global_id: GlobalId::User(global_id_gen.allocate_id()),
create_sql: format!(
"CREATE MATERIALIZED VIEW materialize.public.mv ({column_list}) AS SELECT * FROM {table_list}"
),
raw_expr: mz_sql::plan::HirRelationExpr::constant(
Vec::new(),
RelationType {
column_types: Vec::new(),
keys: Vec::new(),
},
).into(),
dependencies: DependencyIds(Default::default()),
optimized_expr: OptimizedMirRelationExpr(MirRelationExpr::Constant {
rows: Ok(Vec::new()),
typ: RelationType {
column_types: Vec::new(),
keys: Vec::new(),
},
}).into(),
desc: RelationDesc::builder()
.with_column("a", ScalarType::Int32.nullable(true))
.with_key(vec![0])
.finish(),
resolved_ids: resolved_ids.into_iter().collect(),
cluster_id: ClusterId::User(1),
non_null_assertions: vec![],
custom_logical_compaction_window: None,
refresh_schedule: None,
initial_as_of: None,
})
}
SimplifiedItem::Index { on } => {
let on_item_id = item_id_mapping[&on];
let on_gid = global_id_mapping[&on];
CatalogItem::Index(Index {
create_sql: format!("CREATE INDEX idx ON materialize.public.{on} (a)"),
global_id: GlobalId::User(global_id_gen.allocate_id()),
on: on_gid,
keys: Default::default(),
conn_id: None,
resolved_ids: [(on_item_id, on_gid)].into_iter().collect(),
cluster_id: ClusterId::User(1),
custom_logical_compaction_window: None,
is_retained_metrics_object: false,
})
}
};
(self.name, self.namespace, item)
}
}
struct BuiltinMigrationTestCase {
test_name: &'static str,
initial_state: Vec<SimplifiedCatalogEntry>,
migrated_names: Vec<String>,
expected_previous_storage_collection_names: Vec<String>,
expected_migrated_system_object_mappings: Vec<String>,
expected_user_item_drop_ops: Vec<String>,
expected_user_item_create_ops: Vec<String>,
}
async fn add_item(
catalog: &mut Catalog,
name: String,
item: CatalogItem,
item_namespace: ItemNamespace,
) -> (CatalogItemId, GlobalId) {
let (item_id, global_id) = match item_namespace {
ItemNamespace::User => catalog
.allocate_user_id()
.await
.expect("cannot fail to allocate user ids"),
ItemNamespace::System => catalog
.allocate_system_id()
.await
.expect("cannot fail to allocate system ids"),
};
let database_id = catalog
.resolve_database(DEFAULT_DATABASE_NAME)
.expect("failed to resolve default database")
.id();
let database_spec = ResolvedDatabaseSpecifier::Id(database_id);
let schema_spec = catalog
.resolve_schema_in_database(&database_spec, DEFAULT_SCHEMA, &SYSTEM_CONN_ID)
.expect("failed to resolve default schemas")
.id
.clone();
catalog
.transact(
None,
mz_repr::Timestamp::MIN,
None,
vec![Op::CreateItem {
id: item_id,
name: QualifiedItemName {
qualifiers: ItemQualifiers {
database_spec,
schema_spec,
},
item: name,
},
item,
owner_id: MZ_SYSTEM_ROLE_ID,
}],
)
.await
.expect("failed to transact");
(item_id, global_id)
}
fn convert_names_to_ids(
name_vec: Vec<String>,
item_id_lookup: &BTreeMap<String, CatalogItemId>,
global_id_lookup: &BTreeMap<String, GlobalId>,
) -> BTreeMap<CatalogItemId, GlobalId> {
name_vec
.into_iter()
.map(|name| {
let item_id = item_id_lookup[&name];
let global_id = global_id_lookup[&name];
(item_id, global_id)
})
.collect()
}
fn convert_ids_to_names<I: IntoIterator<Item = CatalogItemId>>(
ids: I,
name_lookup: &BTreeMap<CatalogItemId, String>,
) -> BTreeSet<String> {
ids.into_iter().map(|id| name_lookup[&id].clone()).collect()
}
async fn run_test_case(test_case: BuiltinMigrationTestCase) {
Catalog::with_debug(|mut catalog| async move {
let mut item_id_mapping = BTreeMap::new();
let mut name_mapping = BTreeMap::new();
let mut global_id_gen = Gen::<u64>::default();
let mut global_id_mapping = BTreeMap::new();
for entry in test_case.initial_state {
let (name, namespace, item) =
entry.to_catalog_item(&item_id_mapping, &global_id_mapping, &mut global_id_gen);
let (item_id, global_id) =
add_item(&mut catalog, name.clone(), item, namespace).await;
item_id_mapping.insert(name.clone(), item_id);
global_id_mapping.insert(name.clone(), global_id);
name_mapping.insert(item_id, name);
}
let migrated_ids = test_case
.migrated_names
.into_iter()
.map(|name| item_id_mapping[&name])
.collect();
let id_fingerprint_map: BTreeMap<CatalogItemId, String> = item_id_mapping
.iter()
.filter(|(_name, id)| id.is_system())
.map(|(_name, id)| (*id, "".to_string()))
.collect();
let migration_metadata = {
let state = catalog.state.clone();
let mut storage = catalog.storage().await;
let mut txn = storage
.transaction()
.await
.expect("failed to create transaction");
Catalog::generate_builtin_migration_metadata(
&state,
&mut txn,
migrated_ids,
id_fingerprint_map,
)
.expect("failed to generate builtin migration metadata")
};
assert_eq!(
convert_ids_to_names(
migration_metadata
.previous_storage_collection_ids
.into_iter(),
&name_mapping
),
test_case
.expected_previous_storage_collection_names
.into_iter()
.collect(),
"{} test failed with wrong previous collection_names",
test_case.test_name
);
assert_eq!(
migration_metadata
.migrated_system_object_mappings
.values()
.map(|mapping| mapping.description.object_name.clone())
.collect::<BTreeSet<_>>(),
test_case
.expected_migrated_system_object_mappings
.into_iter()
.collect(),
"{} test failed with wrong migrated system object mappings",
test_case.test_name
);
assert_eq!(
convert_ids_to_names(
migration_metadata.user_item_drop_ops.into_iter(),
&name_mapping
),
test_case.expected_user_item_drop_ops.into_iter().collect(),
"{} test failed with wrong user drop ops",
test_case.test_name
);
assert_eq!(
migration_metadata
.user_item_create_ops
.into_iter()
.map(|create_op| create_op.name)
.collect::<BTreeSet<_>>(),
test_case
.expected_user_item_create_ops
.into_iter()
.collect(),
"{} test failed with wrong user create ops",
test_case.test_name
);
catalog.expire().await;
})
.await
}
#[mz_ore::test(tokio::test)]
#[cfg_attr(miri, ignore)] async fn test_builtin_migration_no_migrations() {
let test_case = BuiltinMigrationTestCase {
test_name: "no_migrations",
initial_state: vec![SimplifiedCatalogEntry {
name: "s1".to_string(),
namespace: ItemNamespace::System,
item: SimplifiedItem::Table,
}],
migrated_names: vec![],
expected_previous_storage_collection_names: vec![],
expected_migrated_system_object_mappings: vec![],
expected_user_item_drop_ops: vec![],
expected_user_item_create_ops: vec![],
};
run_test_case(test_case).await;
}
#[mz_ore::test(tokio::test)]
#[cfg_attr(miri, ignore)] async fn test_builtin_migration_single_migrations() {
let test_case = BuiltinMigrationTestCase {
test_name: "single_migrations",
initial_state: vec![SimplifiedCatalogEntry {
name: "s1".to_string(),
namespace: ItemNamespace::System,
item: SimplifiedItem::Table,
}],
migrated_names: vec!["s1".to_string()],
expected_previous_storage_collection_names: vec!["s1".to_string()],
expected_migrated_system_object_mappings: vec!["s1".to_string()],
expected_user_item_drop_ops: vec![],
expected_user_item_create_ops: vec![],
};
run_test_case(test_case).await;
}
#[mz_ore::test(tokio::test)]
#[cfg_attr(miri, ignore)] async fn test_builtin_migration_child_migrations() {
let test_case = BuiltinMigrationTestCase {
test_name: "child_migrations",
initial_state: vec![
SimplifiedCatalogEntry {
name: "s1".to_string(),
namespace: ItemNamespace::System,
item: SimplifiedItem::Table,
},
SimplifiedCatalogEntry {
name: "u1".to_string(),
namespace: ItemNamespace::User,
item: SimplifiedItem::MaterializedView {
referenced_names: vec!["s1".to_string()],
},
},
],
migrated_names: vec!["s1".to_string()],
expected_previous_storage_collection_names: vec!["u1".to_string(), "s1".to_string()],
expected_migrated_system_object_mappings: vec!["s1".to_string()],
expected_user_item_drop_ops: vec!["u1".to_string()],
expected_user_item_create_ops: vec!["u1".to_string()],
};
run_test_case(test_case).await;
}
#[mz_ore::test(tokio::test)]
#[cfg_attr(miri, ignore)] async fn test_builtin_migration_multi_child_migrations() {
let test_case = BuiltinMigrationTestCase {
test_name: "multi_child_migrations",
initial_state: vec![
SimplifiedCatalogEntry {
name: "s1".to_string(),
namespace: ItemNamespace::System,
item: SimplifiedItem::Table,
},
SimplifiedCatalogEntry {
name: "u1".to_string(),
namespace: ItemNamespace::User,
item: SimplifiedItem::MaterializedView {
referenced_names: vec!["s1".to_string()],
},
},
SimplifiedCatalogEntry {
name: "u2".to_string(),
namespace: ItemNamespace::User,
item: SimplifiedItem::MaterializedView {
referenced_names: vec!["s1".to_string()],
},
},
],
migrated_names: vec!["s1".to_string()],
expected_previous_storage_collection_names: vec![
"u1".to_string(),
"u2".to_string(),
"s1".to_string(),
],
expected_migrated_system_object_mappings: vec!["s1".to_string()],
expected_user_item_drop_ops: vec!["u1".to_string(), "u2".to_string()],
expected_user_item_create_ops: vec!["u2".to_string(), "u1".to_string()],
};
run_test_case(test_case).await;
}
#[mz_ore::test(tokio::test)]
#[cfg_attr(miri, ignore)] async fn test_builtin_migration_topological_sort() {
let test_case = BuiltinMigrationTestCase {
test_name: "topological_sort",
initial_state: vec![
SimplifiedCatalogEntry {
name: "s1".to_string(),
namespace: ItemNamespace::System,
item: SimplifiedItem::Table,
},
SimplifiedCatalogEntry {
name: "s2".to_string(),
namespace: ItemNamespace::System,
item: SimplifiedItem::Table,
},
SimplifiedCatalogEntry {
name: "u1".to_string(),
namespace: ItemNamespace::User,
item: SimplifiedItem::MaterializedView {
referenced_names: vec!["s2".to_string()],
},
},
SimplifiedCatalogEntry {
name: "u2".to_string(),
namespace: ItemNamespace::User,
item: SimplifiedItem::MaterializedView {
referenced_names: vec!["s1".to_string(), "u1".to_string()],
},
},
],
migrated_names: vec!["s1".to_string(), "s2".to_string()],
expected_previous_storage_collection_names: vec![
"u2".to_string(),
"u1".to_string(),
"s1".to_string(),
"s2".to_string(),
],
expected_migrated_system_object_mappings: vec!["s1".to_string(), "s2".to_string()],
expected_user_item_drop_ops: vec!["u2".to_string(), "u1".to_string()],
expected_user_item_create_ops: vec!["u1".to_string(), "u2".to_string()],
};
run_test_case(test_case).await;
}
#[mz_ore::test(tokio::test)]
#[cfg_attr(miri, ignore)] async fn test_builtin_migration_topological_sort_complex() {
let test_case = BuiltinMigrationTestCase {
test_name: "topological_sort_complex",
initial_state: vec![
SimplifiedCatalogEntry {
name: "s273".to_string(),
namespace: ItemNamespace::System,
item: SimplifiedItem::Table,
},
SimplifiedCatalogEntry {
name: "s322".to_string(),
namespace: ItemNamespace::System,
item: SimplifiedItem::Table,
},
SimplifiedCatalogEntry {
name: "s317".to_string(),
namespace: ItemNamespace::System,
item: SimplifiedItem::Table,
},
SimplifiedCatalogEntry {
name: "s349".to_string(),
namespace: ItemNamespace::System,
item: SimplifiedItem::MaterializedView {
referenced_names: vec!["s273".to_string()],
},
},
SimplifiedCatalogEntry {
name: "s421".to_string(),
namespace: ItemNamespace::System,
item: SimplifiedItem::MaterializedView {
referenced_names: vec!["s273".to_string()],
},
},
SimplifiedCatalogEntry {
name: "s295".to_string(),
namespace: ItemNamespace::System,
item: SimplifiedItem::MaterializedView {
referenced_names: vec!["s273".to_string()],
},
},
SimplifiedCatalogEntry {
name: "s296".to_string(),
namespace: ItemNamespace::System,
item: SimplifiedItem::MaterializedView {
referenced_names: vec!["s295".to_string()],
},
},
SimplifiedCatalogEntry {
name: "s320".to_string(),
namespace: ItemNamespace::System,
item: SimplifiedItem::MaterializedView {
referenced_names: vec!["s295".to_string()],
},
},
SimplifiedCatalogEntry {
name: "s340".to_string(),
namespace: ItemNamespace::System,
item: SimplifiedItem::MaterializedView {
referenced_names: vec!["s295".to_string()],
},
},
SimplifiedCatalogEntry {
name: "s318".to_string(),
namespace: ItemNamespace::System,
item: SimplifiedItem::MaterializedView {
referenced_names: vec!["s295".to_string()],
},
},
SimplifiedCatalogEntry {
name: "s323".to_string(),
namespace: ItemNamespace::System,
item: SimplifiedItem::MaterializedView {
referenced_names: vec!["s295".to_string(), "s322".to_string()],
},
},
SimplifiedCatalogEntry {
name: "s330".to_string(),
namespace: ItemNamespace::System,
item: SimplifiedItem::MaterializedView {
referenced_names: vec!["s318".to_string(), "s317".to_string()],
},
},
SimplifiedCatalogEntry {
name: "s321".to_string(),
namespace: ItemNamespace::System,
item: SimplifiedItem::MaterializedView {
referenced_names: vec!["s318".to_string()],
},
},
SimplifiedCatalogEntry {
name: "s315".to_string(),
namespace: ItemNamespace::System,
item: SimplifiedItem::MaterializedView {
referenced_names: vec!["s296".to_string()],
},
},
SimplifiedCatalogEntry {
name: "s354".to_string(),
namespace: ItemNamespace::System,
item: SimplifiedItem::MaterializedView {
referenced_names: vec!["s296".to_string()],
},
},
SimplifiedCatalogEntry {
name: "s327".to_string(),
namespace: ItemNamespace::System,
item: SimplifiedItem::MaterializedView {
referenced_names: vec!["s296".to_string()],
},
},
SimplifiedCatalogEntry {
name: "s339".to_string(),
namespace: ItemNamespace::System,
item: SimplifiedItem::MaterializedView {
referenced_names: vec!["s296".to_string()],
},
},
SimplifiedCatalogEntry {
name: "s355".to_string(),
namespace: ItemNamespace::System,
item: SimplifiedItem::MaterializedView {
referenced_names: vec!["s315".to_string()],
},
},
],
migrated_names: vec![
"s273".to_string(),
"s317".to_string(),
"s318".to_string(),
"s320".to_string(),
"s321".to_string(),
"s322".to_string(),
"s323".to_string(),
"s330".to_string(),
"s339".to_string(),
"s340".to_string(),
],
expected_previous_storage_collection_names: vec![
"s349".to_string(),
"s421".to_string(),
"s355".to_string(),
"s315".to_string(),
"s354".to_string(),
"s327".to_string(),
"s339".to_string(),
"s296".to_string(),
"s320".to_string(),
"s340".to_string(),
"s330".to_string(),
"s321".to_string(),
"s318".to_string(),
"s323".to_string(),
"s295".to_string(),
"s273".to_string(),
"s317".to_string(),
"s322".to_string(),
],
expected_migrated_system_object_mappings: vec![
"s322".to_string(),
"s317".to_string(),
"s273".to_string(),
"s295".to_string(),
"s323".to_string(),
"s318".to_string(),
"s321".to_string(),
"s330".to_string(),
"s340".to_string(),
"s320".to_string(),
"s296".to_string(),
"s339".to_string(),
"s327".to_string(),
"s354".to_string(),
"s315".to_string(),
"s355".to_string(),
"s421".to_string(),
"s349".to_string(),
],
expected_user_item_drop_ops: vec![],
expected_user_item_create_ops: vec![],
};
run_test_case(test_case).await;
}
#[mz_ore::test(tokio::test)]
#[cfg_attr(miri, ignore)] async fn test_builtin_migration_system_child_migrations() {
let test_case = BuiltinMigrationTestCase {
test_name: "system_child_migrations",
initial_state: vec![
SimplifiedCatalogEntry {
name: "s1".to_string(),
namespace: ItemNamespace::System,
item: SimplifiedItem::Table,
},
SimplifiedCatalogEntry {
name: "s2".to_string(),
namespace: ItemNamespace::System,
item: SimplifiedItem::Index {
on: "s1".to_string(),
},
},
],
migrated_names: vec!["s1".to_string()],
expected_previous_storage_collection_names: vec!["s1".to_string()],
expected_migrated_system_object_mappings: vec!["s1".to_string(), "s2".to_string()],
expected_user_item_drop_ops: vec![],
expected_user_item_create_ops: vec![],
};
run_test_case(test_case).await;
}
}