use std::{collections::BTreeMap, time::Duration};
use anyhow::bail;
use k8s_openapi::{
api::{
apps::v1::{
Deployment, DeploymentSpec, DeploymentStrategy, RollingUpdateDeployment, StatefulSet,
StatefulSetSpec, StatefulSetUpdateStrategy,
},
core::v1::{
Capabilities, Container, ContainerPort, EnvVar, EnvVarSource, EphemeralVolumeSource,
HTTPGetAction, PersistentVolumeClaimSpec, PersistentVolumeClaimTemplate, Pod,
PodSecurityContext, PodSpec, PodTemplateSpec, Probe, SeccompProfile, SecretKeySelector,
SecurityContext, Service, ServiceAccount, ServicePort, ServiceSpec, TCPSocketAction,
Toleration, Volume, VolumeMount, VolumeResourceRequirements,
},
networking::v1::{
IPBlock, NetworkPolicy, NetworkPolicyEgressRule, NetworkPolicyIngressRule,
NetworkPolicyPeer, NetworkPolicyPort, NetworkPolicySpec,
},
rbac::v1::{PolicyRule, Role, RoleBinding, RoleRef, Subject},
},
apimachinery::pkg::{apis::meta::v1::LabelSelector, util::intstr::IntOrString},
};
use kube::{api::ObjectMeta, runtime::controller::Action, Api, Client, ResourceExt};
use maplit::btreemap;
use rand::{thread_rng, Rng};
use serde::{Deserialize, Serialize};
use sha2::{Digest, Sha256};
use tracing::trace;
use super::{matching_image_from_environmentd_image_ref, CloudProvider};
use crate::k8s::{apply_resource, delete_resource, get_resource};
use mz_cloud_resources::crd::materialize::v1alpha1::Materialize;
use mz_environmentd::DeploymentStatus;
use mz_orchestrator_tracing::TracingCliArgs;
use mz_ore::instrument;
#[derive(Debug, Serialize)]
pub struct Resources {
generation: u64,
environmentd_network_policies: Vec<NetworkPolicy>,
service_account: Box<ServiceAccount>,
role: Box<Role>,
role_binding: Box<RoleBinding>,
public_service: Box<Service>,
generation_service: Box<Service>,
persist_pubsub_service: Box<Service>,
environmentd_statefulset: Box<StatefulSet>,
balancerd_deployment: Option<Box<Deployment>>,
balancerd_service: Option<Box<Service>>,
}
impl Resources {
pub fn new(
config: &super::Args,
tracing: &TracingCliArgs,
orchestratord_namespace: &str,
mz: &Materialize,
generation: u64,
) -> Self {
let environmentd_network_policies =
create_environmentd_network_policies(config, mz, orchestratord_namespace);
let service_account = Box::new(create_service_account_object(config, mz));
let role = Box::new(create_role_object(mz));
let role_binding = Box::new(create_role_binding_object(mz));
let public_service = Box::new(create_public_service_object(config, mz, generation));
let generation_service = Box::new(create_generation_service_object(config, mz, generation));
let persist_pubsub_service =
Box::new(create_persist_pubsub_service(config, mz, generation));
let environmentd_statefulset = Box::new(create_environmentd_statefulset_object(
config, tracing, mz, generation,
));
let balancerd_deployment = config
.create_balancers
.then(|| Box::new(create_balancerd_deployment_object(config, mz)));
let balancerd_service = config
.create_balancers
.then(|| Box::new(create_balancerd_service_object(config, mz)));
Self {
generation,
environmentd_network_policies,
service_account,
role,
role_binding,
public_service,
generation_service,
persist_pubsub_service,
environmentd_statefulset,
balancerd_deployment,
balancerd_service,
}
}
#[instrument]
pub async fn apply(
&self,
client: &Client,
args: &super::Args,
increment_generation: bool,
namespace: &str,
) -> Result<Option<Action>, anyhow::Error> {
let environmentd_network_policy_api: Api<NetworkPolicy> =
Api::namespaced(client.clone(), namespace);
let service_api: Api<Service> = Api::namespaced(client.clone(), namespace);
let service_account_api: Api<ServiceAccount> = Api::namespaced(client.clone(), namespace);
let role_api: Api<Role> = Api::namespaced(client.clone(), namespace);
let role_binding_api: Api<RoleBinding> = Api::namespaced(client.clone(), namespace);
let statefulset_api: Api<StatefulSet> = Api::namespaced(client.clone(), namespace);
let pod_api: Api<Pod> = Api::namespaced(client.clone(), namespace);
for policy in &self.environmentd_network_policies {
trace!("applying network policy {}", policy.name_unchecked());
apply_resource(&environmentd_network_policy_api, policy).await?;
}
trace!("applying environmentd service account");
apply_resource(&service_account_api, &*self.service_account).await?;
trace!("applying environmentd role");
apply_resource(&role_api, &*self.role).await?;
trace!("applying environmentd role binding");
apply_resource(&role_binding_api, &*self.role_binding).await?;
trace!("applying environmentd per-generation service");
apply_resource(&service_api, &*self.generation_service).await?;
trace!("creating persist pubsub service");
apply_resource(&service_api, &*self.persist_pubsub_service).await?;
trace!("creating new environmentd statefulset");
apply_resource(&statefulset_api, &*self.environmentd_statefulset).await?;
if increment_generation {
let retry_action = Action::requeue(Duration::from_secs(thread_rng().gen_range(5..10)));
let statefulset = get_resource(
&statefulset_api,
&self.environmentd_statefulset.name_unchecked(),
)
.await?;
if statefulset
.and_then(|statefulset| statefulset.status)
.and_then(|status| status.ready_replicas)
.unwrap_or(0)
== 0
{
trace!("environmentd statefulset is not ready yet...");
return Ok(Some(retry_action));
}
let environmentd_url =
environmentd_internal_http_address(args, namespace, &*self.generation_service);
let http_client = reqwest::Client::builder()
.timeout(std::time::Duration::from_secs(10))
.build()
.unwrap();
let status_url =
reqwest::Url::parse(&format!("http://{}/api/leader/status", environmentd_url))
.unwrap();
match http_client.get(status_url.clone()).send().await {
Ok(response) => {
let response: BTreeMap<String, DeploymentStatus> = response.json().await?;
if response["status"] == DeploymentStatus::Initializing {
trace!("environmentd is still initializing, retrying...");
return Ok(Some(retry_action));
} else {
trace!("environmentd is ready");
}
}
Err(e) => {
trace!("failed to connect to environmentd, retrying... ({e})");
return Ok(Some(retry_action));
}
}
let promote_url =
reqwest::Url::parse(&format!("http://{}/api/leader/promote", environmentd_url))
.unwrap();
trace!("promoting new environmentd to leader");
let response = http_client.post(promote_url).send().await?;
let response: BecomeLeaderResponse = response.json().await?;
if let BecomeLeaderResult::Failure { message } = response.result {
bail!("failed to promote new environmentd: {message}");
}
match http_client.get(status_url.clone()).send().await {
Ok(response) => {
let response: BTreeMap<String, DeploymentStatus> = response.json().await?;
if response["status"] != DeploymentStatus::IsLeader {
trace!(
"environmentd is still promoting (status: {:?}), retrying...",
response["status"]
);
return Ok(Some(retry_action));
} else {
trace!("environmentd is ready");
}
}
Err(e) => {
trace!("failed to connect to environmentd, retrying... ({e})");
return Ok(Some(retry_action));
}
}
} else {
trace!("restarting environmentd pod to pick up statefulset changes");
delete_resource(
&pod_api,
&statefulset_pod_name(&*self.environmentd_statefulset, 0),
)
.await?;
}
Ok(None)
}
#[instrument]
pub async fn promote_services(
&self,
client: &Client,
namespace: &str,
) -> Result<(), anyhow::Error> {
let service_api: Api<Service> = Api::namespaced(client.clone(), namespace);
let deployment_api: Api<Deployment> = Api::namespaced(client.clone(), namespace);
trace!("applying environmentd public service");
apply_resource(&service_api, &*self.public_service).await?;
if let Some(balancerd_deployment) = &self.balancerd_deployment {
trace!("creating new balancerd deployment");
apply_resource(&deployment_api, &*balancerd_deployment).await?;
}
if let Some(balancerd_service) = &self.balancerd_service {
trace!("creating new balancerd service");
apply_resource(&service_api, &*balancerd_service).await?;
}
Ok(())
}
#[instrument]
pub async fn teardown_generation(
&self,
client: &Client,
mz: &Materialize,
generation: u64,
) -> Result<(), anyhow::Error> {
let service_api: Api<Service> = Api::namespaced(client.clone(), &mz.namespace());
let statefulset_api: Api<StatefulSet> = Api::namespaced(client.clone(), &mz.namespace());
trace!("deleting environmentd statefulset for generation {generation}");
delete_resource(
&statefulset_api,
&mz.environmentd_statefulset_name(generation),
)
.await?;
trace!("deleting persist pubsub service for generation {generation}");
delete_resource(&service_api, &mz.persist_pubsub_service_name(generation)).await?;
trace!("deleting environmentd per-generation service for generation {generation}");
delete_resource(
&service_api,
&mz.environmentd_generation_service_name(generation),
)
.await?;
Ok(())
}
pub fn generate_hash(&self) -> String {
let mut hasher = Sha256::new();
hasher.update(&serde_json::to_string(self).unwrap());
format!("{:x}", hasher.finalize())
}
}
fn create_environmentd_network_policies(
config: &super::Args,
mz: &Materialize,
orchestratord_namespace: &str,
) -> Vec<NetworkPolicy> {
let mut network_policies = Vec::new();
if config.network_policies.internal_enabled {
let environmentd_label_selector = LabelSelector {
match_labels: Some(
mz.default_labels()
.into_iter()
.chain([(
"materialize.cloud/app".to_owned(),
mz.environmentd_app_name(),
)])
.collect(),
),
..Default::default()
};
let orchestratord_label_selector = LabelSelector {
match_labels: Some(
config
.orchestratord_pod_selector_labels
.iter()
.cloned()
.map(|kv| (kv.key, kv.value))
.collect(),
),
..Default::default()
};
let all_pods_label_selector = LabelSelector {
match_labels: Some(mz.default_labels()),
..Default::default()
};
network_policies.extend([
NetworkPolicy {
metadata: mz
.managed_resource_meta(mz.name_prefixed("allow-all-within-environment")),
spec: Some(NetworkPolicySpec {
egress: Some(vec![NetworkPolicyEgressRule {
to: Some(vec![NetworkPolicyPeer {
pod_selector: Some(all_pods_label_selector.clone()),
..Default::default()
}]),
..Default::default()
}]),
ingress: Some(vec![NetworkPolicyIngressRule {
from: Some(vec![NetworkPolicyPeer {
pod_selector: Some(all_pods_label_selector.clone()),
..Default::default()
}]),
..Default::default()
}]),
pod_selector: all_pods_label_selector.clone(),
policy_types: Some(vec!["Ingress".to_owned(), "Egress".to_owned()]),
..Default::default()
}),
},
NetworkPolicy {
metadata: mz.managed_resource_meta(mz.name_prefixed("allow-orchestratord")),
spec: Some(NetworkPolicySpec {
ingress: Some(vec![NetworkPolicyIngressRule {
from: Some(vec![NetworkPolicyPeer {
namespace_selector: Some(LabelSelector {
match_labels: Some(btreemap! {
"kubernetes.io/metadata.name".into()
=> orchestratord_namespace.into(),
}),
..Default::default()
}),
pod_selector: Some(orchestratord_label_selector),
..Default::default()
}]),
ports: Some(vec![NetworkPolicyPort {
port: Some(IntOrString::Int(config.environmentd_internal_http_port)),
protocol: Some("TCP".to_string()),
..Default::default()
}]),
..Default::default()
}]),
pod_selector: environmentd_label_selector,
policy_types: Some(vec!["Ingress".to_owned()]),
..Default::default()
}),
},
]);
}
if config.network_policies.ingress_enabled {
let mut ingress_label_selector = mz.default_labels();
ingress_label_selector.insert("materialize.cloud/app".to_owned(), mz.balancerd_app_name());
network_policies.extend([NetworkPolicy {
metadata: mz.managed_resource_meta(mz.name_prefixed("sql-and-http-ingress")),
spec: Some(NetworkPolicySpec {
ingress: Some(vec![NetworkPolicyIngressRule {
from: Some(
config
.network_policies
.ingress_cidrs
.iter()
.map(|cidr| NetworkPolicyPeer {
ip_block: Some(IPBlock {
cidr: cidr.to_owned(),
except: None,
}),
..Default::default()
})
.collect(),
),
ports: Some(vec![
NetworkPolicyPort {
port: Some(IntOrString::Int(config.environmentd_http_port)),
protocol: Some("TCP".to_string()),
..Default::default()
},
NetworkPolicyPort {
port: Some(IntOrString::Int(config.environmentd_sql_port)),
protocol: Some("TCP".to_string()),
..Default::default()
},
]),
..Default::default()
}]),
pod_selector: LabelSelector {
match_expressions: None,
match_labels: Some(ingress_label_selector),
},
policy_types: Some(vec!["Ingress".to_owned()]),
..Default::default()
}),
}]);
}
if config.network_policies.egress_enabled {
network_policies.extend([NetworkPolicy {
metadata: mz.managed_resource_meta(mz.name_prefixed("sources-and-sinks-egress")),
spec: Some(NetworkPolicySpec {
egress: Some(vec![NetworkPolicyEgressRule {
to: Some(
config
.network_policies
.egress_cidrs
.iter()
.map(|cidr| NetworkPolicyPeer {
ip_block: Some(IPBlock {
cidr: cidr.to_owned(),
except: None,
}),
..Default::default()
})
.collect(),
),
..Default::default()
}]),
pod_selector: LabelSelector {
match_expressions: None,
match_labels: Some(mz.default_labels()),
},
policy_types: Some(vec!["Egress".to_owned()]),
..Default::default()
}),
}]);
}
network_policies
}
fn create_service_account_object(config: &super::Args, mz: &Materialize) -> ServiceAccount {
let annotations = if config.cloud_provider == CloudProvider::Aws {
let role_arn = mz
.spec
.environmentd_iam_role_arn
.as_deref()
.or(config.aws_info.environmentd_iam_role_arn.as_deref())
.unwrap()
.to_string();
Some(btreemap! {
"eks.amazonaws.com/role-arn".to_string() => role_arn
})
} else {
None
};
ServiceAccount {
metadata: ObjectMeta {
annotations,
..mz.managed_resource_meta(mz.service_account_name())
},
..Default::default()
}
}
fn create_role_object(mz: &Materialize) -> Role {
Role {
metadata: mz.managed_resource_meta(mz.role_name()),
rules: Some(vec![
PolicyRule {
api_groups: Some(vec!["apps".to_string()]),
resources: Some(vec!["statefulsets".to_string()]),
verbs: vec![
"get".to_string(),
"list".to_string(),
"watch".to_string(),
"create".to_string(),
"update".to_string(),
"patch".to_string(),
"delete".to_string(),
],
..Default::default()
},
PolicyRule {
api_groups: Some(vec!["".to_string()]),
resources: Some(vec![
"persistentvolumeclaims".to_string(),
"pods".to_string(),
"secrets".to_string(),
"services".to_string(),
]),
verbs: vec![
"get".to_string(),
"list".to_string(),
"watch".to_string(),
"create".to_string(),
"update".to_string(),
"patch".to_string(),
"delete".to_string(),
],
..Default::default()
},
PolicyRule {
api_groups: Some(vec!["".to_string()]),
resources: Some(vec!["configmaps".to_string()]),
verbs: vec!["get".to_string()],
..Default::default()
},
PolicyRule {
api_groups: Some(vec!["materialize.cloud".to_string()]),
resources: Some(vec!["vpcendpoints".to_string()]),
verbs: vec![
"get".to_string(),
"list".to_string(),
"watch".to_string(),
"create".to_string(),
"update".to_string(),
"patch".to_string(),
"delete".to_string(),
],
..Default::default()
},
PolicyRule {
api_groups: Some(vec!["metrics.k8s.io".to_string()]),
resources: Some(vec!["pods".to_string()]),
verbs: vec!["get".to_string(), "list".to_string()],
..Default::default()
},
PolicyRule {
api_groups: Some(vec!["custom.metrics.k8s.io".to_string()]),
resources: Some(vec![
"persistentvolumeclaims/kubelet_volume_stats_used_bytes".to_string(),
"persistentvolumeclaims/kubelet_volume_stats_capacity_bytes".to_string(),
]),
verbs: vec!["get".to_string()],
..Default::default()
},
]),
}
}
fn create_role_binding_object(mz: &Materialize) -> RoleBinding {
RoleBinding {
metadata: mz.managed_resource_meta(mz.role_binding_name()),
role_ref: RoleRef {
api_group: "".to_string(),
kind: "Role".to_string(),
name: mz.role_name(),
},
subjects: Some(vec![Subject {
api_group: Some("".to_string()),
kind: "ServiceAccount".to_string(),
name: mz.service_account_name(),
namespace: Some(mz.namespace()),
}]),
}
}
fn create_public_service_object(
config: &super::Args,
mz: &Materialize,
generation: u64,
) -> Service {
create_base_service_object(config, mz, generation, &mz.environmentd_service_name())
}
fn create_generation_service_object(
config: &super::Args,
mz: &Materialize,
generation: u64,
) -> Service {
create_base_service_object(
config,
mz,
generation,
&mz.environmentd_generation_service_name(generation),
)
}
fn create_base_service_object(
config: &super::Args,
mz: &Materialize,
generation: u64,
service_name: &str,
) -> Service {
let ports = vec![
ServicePort {
port: config.environmentd_sql_port,
protocol: Some("TCP".to_string()),
name: Some("sql".to_string()),
..Default::default()
},
ServicePort {
port: config.environmentd_http_port,
protocol: Some("TCP".to_string()),
name: Some("https".to_string()),
..Default::default()
},
ServicePort {
port: config.environmentd_internal_sql_port,
protocol: Some("TCP".to_string()),
name: Some("internal-sql".to_string()),
..Default::default()
},
ServicePort {
port: config.environmentd_internal_http_port,
protocol: Some("TCP".to_string()),
name: Some("internal-http".to_string()),
..Default::default()
},
ServicePort {
port: config.environmentd_balancer_sql_port,
protocol: Some("TCP".to_string()),
name: Some("balancer-sql".to_string()),
..Default::default()
},
ServicePort {
port: config.environmentd_balancer_http_port,
protocol: Some("TCP".to_string()),
name: Some("balancer-http".to_string()),
..Default::default()
},
];
let selector = btreemap! {"materialize.cloud/name".to_string() => mz.environmentd_statefulset_name(generation)};
let spec = if config.local_development {
ServiceSpec {
type_: Some("NodePort".to_string()),
selector: Some(selector),
ports: Some(ports),
external_traffic_policy: Some("Local".to_string()),
..Default::default()
}
} else {
ServiceSpec {
type_: Some("ClusterIP".to_string()),
cluster_ip: Some("None".to_string()),
selector: Some(selector),
ports: Some(ports),
..Default::default()
}
};
Service {
metadata: mz.managed_resource_meta(service_name.to_string()),
spec: Some(spec),
status: None,
}
}
fn create_persist_pubsub_service(
config: &super::Args,
mz: &Materialize,
generation: u64,
) -> Service {
Service {
metadata: mz.managed_resource_meta(mz.persist_pubsub_service_name(generation)),
spec: Some(ServiceSpec {
type_: Some("ClusterIP".to_string()),
cluster_ip: Some("None".to_string()),
selector: Some(btreemap! {
"materialize.cloud/name".to_string() => mz.environmentd_statefulset_name(generation),
}),
ports: Some(vec![ServicePort {
name: Some("grpc".to_string()),
protocol: Some("TCP".to_string()),
port: config.environmentd_internal_persist_pubsub_port,
..Default::default()
}]),
..Default::default()
}),
status: None,
}
}
fn create_environmentd_statefulset_object(
config: &super::Args,
tracing: &TracingCliArgs,
mz: &Materialize,
generation: u64,
) -> StatefulSet {
let mut env = vec![
EnvVar {
name: "MZ_METADATA_BACKEND_URL".to_string(),
value_from: Some(EnvVarSource {
secret_key_ref: Some(SecretKeySelector {
name: Some(mz.backend_secret_name()),
key: "metadata_backend_url".to_string(),
optional: Some(false),
}),
..Default::default()
}),
..Default::default()
},
EnvVar {
name: "MZ_PERSIST_BLOB_URL".to_string(),
value_from: Some(EnvVarSource {
secret_key_ref: Some(SecretKeySelector {
name: Some(mz.backend_secret_name()),
key: "persist_backend_url".to_string(),
optional: Some(false),
}),
..Default::default()
}),
..Default::default()
},
];
env.push(EnvVar {
name: "AWS_REGION".to_string(),
value: Some(config.region.clone()),
..Default::default()
});
if config.local_development {
env.push(EnvVar {
name: "AWS_ACCESS_KEY_ID".to_string(),
value: Some("dummy".to_string()),
..Default::default()
});
env.push(EnvVar {
name: "AWS_SECRET_ACCESS_KEY".to_string(),
value: Some("Dummy".to_string()),
..Default::default()
});
}
env.extend(mz.spec.environmentd_extra_env.iter().flatten().cloned());
let mut args = vec![];
if let Some(helm_chart_version) = &config.helm_chart_version {
args.push(format!("--helm-chart-version={helm_chart_version}"));
}
args.push(format!(
"--environment-id={}",
mz.environment_id(&config.cloud_provider.to_string(), &config.region)
));
args.push(format!(
"--clusterd-image={}",
matching_image_from_environmentd_image_ref(
&mz.spec.environmentd_image_ref,
"clusterd",
None
)
));
args.extend(
[
config
.environmentd_cluster_replica_sizes
.as_ref()
.map(|sizes| format!("--cluster-replica-sizes={sizes}")),
config
.bootstrap_default_cluster_replica_size
.as_ref()
.map(|size| format!("--bootstrap-default-cluster-replica-size={size}")),
config
.bootstrap_builtin_system_cluster_replica_size
.as_ref()
.map(|size| format!("--bootstrap-builtin-system-cluster-replica-size={size}")),
config
.bootstrap_builtin_probe_cluster_replica_size
.as_ref()
.map(|size| format!("--bootstrap-builtin-probe-cluster-replica-size={size}")),
config
.bootstrap_builtin_support_cluster_replica_size
.as_ref()
.map(|size| format!("--bootstrap-builtin-support-cluster-replica-size={size}")),
config
.bootstrap_builtin_catalog_server_cluster_replica_size
.as_ref()
.map(|size| {
format!("--bootstrap-builtin-catalog-server-cluster-replica-size={size}")
}),
config
.bootstrap_builtin_analytics_cluster_replica_size
.as_ref()
.map(|size| format!("--bootstrap-builtin-analytics-cluster-replica-size={size}")),
]
.into_iter()
.flatten(),
);
args.extend([
format!("--sql-listen-addr=0.0.0.0:{}", config.environmentd_sql_port),
format!(
"--http-listen-addr=0.0.0.0:{}",
config.environmentd_http_port
),
format!(
"--internal-sql-listen-addr=0.0.0.0:{}",
config.environmentd_internal_sql_port
),
format!(
"--internal-http-listen-addr=0.0.0.0:{}",
config.environmentd_internal_http_port
),
format!(
"--balancer-sql-listen-addr=0.0.0.0:{}",
config.environmentd_balancer_sql_port
),
format!(
"--balancer-http-listen-addr=0.0.0.0:{}",
config.environmentd_balancer_http_port
),
]);
args.extend(
config
.environmentd_allowed_origins
.iter()
.map(|origin| format!("--cors-allowed-origin={}", origin.to_str().unwrap())),
);
args.push(format!(
"--secrets-controller={}",
config.secrets_controller
));
if config.local_development {
args.extend([
"--system-parameter-default=cluster_enable_topology_spread=false".into(),
"--system-parameter-default=log_filter=mz_pgwire[{conn_uuid}]=debug,mz_server_core[{conn_uuid}]=debug,info".into(),
]);
}
if config.enable_tls {
unimplemented!();
} else {
args.push("--tls-mode=disable".to_string());
}
args.push("--persist-isolated-runtime-threads=-1".to_string());
if config.cloud_provider == CloudProvider::Aws {
if let Some(azs) = config.aws_info.environmentd_availability_zones.as_ref() {
for az in azs {
args.push(format!("--availability-zone={az}"));
}
}
if let Some(environmentd_connection_role_arn) = mz
.spec
.environmentd_connection_role_arn
.as_deref()
.or(config.aws_info.environmentd_connection_role_arn.as_deref())
{
args.push(format!(
"--aws-connection-role-arn={}",
environmentd_connection_role_arn
));
}
if let Some(account_id) = &config.aws_info.aws_account_id {
args.push(format!("--aws-account-id={account_id}"));
}
args.extend([format!(
"--aws-secrets-controller-tags=Environment={}",
mz.name_unchecked()
)]);
args.extend_from_slice(&config.aws_info.aws_secrets_controller_tags);
}
args.extend([
"--orchestrator=kubernetes".into(),
format!(
"--orchestrator-kubernetes-service-account={}",
&mz.service_account_name()
),
format!(
"--orchestrator-kubernetes-image-pull-policy={}",
config.image_pull_policy.as_kebab_case_str(),
),
]);
for selector in &config.clusterd_node_selector {
args.push(format!(
"--orchestrator-kubernetes-service-node-selector={}={}",
selector.key, selector.value,
));
}
if let Some(scheduler_name) = &config.scheduler_name {
args.push(format!(
"--orchestrator-kubernetes-scheduler-name={}",
scheduler_name
));
}
for (key, val) in mz.default_labels() {
args.push(format!(
"--orchestrator-kubernetes-service-label={key}={val}"
));
}
if let Some(status) = &mz.status {
args.push(format!(
"--orchestrator-kubernetes-name-prefix=mz{}-",
status.resource_id
));
}
args.extend(["--log-format=json".into()]);
if let Some(endpoint) = &tracing.opentelemetry_endpoint {
args.extend([
format!("--opentelemetry-endpoint={}", endpoint),
format!(
"--opentelemetry-resource=environment_name={}",
mz.name_unchecked()
),
]);
}
let mut volumes = None;
let mut volume_mounts = None;
if let Some(ephemeral_volume_class) = &config.ephemeral_volume_class {
args.extend([
format!(
"--orchestrator-kubernetes-ephemeral-volume-class={}",
ephemeral_volume_class
),
"--scratch-directory=/scratch".to_string(),
]);
volumes = Some(vec![Volume {
name: "scratch".to_string(),
ephemeral: Some(EphemeralVolumeSource {
volume_claim_template: Some(PersistentVolumeClaimTemplate {
spec: PersistentVolumeClaimSpec {
access_modes: Some(vec!["ReadWriteOnce".to_string()]),
storage_class_name: Some(ephemeral_volume_class.to_string()),
resources: Some(VolumeResourceRequirements {
requests: Some(BTreeMap::from([(
"storage".to_string(),
mz.environmentd_scratch_volume_storage_requirement(),
)])),
..Default::default()
}),
..Default::default()
},
..Default::default()
}),
..Default::default()
}),
..Default::default()
}]);
volume_mounts = Some(vec![VolumeMount {
name: "scratch".to_string(),
mount_path: "/scratch".to_string(),
..Default::default()
}]);
}
args.push("--orchestrator-kubernetes-service-fs-group=999".to_string());
if let Some(sentry_dsn) = &tracing.sentry_dsn {
args.push(format!("--sentry-dsn={}", sentry_dsn));
if let Some(sentry_environment) = &tracing.sentry_environment {
args.push(format!("--sentry-environment={}", sentry_environment));
}
args.push(format!("--sentry-tag=region={}", config.region));
}
args.push(format!(
"--persist-pubsub-url=http://{}:{}",
mz.persist_pubsub_service_name(generation),
config.environmentd_internal_persist_pubsub_port,
));
args.push(format!(
"--internal-persist-pubsub-listen-addr=0.0.0.0:{}",
config.environmentd_internal_persist_pubsub_port
));
args.push(format!("--deploy-generation={}", generation));
args.push(format!(
"--internal-console-redirect-url={}",
&config.internal_console_proxy_url,
));
if let Some(extra_args) = &mz.spec.environmentd_extra_args {
args.extend(extra_args.iter().cloned());
}
let probe = Probe {
initial_delay_seconds: Some(1),
failure_threshold: Some(12),
tcp_socket: Some(TCPSocketAction {
host: None,
port: IntOrString::Int(config.environmentd_sql_port),
}),
..Default::default()
};
let security_context = if config.enable_security_context {
Some(SecurityContext {
run_as_non_root: Some(true),
capabilities: Some(Capabilities {
drop: Some(vec!["ALL".to_string()]),
..Default::default()
}),
seccomp_profile: Some(SeccompProfile {
type_: "RuntimeDefault".to_string(),
..Default::default()
}),
allow_privilege_escalation: Some(false),
..Default::default()
})
} else {
None
};
let ports = vec![
ContainerPort {
container_port: config.environmentd_sql_port,
name: Some("sql".to_owned()),
..Default::default()
},
ContainerPort {
container_port: config.environmentd_internal_sql_port,
name: Some("internal-sql".to_owned()),
..Default::default()
},
ContainerPort {
container_port: config.environmentd_balancer_sql_port,
name: Some("balancer-sql".to_owned()),
..Default::default()
},
ContainerPort {
container_port: config.environmentd_http_port,
name: Some("http".to_owned()),
..Default::default()
},
ContainerPort {
container_port: config.environmentd_internal_http_port,
name: Some("internal-http".to_owned()),
..Default::default()
},
ContainerPort {
container_port: config.environmentd_balancer_http_port,
name: Some("balancer-http".to_owned()),
..Default::default()
},
ContainerPort {
container_port: config.environmentd_internal_persist_pubsub_port,
name: Some("persist-pubsub".to_owned()),
..Default::default()
},
];
let container = Container {
name: "environmentd".to_owned(),
image: Some(mz.spec.environmentd_image_ref.to_owned()),
image_pull_policy: Some(config.image_pull_policy.to_string()),
ports: Some(ports),
args: Some(args),
env: Some(env),
volume_mounts,
liveness_probe: Some(probe.clone()),
readiness_probe: Some(probe),
resources: mz.spec.environmentd_resource_requirements.clone(),
security_context: security_context.clone(),
..Default::default()
};
let mut pod_template_labels = mz.default_labels();
pod_template_labels.insert(
"materialize.cloud/name".to_owned(),
mz.environmentd_statefulset_name(generation),
);
pod_template_labels.insert(
"materialize.cloud/app".to_owned(),
mz.environmentd_app_name(),
);
pod_template_labels.insert("app".to_owned(), "environmentd".to_string());
let annotations = btreemap! {
"cluster-autoscaler.kubernetes.io/safe-to-evict".to_owned() => "false".to_string(),
"karpenter.sh/do-not-evict".to_owned() => "true".to_string(),
"karpenter.sh/do-not-disrupt".to_owned() => "true".to_string(),
"materialize.cloud/generation".to_owned() => generation.to_string(),
};
let tolerations = Some(vec![
Toleration {
effect: Some("NoExecute".into()),
key: Some("node.kubernetes.io/not-ready".into()),
operator: Some("Exists".into()),
toleration_seconds: Some(30),
value: None,
},
Toleration {
effect: Some("NoExecute".into()),
key: Some("node.kubernetes.io/unreachable".into()),
operator: Some("Exists".into()),
toleration_seconds: Some(30),
value: None,
},
]);
let pod_template_spec = PodTemplateSpec {
metadata: Some(ObjectMeta {
labels: Some(pod_template_labels),
annotations: Some(annotations), ..Default::default()
}),
spec: Some(PodSpec {
containers: vec![container],
node_selector: Some(
config
.environmentd_node_selector
.iter()
.map(|selector| (selector.key.clone(), selector.value.clone()))
.collect(),
),
scheduler_name: config.scheduler_name.clone(),
service_account_name: Some(mz.service_account_name()),
volumes,
security_context: Some(PodSecurityContext {
fs_group: Some(999),
run_as_user: Some(999),
run_as_group: Some(999),
..Default::default()
}),
tolerations,
termination_grace_period_seconds: Some(0),
..Default::default()
}),
};
let mut match_labels = BTreeMap::new();
match_labels.insert(
"materialize.cloud/name".to_owned(),
mz.environmentd_statefulset_name(generation),
);
let statefulset_spec = StatefulSetSpec {
replicas: Some(1),
template: pod_template_spec,
update_strategy: Some(StatefulSetUpdateStrategy {
rolling_update: None,
type_: Some("OnDelete".to_owned()),
}),
service_name: mz.environmentd_service_name(),
selector: LabelSelector {
match_expressions: None,
match_labels: Some(match_labels),
},
..Default::default()
};
StatefulSet {
metadata: ObjectMeta {
annotations: Some(btreemap! {
"materialize.cloud/generation".to_owned() => generation.to_string(),
"materialize.cloud/force".to_owned() => mz.spec.force_rollout.to_string(),
}),
..mz.managed_resource_meta(mz.environmentd_statefulset_name(generation))
},
spec: Some(statefulset_spec),
status: None,
}
}
fn create_balancerd_deployment_object(config: &super::Args, mz: &Materialize) -> Deployment {
let security_context = if config.enable_security_context {
Some(SecurityContext {
run_as_non_root: Some(true),
capabilities: Some(Capabilities {
drop: Some(vec!["ALL".to_string()]),
..Default::default()
}),
seccomp_profile: Some(SeccompProfile {
type_: "RuntimeDefault".to_string(),
..Default::default()
}),
allow_privilege_escalation: Some(false),
..Default::default()
})
} else {
None
};
let mut pod_template_labels = mz.default_labels();
pod_template_labels.insert(
"materialize.cloud/name".to_owned(),
mz.balancerd_deployment_name(),
);
pod_template_labels.insert("app".to_owned(), "balancerd".to_string());
pod_template_labels.insert("materialize.cloud/app".to_owned(), mz.balancerd_app_name());
let ports = vec![
ContainerPort {
container_port: config.balancerd_sql_port,
name: Some("pgwire".into()),
protocol: Some("TCP".into()),
..Default::default()
},
ContainerPort {
container_port: config.balancerd_http_port,
name: Some("http".into()),
protocol: Some("TCP".into()),
..Default::default()
},
ContainerPort {
container_port: config.balancerd_internal_http_port,
name: Some("internal-http".into()),
protocol: Some("TCP".into()),
..Default::default()
},
];
let mut args = vec![
"service".to_string(),
format!("--pgwire-listen-addr=0.0.0.0:{}", config.balancerd_sql_port),
format!("--https-listen-addr=0.0.0.0:{}", config.balancerd_http_port),
format!(
"--internal-http-listen-addr=0.0.0.0:{}",
config.balancerd_internal_http_port
),
format!(
"--https-resolver-template={}.{}.svc.cluster.local:{}",
mz.environmentd_service_name(),
mz.namespace(),
config.environmentd_balancer_http_port
),
format!(
"--static-resolver-addr={}.{}.svc.cluster.local:{}",
mz.environmentd_service_name(),
mz.namespace(),
config.environmentd_balancer_sql_port
),
];
if config.enable_tls {
unimplemented!();
} else {
args.push("--tls-mode=disable".to_string());
}
let startup_probe = Probe {
http_get: Some(HTTPGetAction {
port: IntOrString::Int(config.balancerd_internal_http_port),
path: Some("/api/readyz".into()),
..Default::default()
}),
failure_threshold: Some(20),
initial_delay_seconds: Some(3),
period_seconds: Some(3),
success_threshold: Some(1),
timeout_seconds: Some(1),
..Default::default()
};
let readiness_probe = Probe {
http_get: Some(HTTPGetAction {
port: IntOrString::Int(config.balancerd_internal_http_port),
path: Some("/api/readyz".into()),
..Default::default()
}),
failure_threshold: Some(3),
period_seconds: Some(10),
success_threshold: Some(1),
timeout_seconds: Some(1),
..Default::default()
};
let liveness_probe = Probe {
http_get: Some(HTTPGetAction {
port: IntOrString::Int(config.balancerd_internal_http_port),
path: Some("/api/livez".into()),
..Default::default()
}),
failure_threshold: Some(3),
initial_delay_seconds: Some(8),
period_seconds: Some(10),
success_threshold: Some(1),
timeout_seconds: Some(1),
..Default::default()
};
let container = Container {
name: "balancerd".to_owned(),
image: Some(matching_image_from_environmentd_image_ref(
&mz.spec.environmentd_image_ref,
"balancerd",
None,
)),
image_pull_policy: Some(config.image_pull_policy.to_string()),
ports: Some(ports),
args: Some(args),
startup_probe: Some(startup_probe),
readiness_probe: Some(readiness_probe),
liveness_probe: Some(liveness_probe),
resources: mz.spec.balancerd_resource_requirements.clone(),
security_context: security_context.clone(),
..Default::default()
};
let deployment_spec = DeploymentSpec {
replicas: Some(1),
selector: LabelSelector {
match_labels: Some(pod_template_labels.clone()),
..Default::default()
},
strategy: Some(DeploymentStrategy {
rolling_update: Some(RollingUpdateDeployment {
max_surge: Some(IntOrString::String("100%".into())),
..Default::default()
}),
..Default::default()
}),
template: PodTemplateSpec {
metadata: Some(ObjectMeta {
labels: Some(pod_template_labels),
..Default::default()
}),
spec: Some(PodSpec {
containers: vec![container],
node_selector: Some(
config
.balancerd_node_selector
.iter()
.map(|selector| (selector.key.clone(), selector.value.clone()))
.collect(),
),
security_context: Some(PodSecurityContext {
fs_group: Some(999),
run_as_user: Some(999),
run_as_group: Some(999),
..Default::default()
}),
scheduler_name: config.scheduler_name.clone(),
service_account_name: Some(mz.service_account_name()),
..Default::default()
}),
},
..Default::default()
};
Deployment {
metadata: ObjectMeta {
..mz.managed_resource_meta(mz.balancerd_deployment_name())
},
spec: Some(deployment_spec),
status: None,
}
}
fn create_balancerd_service_object(config: &super::Args, mz: &Materialize) -> Service {
let selector =
btreemap! {"materialize.cloud/name".to_string() => mz.balancerd_deployment_name()};
let ports = vec![
ServicePort {
name: Some("http".to_string()),
protocol: Some("TCP".to_string()),
port: config.balancerd_http_port,
target_port: Some(IntOrString::Int(config.balancerd_http_port)),
..Default::default()
},
ServicePort {
name: Some("pgwire".to_string()),
protocol: Some("TCP".to_string()),
port: config.balancerd_sql_port,
target_port: Some(IntOrString::Int(config.balancerd_sql_port)),
..Default::default()
},
];
let spec = if config.local_development {
ServiceSpec {
type_: Some("NodePort".to_string()),
selector: Some(selector),
ports: Some(ports),
external_traffic_policy: Some("Local".to_string()),
..Default::default()
}
} else {
ServiceSpec {
type_: Some("ClusterIP".to_string()),
cluster_ip: Some("None".to_string()),
selector: Some(selector),
ports: Some(ports),
..Default::default()
}
};
Service {
metadata: mz.managed_resource_meta(mz.balancerd_service_name()),
spec: Some(spec),
status: None,
}
}
#[derive(Debug, Deserialize, PartialEq, Eq)]
struct BecomeLeaderResponse {
result: BecomeLeaderResult,
}
#[derive(Debug, Deserialize, PartialEq, Eq)]
enum BecomeLeaderResult {
Success,
Failure { message: String },
}
fn environmentd_internal_http_address(
args: &super::Args,
namespace: &str,
generation_service: &Service,
) -> String {
let host = if let Some(host_override) = &args.environmentd_internal_http_host_override {
host_override.to_string()
} else {
format!(
"{}.{}.svc.cluster.local",
generation_service.name_unchecked(),
namespace,
)
};
format!("{}:{}", host, args.environmentd_internal_http_port)
}
fn statefulset_pod_name(statefulset: &StatefulSet, idx: u64) -> String {
format!("{}-{}", statefulset.name_unchecked(), idx)
}