82% run K8s in prod. GPU scheduling breaks them.
AWS Interconnect GA: Layer 3 Private Multi-Cloud Connectivity
- AWS Interconnect reached GA with two capabilities: Interconnect – multicloud (private VPC-to-VPC links between AWS and other clouds) and Interconnect – last mile (private connections from branch/DC to AWS via telco partners). Traffic stays exclusively on AWS and partner private backbones; IEEE 802.1AE MACsec encrypts physical links.
- Multicloud currently covers AWS ↔ Google Cloud across 5 region pairs (us-east-1, us-west-1, us-west-2, eu-west-2, eu-central-1); Microsoft Azure and OCI support is coming later in 2026. The specification is published on GitHub under Apache 2.0.
- The last-mile option auto-provisions 4 redundant connections across 2 physical locations, auto-configures BGP, enables MACsec and Jumbo Frames by default, and supports 1–100 Gbps without reprovisioning. SLA is 99.99%. Lumen Technologies is the initial partner; AT&T and Megaport are in progress.
EKS Auto Mode Networking: Prefix Delegation Default, DNS FQDN Policies
- An AWS containers deep dive on EKS Auto Mode networking (April 14) details the VPC CNI defaults. Prefix delegation is the default mode — assigning /28 prefixes to ENIs instead of individual IPs. A
c5.4xlargesupports 110 pods vs. 58 in secondary IP mode. Auto Mode falls back to secondary IP if prefix assignment fails (first covered in Issue #3 as an architecture overview; this is the networking-specific operational detail). - DNS Network Policies filter outbound traffic at Layer 7 by FQDN — enabling
allow outbound to api.example.comrules without IP-based allowlists. Admin Network Policies centralize cross-namespace control; Cluster Network Policies set cluster-wide security rules. eBPF-backed Kubernetes Network Policies ship enabled by default. - SNAT is configurable per
NodeClass: random SNAT by default, disableable for full source IP traceability in hybrid/on-prem routing. CNI upgrades are managed by AWS with compatibility validation, automatic rollback, and rolling node replacement.
AKS April GA: One-Command Backup, Pod CIDR Expansion
- AKS one-command backup collapsed an 8-step workflow into a single
az dataprotection enable-backup trigger --datasource-type AzureKubernetesServicecall — replacing the previous 15+ command sequence covering extension install, storage account, vault, policy, trusted access, and backup instance. Four built-in strategies:Week(7-day op store),Month(30-day),DisasterRecovery(7-day op + 90-day vault, cross-region), andCustom(BYO vault/policy). - Pod CIDR expansion in AKS CNI Overlay reached GA on April 7 — clusters can now expand pod address space post-creation without node pool recreation.
- Disable HTTP proxy in AKS also reached GA April 7 — clusters deployed with HTTP proxy can disable it in place, enabling egress simplification as network controls mature. Observability in AKS Namespace and Workload Views is GA, surfacing per-namespace and per-workload metrics natively in the portal.
Kubernetes v1.36 Security Details: Manifest Admission, externalIPs Removal Schedule
- Manifest-based Admission Control Config (alpha) moves
AdmissionConfigfrom etcd to a file-based manifest in kube-apiserver — eliminating the startup gap where admission policies are briefly unavailable at control plane restart, and preventing etcd-stored policy deletion as an attack vector. Removes etcd as a hard dependency for admission. service.spec.externalIPsdeprecation (warning added in rc.0, covered in Issue #4) now has a confirmed removal schedule across v1.40, v1.43, and v1.46. Audit forexternalIPsusage now; the field enables man-in-the-middle attacks on ClusterIP services.- HostNetwork + UserNamespaces alpha — pods with
hostNetwork: truecan now sethostUsers: false. These pods only schedule to nodes explicitly flagged for support; limits container escape blast radius on host-networked pods. Distinct fromUserNamespacesSupportGA (covered in Issue #4), which applies to non-host-networked pods. StrictIPCIDRValidationbeta (on by default): values with leading zeros, IPv4-mapped IPv6 addresses, and other ambiguous forms now hard-error. Check admission webhooks, controllers, and operators that generate or accept CIDR values before upgrading.
CNCF 2026 Survey: 82% Production Kubernetes, GPU Scheduling the Top Friction Point
- The CNCF Annual Cloud Native Survey (April 15) shows 82% of container users run Kubernetes in production — up from two-thirds in 2023. 94% are actively using, piloting, or evaluating it; 98% of organizations employ cloud-native techniques.
- 66% of organizations now run generative AI workloads on Kubernetes (consistent with CNCF data cited at KubeCon EU). GPU scheduling, autoscaling configuration, and shared-cluster governance — policy drift, cost attribution, noisy-neighbor effects — are the top operational friction points as AI workload density increases.
Get Platform and Infra Briefing in your inbox
Subscribe to receive new issues as they're published.