AWS Lambda Managed Instances Now Supports Rust
AWS Lambda Managed Instances added Rust support in March 2026, extending the production-ready Rust-on-Lambda story that achieved general availability in November 2025. Where the November GA allowed Rust functions to run in standard Lambda execution environments, Managed Instances places those same functions on dedicated Lambda-managed Amazon EC2 instances — combining Lambda's operational simplicity with the resource guarantees and pricing flexibility of EC2.
The key capability difference over standard Lambda is parallel request processing within each execution environment: Managed Instances expose the full capacity of the underlying instance rather than constraining each function invocation to a single-threaded execution context. For Rust workloads in particular, where the language's async runtime (Tokio) can efficiently saturate multiple cores, this means a single Managed Instance can handle concurrent requests without the cold-start overhead that would accompany scaling to multiple standard Lambda functions. AWS pairs this with built-in routing, load balancing, and auto-scaling with no operational overhead — the infrastructure behaves like serverless from the developer's perspective while offering EC2 pricing options including Compute Savings Plans and Reserved Instances.
Rust's memory safety guarantees and zero-cost abstractions make it an attractive option for Lambda Managed Instances specifically for CPU-intensive serverless workloads: image processing, cryptography, data parsing, and high-frequency API backends. The managed instances model removes the previous need to choose between Rust's performance benefits and Lambda's operational model. Rust support for Lambda Managed Instances is available in all AWS Regions where Managed Instances is available, with no changes to the cargo-lambda deployment toolchain.
Read more — AWS
Google Cloud Pub/Sub AI Inference Single Message Transform
Google Cloud added an AI Inference Single Message Transform (SMT) to Pub/Sub, allowing developers to inject Vertex AI model inferences directly into streaming message payloads without deploying intermediate compute — no Cloud Run containers, no API retry logic, no additional egress hops. Documentation was updated April 6, 2026.
The SMT operates at the Pub/Sub subscription level: you configure which Vertex AI endpoint (custom model or a Google/partner foundation model) to call, and Pub/Sub appends the inference result to each message before delivering it downstream. The enrichment model covers real-time sentiment analysis, classification, embedding generation, payload enrichment with contextual data, and any other Vertex AI prediction endpoint. Because the transform runs within the Pub/Sub control plane, it scales automatically with message throughput without any capacity planning on the developer side.
From an architecture standpoint, this eliminates a common pattern of Pub/Sub → Cloud Run → AI call → Pub/Sub publish that teams previously assembled to enrich streaming data. The SMT collapses that into a single subscription configuration, removing operational overhead and reducing end-to-end latency by eliminating round-trip network hops between services. The feature is particularly useful for event-driven architectures where every message requires classification or scoring before reaching downstream consumers — for example, fraud detection pipelines, content moderation, or real-time recommendation feeds. Custom models deployed to Vertex AI endpoints are supported alongside Google's hosted models, so proprietary fine-tuned models work the same way as foundation models.
Read more — Google Cloud Documentation
Deep Dive — Google Cloud Community
Links & Sources
- AWS Lambda Managed Instances now supports Rust — AWS
- AI Inference SMT for Pub/Sub — Google Cloud Documentation
- Deep Dive into Google Cloud Pub/Sub Single Message Transforms and AI Inference — Medium / Google Cloud Community