data: {"id":1,"jsonrpc":"2.0","result":{"capabilities":{"completion":true},"tools":[{"description":"Fetches latest AWS news and announcements","name":"aws_news","parameters":{"properties":{"category":{"description":"Optional category to filter news","type":"string"}},"type":"object"}}]}}

data: {"id":0,"jsonrpc":"2.0","result":{"message":{"content":"📢 AWS News Update (2026-04-16T18:27:25+0000)\n\n**Amazon CloudWatch now supports cross-region telemetry auditing and enablement rules**\n\n\u003cp\u003eAmazon CloudWatch now supports auditing telemetry configuration and enabling telemetry from AWS services such as Amazon EC2, Amazon VPC, and AWS CloudTrail across multiple AWS Regions from a single region. Customers can enable the telemetry auditing feature for their account or organization across all supported regions at once and create enablement rules that automatically apply to selected regions or all available regions.\u003c/p\u003e\n\u003cp\u003eWith today's launch, customers can scope enablement rules to specific regions or all supported regions. For example, a central security team can create a single organization-wide enablement rule for VPC Flow Logs that applies across all regions, ensuring consistent telemetry collection for every VPC across every account. Rules configured for all regions automatically expand to include new regions as they become available.\u003c/p\u003e\n\u003cp\u003eCloudWatch's cross-region telemetry configuration and enablement rule is available in all AWS commercial regions. Standard CloudWatch pricing applies for telemetry ingestion. To learn more, visit the \u003ca href=\"https://docs.aws.amazon.com/AmazonCloudWatch/latest/monitoring/telemetry-config-cloudwatch.html\" target=\"_blank\" rel=\"noopener noreferrer\"\u003eAmazon CloudWatch documentation\u003c/a\u003e.\u003c/p\u003e\n\n\nCategory: management-and-governance\nProducts: amazon-cloudwatch","role":"assistant"}}}

data: {"id":0,"jsonrpc":"2.0","result":{"message":{"content":"📢 AWS News Update (2026-04-16T19:52:22+0000)\n\n**AWS Elastic Disaster Recovery is now available in the AWS European Sovereign Cloud**\n\n\u003cp\u003e\u003ca href=\"https://aws.amazon.com/disaster-recovery/\"\u003eAWS Elastic Disaster Recovery\u003c/a\u003e (AWS DRS) is now available in the AWS European Sovereign Cloud, enabling organizations with data sovereignty requirements to protect their mission-critical workloads with disaster recovery on AWS. AWS DRS\u003cbr\u003e\nminimizes downtime and data loss with fast, reliable recovery of on-premises and cloud-based applications using affordable storage, minimal compute, and point-in-time recovery, with Recovery Point Objectives (RPOs) measured in seconds and Recovery Time Objectives (RTOs) typically in minutes.\u003cbr\u003e\n\u003cbr\u003e\nWith AWS DRS, you can recover applications from physical infrastructure, VMware vSphere, Microsoft Hyper-V, and cloud infrastructure. AWS DRS uses a unified process for testing, recovery, and failback for a wide range of applications, including critical databases such as Oracle, MySQL, and SQL Server, and enterprise applications such as SAP.\u003cbr\u003e\n\u003cbr\u003e\nAWS Elastic Disaster Recovery is available in the AWS European Sovereign Cloud (Germany). See the \u003ca href=\"https://aws.amazon.com/about-aws/global-infrastructure/regional-product-services/\"\u003eAWS Regional Services List\u003c/a\u003e for the latest availability information.\u003cbr\u003e\n\u003cbr\u003e\nTo learn more about AWS Elastic Disaster Recovery, visit our \u003ca href=\"https://aws.amazon.com/disaster-recovery/\"\u003eproduct page\u003c/a\u003e or \u003ca href=\"https://docs.aws.eu/drs/latest/userguide/what-is-drs.html\"\u003edocumentation\u003c/a\u003e.\u003c/p\u003e\n\n\nCategory: storage\nProducts: aws-elastic-disaster-recovery","role":"assistant"}}}

data: {"id":0,"jsonrpc":"2.0","result":{"message":{"content":"📢 AWS News Update (2026-03-31T21:55:45+0000)\n\n**Introducing Amazon EC2 C8in and C8ib instances **\n\n\u003cp\u003eAWS is announcing the general availability of Amazon EC2 C8in and C8ib instances powered by custom, sixth generation Intel Xeon Scalable processors, available only on AWS. These instances feature the latest sixth generation AWS Nitro cards. C8in and C8ib instances deliver up to 43% higher performance compared to previous generation C6in instances.\u003cbr\u003e\n\u003cbr\u003e\nC8in and C8ib instances deliver larger sizes and scale up to 384 vCPUs. C8in instances deliver 600 Gbps network bandwidth—the highest among enhanced networking EC2 instances—making them ideal for network-intensive workloads like distributed compute and large-scale data analytics. C8ib instances deliver up to 300 Gbps EBS bandwidth, the highest among non-accelerated compute instances, making them ideal for high-performance commercial databases and file systems.\u003c/p\u003e\n\u003cp\u003eC8in instances are available in US East (N. Virginia), US West (Oregon), Asia Pacific (Tokyo), and Europe (Spain) regions. C8ib instances are available in US East (N. Virginia) and US West (Oregon). Both, C8in and C8ib instances are available via Savings Plans, On-Demand, and Spot instances. For more information, visit the \u003ca href=\"https://aws.amazon.com/ec2/instance-types/c8i/\" target=\"_blank\" rel=\"noopener noreferrer\"\u003eAmazon EC2 C8i instance page\u003c/a\u003e.\u003c/p\u003e\n\n\nCategory: compute\nProducts: amazon-ec2","role":"assistant"}}}

data: {"id":0,"jsonrpc":"2.0","result":{"message":{"content":"📢 AWS News Update (2026-04-09T15:08:39+0000)\n\n** Amazon Quick now supports multi-account sign-in within the same browser**\n\n\u003cp\u003eToday, AWS announces multi-session support for Amazon Quick, which enables customers to access up to five Amazon Quick accounts simultaneously within the same browser. The feature also includes the Amazon Quick account name in all URLs, enabling users to easily access the correct account when opening agents, spaces, flows, research reports, dashboards, and other assets.\u003cbr\u003e\n\u003cbr\u003e\nCustomers use multiple accounts for different environments such as development, testing, and production, and compare insights and resource configurations across multiple accounts for troubleshooting and other application-related jobs. Using multi-session capability in Amazon Quick, customers can now sign in to multiple accounts and manage their resources in a single browser. You can sign in to another account by accessing the Amazon Quick top right menu and selecting the option to sign in to another account. For users accessing global URLs without an account name, Amazon Quick presents an account input page that pre-populates the accounts they are logged into, allowing them to select the desired account. You have the option to log out of the current session in the specific browser tab or log out of all sessions.\u003cbr\u003e\n\u003cbr\u003e\nAmazon Quick multi-account sign-in is available in all \u003ca href=\"https://docs.aws.amazon.com/quicksight/latest/user/regions-qs.html\"\u003esupported Amazon Quick regions\u003c/a\u003e.\u003cbr\u003e\n\u003cbr\u003e\nTo learn more about this, visit \u003ca href=\"https://docs.aws.amazon.com/quicksuite/latest/userguide/signing-in.html\"\u003eAmazon Quick Signing In\u003c/a\u003e\u003c/p\u003e\n\n\nCategory: artificial-intelligence\nProducts: amazon-quick-suite","role":"assistant"}}}

data: {"id":0,"jsonrpc":"2.0","result":{"message":{"content":"📢 AWS News Update (2026-04-15T03:14:39+0000)\n\n**Claude Opus 4.7 is now available in Amazon Bedrock **\n\n\u003cp\u003eAmazon Bedrock, the platform for building AI applications and agents at production scale, now offers Claude Opus 4.7--\u0026nbsp;Anthropic's\u0026nbsp;most capable Opus model to date\u0026nbsp;--\u0026nbsp;delivering meaningful improvements across agentic coding,\u0026nbsp;professional\u0026nbsp;work, and long-running tasks for developers and enterprises building production AI applications.\u0026nbsp;\u003c/p\u003e\n\u003cp\u003eClaude Opus 4.7 is an\u0026nbsp;upgrade from Claude Opus 4.6, with stronger performance across the workflows teams\u0026nbsp;run\u0026nbsp;in production.\u0026nbsp;Opus 4.7 works better through ambiguity, is more thorough in its problem solving, and\u0026nbsp;folllows\u0026nbsp;instructions more precisely.\u0026nbsp;For coding, the model extends agentic capabilities with improved long-horizon autonomy, systems engineering, and complex code reasoning. For knowledge work, Claude Opus 4.7 advances professional tasks such as\u0026nbsp;slides and\u0026nbsp;document creation, financial analysis, and\u0026nbsp;data visualization. For long-running tasks, the model stays on track\u0026nbsp;over longer horizons with improved reasoning and memory capabilities. Claude Opus 4.7 also advances visual\u0026nbsp;capabilities\u0026nbsp;with high-resolution image support\u0026nbsp;improving accuracy on charts, dense documents, and\u0026nbsp;screen\u0026nbsp;UIs where fine detail matters.\u0026nbsp;\u0026nbsp;\u003c/p\u003e\n\u003cp\u003eClaude Opus\u0026nbsp;4.7 is\u0026nbsp;served\u0026nbsp;through\u0026nbsp;Amazon Bedrock's next-generation inference engine, delivering enterprise-grade infrastructure for production workloads.\u0026nbsp;It provides zero operator data access, meaning customer\u0026nbsp;prompts\u0026nbsp;and responses are never visible to Anthropic or AWS operators, keeping sensitive data private. It also enables\u0026nbsp;enhanced\u0026nbsp;availability through\u0026nbsp;dynamic traffic routing with expanded in-region options, along with improved scalability.\u003c/p\u003e\n\u003cp\u003eClaude Opus 4.7 is available in select\u0026nbsp;\u003ca href=\"https://docs.aws.amazon.com/bedrock/latest/userguide/models-region-compatibility.html\" target=\"_blank\" rel=\"noopener noreferrer\"\u003eAWS Regions\u003c/a\u003e.\u0026nbsp;To learn more about Claude Opus 4.7 and other\u0026nbsp;Anthropic\u0026nbsp;models available\u0026nbsp;in\u0026nbsp;Amazon Bedrock, visit the\u0026nbsp;\u003ca href=\"https://aws.amazon.com/bedrock/anthropic/\" target=\"_blank\" rel=\"noopener noreferrer\"\u003eAmazon Bedrock\u0026nbsp;page\u003c/a\u003e. To get started, see the\u0026nbsp;\u003ca href=\"https://docs.aws.amazon.com/bedrock/latest/userguide/what-is-bedrock.html\" target=\"_blank\" rel=\"noopener noreferrer\"\u003eAmazon Bedrock documentation\u003c/a\u003e.\u0026nbsp;\u003c/p\u003e\n\n\nCategory: artificial-intelligence\nProducts: amazon-bedrock","role":"assistant"}}}

data: {"id":0,"jsonrpc":"2.0","result":{"message":{"content":"📢 AWS News Update (2026-04-15T20:38:56+0000)\n\n**AWS Payment Cryptography now available in South America (São Paulo)**\n\n\u003cp\u003eAWS Payment Cryptography has expanded its global presence with availability in South America (São Paulo). This expansion enables customers with latency-sensitive payment applications to build, deploy or migrate into additional AWS Regions without depending on cross-region support.\u003c/p\u003e\n\u003cp\u003eAWS Payment Cryptography is a fully managed service that simplifies payment-specific cryptographic operations and key management for cloud-hosted payment applications. The service scales elastically with your business needs and is assessed as compliant with PCI PIN and PCI P2PE requirements, eliminating the need to maintain dedicated payment HSM instances. Organizations performing payment functions - including acquirers, payment facilitators, networks, switches, processors, and banks can now position their payment cryptographic operations closer to their applications while reducing dependencies on auxiliary data centers with dedicated payment HSMs.\u003c/p\u003e\n\u003cp\u003eAWS Payment Cryptography is available in the following AWS Regions: Canada (Montreal), US East (Ohio, N. Virginia), US West (Oregon), Europe (Ireland, Frankfurt, London, Paris), South America (São Paulo), Africa (Cape Town) and Asia Pacific (Singapore, Tokyo, Osaka, Mumbai, Hyderabad).\u003c/p\u003e\n\u003cp\u003eTo start using the service, please download the latest \u003ca href=\"https://builder.aws.com/build/tools\"\u003eAWS CLI/SDK\u003c/a\u003e and see the AWS Payment Cryptography \u003ca href=\"https://docs.aws.amazon.com/payment-cryptography/latest/userguide/what-is.html\"\u003euser guide\u003c/a\u003e for more information.\u003c/p\u003e\n\n\nCategory: security-identity-and-compliance\nProducts: aws-paymemt-cryptography","role":"assistant"}}}

data: {"id":0,"jsonrpc":"2.0","result":{"message":{"content":"📢 AWS News Update (2026-04-03T16:52:06+0000)\n\n**Amazon Quick Introduces Sheet Tooltips for Rich, Contextual Data Exploration**\n\n\u003cp\u003e\u003ca href=\"https://aws.amazon.com/quick/quicksight/\" target=\"_blank\" rel=\"noopener noreferrer\"\u003eQuick Sight in Amazon Quick\u003c/a\u003e now supports sheet tooltips, enabling authors to surface rich, contextual detail when viewers hover over data points — without disrupting their analysis flow. Sheet tooltips allow authors to create dedicated tooltip sheets containing visuals, text boxes, and images arranged in a free-form layout. When a viewer hovers over a data point, the tooltip sheet automatically inherits all filters from the source visual and applies an additional filter for the specific data point, delivering an instant, focused breakdown.\u003c/p\u003e\n\u003cp\u003eThis enhancement helps organizations build more intuitive dashboards that reduce the need for multiple sheets or manual navigation. For example, a bar chart showing sales by product category can surface a trend line of monthly sales, a year-over-year growth KPI, and a text box with the category name — all filtered to whichever category the viewer hovers over. Authors can assign one tooltip sheet to multiple visuals, switch between basic, detailed, and sheet tooltip types at any time, and tables and pivot tables are also supported. Sheet tooltips are available on interactive sheets only.\u003c/p\u003e\n\u003cp\u003eThis feature is now available in all Amazon Quick \u003ca href=\"https://docs.aws.amazon.com/quicksuite/latest/userguide/regions.html\" target=\"_blank\" rel=\"noopener noreferrer\"\u003eregions where Quick Sight is supported.\u003c/a\u003e Learn more about how to \u003ca href=\"https://docs.aws.amazon.com/quick/latest/userguide/customizing-visual-tooltips.html#customizing-visual-tooltips-sheet\" target=\"_blank\" rel=\"noopener noreferrer\"\u003euse sheet tooltips\u003c/a\u003e in Amazon Quick and read more about this new feature in \u003ca href=\"https://aws.amazon.com/blogs/machine-learning/create-rich-custom-tooltips-in-amazon-quick-sight/\" target=\"_blank\" rel=\"noopener noreferrer\"\u003eour blog post\u003c/a\u003e.\u003c/p\u003e\n\n\nCategory: analytics\nProducts: amazon-quick-suite, amazon-quicksight","role":"assistant"}}}

data: {"id":0,"jsonrpc":"2.0","result":{"message":{"content":"📢 AWS News Update (2026-04-14T21:46:36+0000)\n\n**Amazon EC2 P6-B300 instances are now available in the AWS GovCloud (US-East) Region**\n\n\u003cp\u003eStarting today, Amazon Elastic Cloud Compute (Amazon EC2) P6-B300 instances are available in the AWS GovCloud (US-East) Region. P6-B300 instances provide 8x NVIDIA Blackwell Ultra GPUs with 2.1 TB high bandwidth GPU memory, 6.4 Tbps EFA networking, 300 Gbps dedicated ENA throughput, and 4 TB of system memory.\u003cbr\u003e\n\u003cbr\u003e\nP6-B300 instances deliver 2x networking bandwidth, 1.5x GPU memory size, and 1.5x GPU TFLOPS (at FP4, without sparsity) compared to P6-B200 instances, making them well suited to train and deploy large trillion-parameter foundation models (FMs) and large language models (LLMs) with sophisticated techniques. The higher networking and larger memory deliver faster training times and more token throughput for AI workloads. \u003cbr\u003e\n\u003cbr\u003e\nP6-B300 instances are now available in p6-b300.48xlarge size in the following AWS Regions: US West (Oregon) and AWS GovCloud (US-East). To learn more about P6-B300 instances, visit \u003ca href=\"https://aws.amazon.com/ec2/instance-types/p6/\" target=\"_blank\"\u003eAmazon EC2 P6 instances\u003c/a\u003e.\u003c/p\u003e\n\n\nCategory: artificial-intelligence\nProducts: aws-govcloud-us, amazon-ec2","role":"assistant"}}}

data: {"id":0,"jsonrpc":"2.0","result":{"message":{"content":"📢 AWS News Update (2026-04-14T00:17:40+0000)\n\n**AWS Transform is now available in Kiro and VS Code**\n\n\u003cp\u003eAWS Transform is now available through two additional developer tools — including Kiro and VS Code. \u003ca href=\"https://aws.amazon.com/transform\"\u003eAWS Transform\u003c/a\u003e is an agentic migration and modernization factory designed to compress enterprise transformation timelines from years to months — handling everything from large-scale infrastructure migrations to continuous tech debt reduction, without the manual handoffs and lost context that commonly stall these programs..\u003c/p\u003e\n\u003cp\u003eWith today’s launch, you can get started with \u003ca href=\"https://aws.amazon.com/transform/custom/\"\u003eAWS Transform custom\u003c/a\u003e transformations from wherever you already work: install the AWS Transform Power in \u003ca href=\"https://github.com/kirodotdev/powers/tree/main/aws-transform\"\u003eKiro\u003c/a\u003e, or install the AWS Transform extension in \u003ca href=\"https://marketplace.visualstudio.com/items?itemName=AmazonWebServices.aws-transform-plugin\"\u003eVS Code\u003c/a\u003e . AWS Transform custom transformations help you crush tech debt at scale — choose from AWS-managed transformations for common patterns like Java, Python, and Node.js version upgrades, AWS SDK migrations (boto2 to boto3, Java SDK v1 to v2, JS SDK v2 to v3), or define your own. These new surfaces make it easier to discover additional capabilities as they become available, build and iterate on your own custom transformations, and run any agent repeatedly or across thousands of repositories at once. The custom transformations are the first in a growing library of playbooks coming to developer tools, complementing the existing AWS Transform web console and CLI so you can start a job in your IDE, track progress in the web console, and finish transformations wherever it makes sense — with job state and context shared across every surface.\u003c/p\u003e\n\u003cp\u003eAWS Transform supports deploying to all AWS commercial regions,and AWS Transform custom is available in US East (N. Virginia) and Europe (Frankfurt). To learn more, visit the AWS Transform \u003ca href=\"https://aws.amazon.com/transform\"\u003eproduct page\u003c/a\u003e and \u003ca href=\"https://docs.aws.amazon.com/transform/latest/userguide/what-is-service.html\"\u003euser guide\u003c/a\u003e.\u003c/p\u003e\n\n\nCategory: migration\nProducts: aws-transform ","role":"assistant"}}}

data: {"id":0,"jsonrpc":"2.0","result":{"message":{"content":"📢 AWS News Update (2026-04-14T20:12:44+0000)\n\n**AWS Secrets Manager now supports hybrid post-quantum TLS to protect secrets from quantum threats **\n\n\u003cp\u003eAWS Secrets Manager now\u0026nbsp;supports\u0026nbsp;hybrid post-quantum key exchange using ML-KEM (Module-Lattice-based Key-Encapsulation Mechanism) to secure TLS connections for retrieving and managing secrets. This protection is automatically enabled in Secrets Manager Agent (version 2.0.0+), AWS Lambda Extension (version 19+), and Secrets Manager CSI Driver (version 2.0.0+). For SDK-based clients, hybrid post-quantum key exchange is available in supported AWS SDKs including Rust, Go, Node.js, Kotlin, Python (with OpenSSL 3.5+), and Java v2 (v2.35.11+).\u003c/p\u003e\n\u003cp\u003eWith this launch, your applications retrieve secrets over TLS connections that combine classical key exchange with post-quantum cryptography, helping protect against both traditional cryptographic attacks and future quantum computing threats known as \u0026quot;harvest now, decrypt later\u0026quot; (HNDL). No code changes, configuration updates, or migration effort are\u0026nbsp;required\u0026nbsp;for customers using the latest client versions\u0026nbsp;except for Java v2. For example, a microservice requiring multiple secrets at startup can now retrieve them over quantum-resistant TLS connections by simply upgrading to the latest Secrets Manager Agent version. You can verify hybrid post-quantum key exchange is active by checking CloudTrail logs for the \u0026quot;X25519MLKEM768\u0026quot; key exchange algorithm in the\u0026nbsp;tlsDetails\u0026nbsp;field of\u0026nbsp;GetSecretValue\u0026nbsp;API calls.\u003c/p\u003e\n\u003cp\u003eHybrid post-quantum key exchange using ML-KEM for AWS Secrets Manager is available in all AWS Regions where AWS Secrets Manager is supported. To learn more, visit the\u0026nbsp;\u003ca href=\"https://docs.aws.amazon.com/secretsmanager/latest/userguide/intro.html\"\u003eAWS Secrets Manager documentation\u003c/a\u003e\u0026nbsp;and the\u0026nbsp;\u003ca href=\"https://aws.amazon.com/security/post-quantum-cryptography/migrating-to-post-quantum-cryptography/\"\u003eAWS Post-Quantum Cryptography migration page\u003c/a\u003e.\u003c/p\u003e\n\n\nCategory: TBD\nProducts: aws-secrets-manager","role":"assistant"}}}

data: {"id":0,"jsonrpc":"2.0","result":{"message":{"content":"📢 AWS News Update (2026-04-08T04:29:42+0000)\n\n**AWS announces general availability of AWS Interconnect - multicloud**\n\n\u003cp\u003eAWS announces general availability (GA) of AWS Interconnect - multicloud, providing simple, resilient, high-speed private connections to other cloud service providers (CSPs). With GA comes Google Cloud as the first launch partner, with Microsoft Azure and Oracle Cloud Infrastructure (OCI) coming later in 2026.\u003c/p\u003e\n\u003cp\u003eCustomers have been adopting multicloud strategies while migrating more applications to the cloud. They do so for many reasons including interoperability requirements, the freedom to choose technology that best suits their needs, and the ability to build and deploy applications on any environment with greater ease and speed. Previously, when interconnecting workloads across multiple cloud providers, customers had to go the route of a ‘do-it-yourself’ multicloud approach, leading to complexities of managing global multi-layered networks at scale. AWS Interconnect - multicloud is the first purpose-built product of its kind and a new way of how clouds connect and talk to each other. Simplifying connectivity into AWS, Interconnect - multicloud enables customers to quickly establish private, secure, high-speed network connections with dedicated bandwidth and built-in resiliency between their Amazon VPCs and other cloud environments. Interconnect - multicloud makes it easy to connect AWS resources or VPCs to other CSPs. Customers can also quickly scale connectivity to multiple VPCs or Regions via associating Interconnect with other networking services such as AWS Transit Gateway and AWS Cloud WAN, instead of taking weeks or months. Interconnect - multicloud introduces a new, single-fee pricing structure based on the customer’s selected bandwidth and the geographical scope of the connectivity to other CSPs. Customers can also use one free, local 500Mbps interconnect per Region starting in May. To learn more please see the Interconnect - multicloud Pricing documentation page.\u003c/p\u003e\n\u003cp\u003eInterconnect - multicloud is available in five AWS Regions. You can enable this capability using the AWS Management Console, Command Line Interface (CLI), or API, and CSPs can also adopt via a published \u003ca href=\"https://github.com/aws/Interconnect\" target=\"_blank\" rel=\"noopener noreferrer\"\u003eopen API package on GitHub\u003c/a\u003e. For more information, see the \u003ca href=\"https://docs.aws.amazon.com/interconnect/\" target=\"_blank\" rel=\"noopener noreferrer\"\u003eAWS Interconnect - multicloud documentation\u003c/a\u003e\u0026nbsp;and \u003ca href=\"https://aws.amazon.com/interconnect/multicloud/pricing\" target=\"_blank\" rel=\"noopener noreferrer\"\u003epricing\u003c/a\u003e pages.\u003c/p\u003e\n\n\nCategory: networking-and-content-delivery\nProducts: aws-direct-connect","role":"assistant"}}}

data: {"id":0,"jsonrpc":"2.0","result":{"message":{"content":"📢 AWS News Update (2026-04-10T19:06:19+0000)\n\n**Amazon EC2 C8gn, M8gn, and R8gn instances now support higher Amazon EBS-optimized performance**\n\n\u003cp\u003eToday, AWS announces increased Amazon Elastic Block Store (Amazon EBS) performance for Amazon EC2 C8gn, M8gn, and R8gn instances in 48xlarge and metal-48xl sizes.\u003c/p\u003e\n\u003cp\u003eEC2 C8gn, M8gn, and R8gn instances are network optimized instances powered by AWS Graviton4 processors and latest 6th generation AWS Nitro Cards. With the latest enhancements to AWS Nitro System, we have doubled the maximum EBS performance on these instances in 48xlarge and metal-48xl sizes, from 60 Gbps of EBS bandwidth and 240,000 IOPS to 120 Gbps of EBS bandwidth and 480,000 IOPS. Customers running network-intensive workloads while requiring additional block storage performance such as data analytics and high-performance file systems can benefit from the improved EBS performance.\u003c/p\u003e\n\u003cp\u003eAll existing and new C8gn, M8gn, and R8gn instances in 48xlarge and metal-48xl sizes launched starting today will benefit from this performance increase at no additional cost. For running instances, customers can stop and start instances to enable this performance increase. The higher EBS performance is available in all AWS regions where these instance types are generally available today.\u003c/p\u003e\n\u003cp\u003eTo learn more, see Amazon \u003ca href=\"https://aws.amazon.com/ec2/instance-types/c8g/\"\u003eC8gn\u003c/a\u003e, \u003ca href=\"https://aws.amazon.com/ec2/instance-types/m8g/\"\u003eM8gn\u003c/a\u003e, and \u003ca href=\"https://aws.amazon.com/ec2/instance-types/r8g\"\u003eR8gn\u003c/a\u003e Instances and \u003ca href=\"https://docs.aws.amazon.com/AWSEC2/latest/UserGuide/ebs-optimized.html\"\u003eEBS-optimized instance types\u003c/a\u003e.\u0026nbsp;\u003c/p\u003e\n\n\nCategory: storage\nProducts: amazon-elastic-block-store","role":"assistant"}}}

data: {"id":0,"jsonrpc":"2.0","result":{"message":{"content":"📢 AWS News Update (2026-04-14T15:54:55+0000)\n\n**NVIDIA Nemotron-3-Super-120B, Qwen3.5-9B, and Qwen3.5-27B models now available on Amazon SageMaker JumpStart**\n\n\u003cp\u003eNVIDIA’s Nemotron-3-Super-120B, Qwen3.5-9B, and Qwen3.5-27B models are now available on Amazon SageMaker JumpStart, expanding the portfolio of foundation models available to AWS customers. These three models bring specialized capabilities spanning agentic reasoning, multilingual coding, and advanced instruction following, enabling customers\u003cb\u003e\u0026nbsp;\u003c/b\u003eto deploy high-performance, scalable AI solutions on AWS infrastructure.\u003cbr\u003e\n\u003cbr\u003e\nThese models address different enterprise AI challenges with specialized capabilities:\u003cbr\u003e\nNemotron-3-Super-120B is optimized for collaborative agents and high-volume workloads such as IT ticket automation. It employs a hybrid Latent Mixture-of-Experts (LatentMoE) architecture with Mamba-2 and MoE layers, enabling strong agentic, reasoning, and conversational capabilities useful for multi-agent applications like software development and cybersecurity triaging.\u003cbr\u003e\nQwen 3.5 9B excels in multilingual coding, instruction following, and long-horizon planning, automating software development workflows and executing complex, multi-step office tasks. Its compact design balances efficiency and performance for resource-constrained environments.\u003cbr\u003e\nQwen 3.5 27B provides deeper contextual understanding, extended reasoning capabilities, and enhanced spatial/complex scenario comprehension, ideal for advanced multimodal reasoning and large-scale document processing.\u003cbr\u003e\nWith SageMaker JumpStart, customers can deploy any of these models with just a few clicks to address their specific AI use cases.\u003cbr\u003e\n\u003cbr\u003e\nTo get started with these models, navigate to the SageMaker JumpStart model catalog in the SageMaker console or use the SageMaker Python SDK to deploy the models to your AWS account. For more information about deploying and using foundation models in SageMaker JumpStart, see the\u0026nbsp;\u003ca href=\"https://docs.aws.amazon.com/sagemaker/latest/dg/studio-jumpstart.html\"\u003eAmazon SageMaker JumpStart documentation\u003c/a\u003e.\u003c/p\u003e\n\n\nCategory: TBD\nProducts: amazon-sagemaker-jumpstart, aiml","role":"assistant"}}}

data: {"id":0,"jsonrpc":"2.0","result":{"message":{"content":"📢 AWS News Update (2026-04-13T20:56:55+0000)\n\n**Amazon OpenSearch Serverless now supports Derived Source for storage optimization **\n\n\u003cp\u003eAmazon OpenSearch Serverless introduces support for Derived Source, a new feature that can help reduce the amount of storage required for your OpenSearch Service collections. With derived source support, you can skip storing source fields and dynamically derive them when required.\u0026nbsp;\u003c/p\u003e\n\u003cp\u003eWith Derived Source, OpenSearch Serverless reconstructs the _source field on the fly using the values already stored in the index, eliminating the need to maintain a separate copy of the original document. This can significantly reduce storage consumption, particularly for time-series and log analytics collections where documents contain many indexed fields. You can enable derived source at the index level when creating or updating index mappings.\u003c/p\u003e\n\u003cp\u003eDerived Source support is available today in all AWS Regions where Amazon OpenSearch Serverless is\u0026nbsp;\u003ca href=\"https://docs.aws.amazon.com/general/latest/gr/opensearch-service.html\" target=\"_blank\"\u003esupported\u003c/a\u003e. For more information, see the Amazon OpenSearch Serverless\u0026nbsp;\u003ca href=\"https://docs.aws.amazon.com/opensearch-service/latest/developerguide/serverless-derived-source.html\" target=\"_blank\"\u003edocumentation\u003c/a\u003e.\u003c/p\u003e\n\n\nCategory: analytics\nProducts: amazon-opensearch-service","role":"assistant"}}}

data: {"id":0,"jsonrpc":"2.0","result":{"message":{"content":"📢 AWS News Update (2026-04-13T22:25:43+0000)\n\n**Amazon Redshift introduces key  performance  optimization for Top-K queries**\n\n\u003cp\u003e\u003ca href=\"https://aws.amazon.com/redshift/\"\u003eAmazon Redshift\u003c/a\u003e further optimizes the processing of top-k queries (queries with ORDER BY and LIMIT clauses) by intelligently skipping irrelevant data blocks to return results faster, dramatically reducing the amount of data processed. This optimization reorders and efficiently adjusts the data blocks to be read based on the ORDER BY column's min/max values, maintaining only the K most qualifying rows in memory. When the ORDER BY column is sorted or partially sorted, Amazon Redshift now processes only the minimal data blocks needed rather than scanning entire tables, eliminating unnecessary I/O and compute overhead.\u003c/p\u003e\n\u003cp\u003eThis enhancement particularly benefits top-k queries when the data permanently stores in descending order (ORDER BY ... DESC LIMIT K) on large tables where qualifying rows are appended at the end of the data storage. Common examples include:\u003cbr\u003e\n\u003c/p\u003e\n\u003cul\u003e\n\u003cli\u003eFinding the k most recent orders from millions or billions of transactions\u003c/li\u003e\n\u003cli\u003eRetrieving top-k best performing products or k worst performing products (top-k in descending order) from your sales catalog containing hundreds of thousands stock keeping units (SKUs) and millions or billions of sales transactions associated with all product SKUs in your sales catalog\u003c/li\u003e\n\u003cli\u003eFinding the top-k most recent or top-k oldest (top k in descending order) prompts inferred by a foundational large language model (LLM) out of billions of prompts.\u003c/li\u003e\n\u003c/ul\u003e\n\u003cp\u003eWith this new optimization, top-k query performance improves dramatically. This optimization for top-k queries is now available in Amazon Redshift at no additional cost starting with \u003ca href=\"https://docs.aws.amazon.com/redshift/latest/mgmt/cluster-versions.html#cluster-version-199\"\u003epatch release P199\u003c/a\u003e across all AWS regions where Amazon Redshift is available. This optimization automatically applies to eligible queries without requiring any query rewrites or configuration changes.\u003c/p\u003e\n\n\nCategory: analytics\nProducts: amazon-redshift","role":"assistant"}}}

data: {"id":0,"jsonrpc":"2.0","result":{"message":{"content":"📢 AWS News Update (2026-04-14T14:06:17+0000)\n\n**Amazon Quick now supports document-level access controls for Google Drive knowledge bases**\n\n\u003cp\u003eAmazon Quick now supports document-level access controls (ACLs) for Google Drive knowledge bases, enabling organizations to maintain native Google Drive permissions when indexing content. Quick combines ACL replication for efficient pre-retrieval filtering with an additional layer of real-time permission checks directly with Google Drive at query time. This dual approach means you get the performance benefits of indexed ACLs while also guarding against stale or incorrectly mapped permission data. When a user submits a query, Quick verifies their current permissions with Google Drive before generating a response—ensuring answers are based on live access rights.\u003cbr\u003e\n\u003cbr\u003e\nWith document-level access controls, Amazon Quick now respects individual file and folder permissions from Google Drive. This feature is available in all AWS \u003ca href=\"https://docs.aws.amazon.com/quick/latest/userguide/regions.html#regions-qs\" target=\"_blank\"\u003eRegions where Amazon Quick is available\u003c/a\u003e. \u003c/p\u003e\n\u003cp\u003eTo get started, create or update a Google Drive knowledge base in the Amazon Quick console and configure document-level access controls in your integration settings. For more information, see \u003ca href=\"https://docs.aws.amazon.com/quick/latest/userguide/google-drive-integration.html\" target=\"_blank\" rel=\"noopener noreferrer\"\u003eGoogle Drive integration\u003c/a\u003e in the Amazon Quick User Guide.\u003c/p\u003e\n\n\nCategory: artificial-intelligence\nProducts: amazon-quick-suite","role":"assistant"}}}

data: {"id":0,"jsonrpc":"2.0","result":{"message":{"content":"📢 AWS News Update (2026-04-13T17:48:17+0000)\n\n**Aurora DSQL launches connector that simplifies building PHP applications**\n\n\u003cp\u003eToday we are announcing the release of the Aurora DSQL Connector for PHP (PDO_PGSQL) that makes it easy to build PHP applications on Aurora DSQL. The PHP Connector streamlines authentication and eliminates security risks associated with traditional user-generated passwords by automatically generating tokens for each connection, ensuring valid tokens are always used while maintaining full compatibility with existing PDO_PGSQL features.\u003cbr\u003e\n\u003cbr\u003e\nThe connector handles IAM token generation, SSL configuration, and connection pooling, enabling customers to scale from simple scripts to production workloads without changing their authentication approach. It also provides opt-in optimistic concurrency control (OCC) retry with exponential backoff, custom IAM credential providers, and AWS profile support, making it easier to develop client retry logic and manage AWS credentials.\u003cbr\u003e\n\u003cbr\u003e\nTo get started, visit the \u003ca href=\"https://docs.aws.amazon.com/aurora-dsql/latest/userguide/SECTION_connectors.html\"\u003eConnectors for Aurora DSQL documentation\u003c/a\u003e page. For code examples, visit our \u003ca href=\"https://github.com/awslabs/aurora-dsql-connectors/blob/main/php/pdo_pgsql/README.md\"\u003eGitHub page for the PHP connector\u003c/a\u003e. Get started with Aurora DSQL for free with the \u003ca href=\"https://aws.amazon.com/free/?all-free-tier.sort-by=item.additionalFields.SortRank\\\u0026amp;all-free-tier.sort-order=asc\\\u0026amp;awsf.Free%20Tier%20Types=*all\\\u0026amp;awsf.Free%20Tier%20Categories=categories%23databases\"\u003eAWS Free Tier\u003c/a\u003e. To learn more about Aurora DSQL, visit the \u003ca href=\"https://aws.amazon.com/rds/aurora/dsql/\"\u003ewebpage\u003c/a\u003e.\u0026nbsp;\u0026nbsp;\u0026nbsp;\u0026nbsp;\u003c/p\u003e\n\n\nCategory: databases\nProducts: amazon-aurora-dsql","role":"assistant"}}}

data: {"id":0,"jsonrpc":"2.0","result":{"message":{"content":"📢 AWS News Update (2026-04-10T21:41:16+0000)\n\n**Amazon EC2 M8i and M8i-flex instances are now available in AWS GovCloud (US-West) Region**\n\n\u003cp\u003eStarting today, Amazon EC2 M8i and M8i-flex instances are now available in AWS GovCloud (US-West)\u003cb\u003e \u003c/b\u003eRegion. These instances are powered by custom Intel Xeon 6 processors, available only on AWS, delivering the highest performance and fastest memory bandwidth among comparable Intel processors in the cloud. The M8i and M8i-flex instances offer up to 15% better price-performance, and 2.5x more memory bandwidth compared to previous generation Intel-based instances. They deliver up to 20% better performance than M7i and M7i-flex instances, with even higher gains for specific workloads. The M8i and M8i-flex instances are up to 30% faster for PostgreSQL databases, up to 60% faster for NGINX web applications, and up to 40% faster for AI deep learning recommendation models compared to M7i and M7i-flex instances.\u003cbr\u003e\n\u003cbr\u003e\nM8i-flex are the easiest way to get price performance benefits for a majority of general-purpose workloads like web and application servers, microservices, small and medium data stores, virtual desktops, and enterprise applications. They offer the most common sizes, from large to 16xlarge, and are a great first choice for applications that don't fully utilize all compute resources.\u003cbr\u003e\n\u003cbr\u003e\nM8i instances are a great choice for all general purpose workloads, especially for workloads that need the largest instance sizes or continuous high CPU usage. The SAP-certified M8i instances offer 13 sizes including 2 bare metal sizes and the new 96xlarge size for the largest applications.\u003cbr\u003e\n\u003cbr\u003e\nTo get started, sign in to the \u003ca href=\"https://aws.amazon.com/console/\" target=\"_blank\" rel=\"noopener noreferrer\"\u003eAWS Management Console\u003c/a\u003e. For more information about the new instances, visit the\u003ca href=\"https://aws.amazon.com/ec2/instance-types/m8i/\" target=\"_blank\" rel=\"noopener noreferrer\"\u003e M8i and M8i-flex instance page\u003c/a\u003e or visit the AWS News \u003ca href=\"https://aws.amazon.com/blogs/aws/new-general-purpose-amazon-ec2-m8i-and-m8i-flex-instances-are-now-available/\" target=\"_blank\" rel=\"noopener noreferrer\"\u003eblog\u003c/a\u003e.\u003c/p\u003e\n\n\nCategory: compute\nProducts: aws-govcloud-us, amazon-ec2","role":"assistant"}}}

data: {"id":0,"jsonrpc":"2.0","result":{"message":{"content":"📢 AWS News Update (2026-04-13T23:49:12+0000)\n\n**Amazon CloudWatch Logs Insights now supports saved queries with parameters**\n\n\u003cp\u003eAmazon CloudWatch Logs Insights saved queries now support parameters, allowing you to pass values to reusable query templates with placeholders. This eliminates the need to maintain multiple copies of nearly identical queries that differ only in specific values such as log levels, service names, or time intervals.\u003cbr\u003e\n\u003cbr\u003e\nYou can define up to 20 parameters in a query, with each parameter supporting optional default values. For example, you can create a single template to query logs by severity level (such as ERROR or WARN) and pass different service names each time you run it. To execute a query with parameters, invoke it using the query name prefixed with $ and pass your parameter values, such as $ErrorsByService(logLevel=\u0026quot;ERROR\u0026quot;, serviceName=\u0026quot;OrderEntry\u0026quot;). You can also use multiple saved queries with parameters together for complex log analysis, significantly reducing query maintenance overhead while improving reusability.\u003cbr\u003e\n\u003cbr\u003e\nSaved queries with parameters are available in all commercial AWS regions. You can create and use saved queries with parameters using the Amazon CloudWatch console, AWS Command Line Interface (AWS CLI), AWS Cloud Development Kit (AWS CDK), and AWS SDKs. To learn more, see the Amazon CloudWatch Logs \u003ca href=\"https://docs.aws.amazon.com/AmazonCloudWatch/latest/logs/CWL_Insights-Saving-Queries.html\"\u003edocumentation\u003c/a\u003e.\u003c/p\u003e\n\n\nCategory: management-and-governance\nProducts: amazon-cloudwatch, amazon-cloudwatch-logs","role":"assistant"}}}

data: {"id":0,"jsonrpc":"2.0","result":{"message":{"content":"📢 AWS News Update (2026-04-10T21:53:45+0000)\n\n**Amazon EC2 R8i and R8i-flex instances are now available in AWS GovCloud (US-West) Region**\n\n\u003cp\u003eStarting today, Amazon Elastic Compute Cloud (Amazon EC2) R8i and R8i-flex instances are available in the AWS GovCloud (US-West) Region. These instances are powered by custom Intel Xeon 6 processors, available only on AWS, delivering the highest performance and fastest memory bandwidth among comparable Intel processors in the cloud. The R8i and R8i-flex instances offer up to 15% better price-performance, and 2.5x more memory bandwidth compared to previous generation Intel-based instances. They deliver 20% higher performance than R7i instances, with even higher gains for specific workloads. They are up to 30% faster for PostgreSQL databases, up to 60% faster for NGINX web applications, and up to 40% faster for AI deep learning recommendation models compared to R7i.\u003cbr\u003e\n\u003cbr\u003e\nR8i-flex, our first memory-optimized Flex instances, are the easiest way to get price performance benefits for a majority of memory-intensive workloads. They offer the most common sizes, from large to 16xlarge, and are a great first choice for applications that don't fully utilize all compute resources.\u003cbr\u003e\n\u003cbr\u003e\nR8i instances are a great choice for all memory-intensive workloads, especially for workloads that need the largest instance sizes or continuous high CPU usage. R8i instances offer 13 sizes including 2 bare metal sizes and the new 96xlarge size for the largest applications. \u003ca href=\"https://docs.aws.amazon.com/sap/latest/general/sap-hana-aws-ec2.html\" target=\"_blank\" rel=\"noopener noreferrer\"\u003eR8i instances are SAP-certified \u003c/a\u003eand deliver 142,100 aSAPS, delivering exceptional performance for mission-critical SAP workloads.\u003cbr\u003e\n\u003cbr\u003e\nTo get started, sign in to the \u003ca href=\"https://aws.amazon.com/console/\" target=\"_blank\" rel=\"noopener noreferrer\"\u003eAWS Management Console\u003c/a\u003e. For more information about the \u003ca href=\"https://aws.amazon.com/ec2/instance-types/r8i\" target=\"_blank\" rel=\"noopener noreferrer\"\u003eR8i and R8i-flex \u003c/a\u003einstances visit the AWS News \u003ca\u003e\u003c/a\u003e\u003ca href=\"https://aws.amazon.com/blogs/aws/best-performance-and-fastest-memory-with-the-new-amazon-ec2-r8i-and-r8i-flex-instances/\" target=\"_blank\" rel=\"noopener noreferrer\"\u003eblog\u003c/a\u003e.\u003c/p\u003e\n\n\nCategory: compute\nProducts: aws-govcloud-us, amazon-ec2","role":"assistant"}}}

data: {"id":0,"jsonrpc":"2.0","result":{"message":{"content":"📢 AWS News Update (2026-04-13T18:01:02+0000)\n\n**AWS IoT is now available in Israel (Tel Aviv) and Europe (Milan) AWS Regions**\n\n\u003cp\u003eAWS IoT Core and AWS IoT Device Management services are now available in the Israel (Tel Aviv) and Europe (Milan) AWS Regions. With this expansion, organizations operating in these regions can better serve their local customers and unlock multiple benefits, including faster response times, stronger data residency controls, and reduced data transfer expenses.\u003cbr\u003e\n\u003cbr\u003e\n\u003ca href=\"https://aws.amazon.com/iot-core/\"\u003eAWS IoT Core\u003c/a\u003e is a managed cloud service that lets you securely connect billions of Internet of Things (IoT) devices to the cloud and manage them at scale. It routes trillions of messages to IoT devices and AWS endpoints, through bi-directional industry standard protocols, such as MQTT, HTTPS, LoRaWAN (select regions). \u003ca href=\"https://aws.amazon.com/iot-device-management/\"\u003eAWS IoT Device Management\u003c/a\u003e allows customers to search, organize, monitor and remotely manage connected devices at scale.\u003cbr\u003e\n\u003cbr\u003e\nWith the expansion to these regions, AWS IoT is now available in \u003ca href=\"https://aws.amazon.com/about-aws/global-infrastructure/regional-product-services/\"\u003e27 AWS Regions\u003c/a\u003e worldwide. To get started and to learn more, refer to the technical documentation for \u003ca href=\"https://docs.aws.amazon.com/iot/\"\u003eAWS IoT Core \u003c/a\u003eand \u003ca href=\"https://docs.aws.amazon.com/iot-device-management/\"\u003eAWS IoT Device Management\u003c/a\u003e.\u003c/p\u003e\n\n\nCategory: internet-of-things\nProducts: aws-iot-core, aws-iot-device-management, aws-iot","role":"assistant"}}}

data: {"id":0,"jsonrpc":"2.0","result":{"message":{"content":"📢 AWS News Update (2026-04-08T20:31:14+0000)\n\n**Amazon FSx now supports copying file system backups across AWS opt-in Regions **\n\n\u003cp\u003eAmazon\u0026nbsp;FSx\u0026nbsp;now supports copying file system backups across\u0026nbsp;\u003ca href=\"https://docs.aws.amazon.com/glossary/latest/reference/glos-chap.html#optinregion\"\u003eopt-in Regions\u003c/a\u003e\u0026nbsp;(AWS Regions that are disabled by default) for Amazon\u0026nbsp;FSx\u0026nbsp;for Windows File Server, Amazon\u0026nbsp;FSx\u0026nbsp;for\u0026nbsp;Lustre, and Amazon\u0026nbsp;FSx\u0026nbsp;for\u0026nbsp;OpenZFS. This launch makes it easier for customers to meet business continuity, disaster recovery, and compliance requirements by extending cross-Region, cross-account backup and recovery capabilities beyond AWS Regions that are enabled by default.\u003c/p\u003e\n\u003cp\u003eAmazon\u0026nbsp;FSx\u0026nbsp;is a fully managed service that makes it easy and cost-effective to launch, run, and scale feature-rich, high-performance file systems in the AWS Cloud. Opt-in Regions are AWS Regions that are disabled by default, in contrast to regions that are enabled by default. Previously, customers could copy Amazon\u0026nbsp;FSx\u0026nbsp;file system backups across regions enabled by default, within the same AWS account or across AWS accounts in the same AWS Organization. Starting today, you can copy backups into and out of opt-in Regions within the same AWS account using the Amazon FSx console, API, or CLI, or across AWS accounts in the same AWS Organization using AWS Backup. This allows you to design resilient,\u0026nbsp;multi-account,\u0026nbsp;cross-Region\u0026nbsp;backup\u0026nbsp;and recovery architectures across a broader set of AWS Regions.\u003c/p\u003e\n\u003cp\u003eTo get started,\u0026nbsp;visit\u0026nbsp;the\u0026nbsp;\u003ca href=\"https://console.aws.amazon.com/fsx\"\u003eAmazon\u0026nbsp;FSx console\u003c/a\u003e\u0026nbsp;or the\u0026nbsp;\u003ca href=\"https://console.aws.amazon.com/backup\"\u003eAWS Backup console\u003c/a\u003e. For more details, see the\u0026nbsp;\u003ca href=\"https://aws.amazon.com/fsx/\"\u003eAmazon FSx product page\u003c/a\u003e\u0026nbsp;and the\u0026nbsp;\u003ca href=\"https://aws.amazon.com/backup/\"\u003eAWS Backup product page\u003c/a\u003e.\u003c/p\u003e\n\n\nCategory: storage\nProducts: aws-backup, amazon-fsx-for-windows-file-server, amazon-fsx-for-openzfs, amazon-fsx-for-lustre","role":"assistant"}}}

data: {"id":0,"jsonrpc":"2.0","result":{"message":{"content":"📢 AWS News Update (2026-04-13T18:10:08+0000)\n\n**AWS Elastic Disaster Recovery now supports IPv6 **\n\n\u003cp\u003eAWS Elastic Disaster Recovery (AWS DRS) now supports IPv6 for both data replication and control plane connections. Customers operating in IPv6-only or dual-stack network environments can now configure AWS DRS to replicate using IPv6, eliminating the need for IPv4 addresses in their disaster recovery setup.\u003cbr\u003e\n\u003cbr\u003e\nAWS DRS minimizes downtime and data loss with fast, reliable recovery of on-premises and cloud-based applications using affordable storage, minimal compute, and point-in-time recovery. Previously, AWS DRS required IPv4 connectivity for all replication and service communication. Now, customers can set the internet protocol to IPv6 in their replication configuration to use dual-stack endpoints for agent-to-service communication and data replication. This helps customers meet network modernization requirements and enables disaster recovery in environments where IPv4 addresses are unavailable or restricted. Existing replication configurations are not affected and continue to use IPv4 by default.\u003cbr\u003e\n\u003cbr\u003e\nThis capability is available in all AWS Regions where AWS DRS is available and where Amazon EC2 supports IPv6. See the \u003ca href=\"https://aws.amazon.com/about-aws/global-infrastructure/regional-product-services/\"\u003eAWS Regional Services List\u003c/a\u003e for the latest availability information.\u003cbr\u003e\n\u003cbr\u003e\nTo learn more about AWS DRS, visit our \u003ca href=\"https://aws.amazon.com/disaster-recovery/\"\u003eproduct page\u003c/a\u003e or \u003ca href=\"https://docs.aws.amazon.com/drs/latest/userguide/what-is-drs.html\"\u003edocumentation\u003c/a\u003e. To get started, sign in to the \u003ca href=\"https://console.aws.amazon.com/drs/home\"\u003eAWS Elastic Disaster Recovery Console\u003c/a\u003e.\u003c/p\u003e\n\n\nCategory: storage\nProducts: aws-elastic-disaster-recovery","role":"assistant"}}}

data: {"id":0,"jsonrpc":"2.0","result":{"message":{"content":"📢 AWS News Update (2026-04-08T04:04:45+0000)\n\n**AWS announces general availability of AWS Interconnect - last mile**\n\n\u003cp\u003eAWS launches AWS Interconnect - last mile, a fully managed connectivity offering that allows customers to connect their branch offices, data centers, and remote locations to AWS with just a few clicks, eliminating the friction and complexity of network setup. As a milestone collaboration between AWS and Lumen, AWS Interconnect - last mile combines AWS cloud innovation with Lumen’s extensive network footprint to redefine how businesses connect to the cloud.\u003c/p\u003e\n\u003cp\u003eThrough the AWS Console, customers can now instantly establish private, high-speed connections to AWS by simply choosing their preferred AWS Region, bandwidth speed, Direct Connect Gateway ID and partner subscriber ID. Once initiated, AWS generates an activation key to complete provisioning with Lumen. The launch simplifies the connectivity experience by pre-provisioning capacity and automating complex network configuration including BGP peering, VLAN configuration, and ASN assignment. Customers can dynamically scale bandwidth from 1 Gbps to 100 Gbps through the AWS Console and benefit from zero down-time maintenance. The service is designed for high availability and backed by SLA. MACsec encryption is enabled by default for enhanced security between AWS and partner devices.\u003c/p\u003e\n\u003cp\u003eAWS Interconnect - last mile is available in the US through our launch partner Lumen. Partners can also easily adopt via a published open API package on \u003ca href=\"https://github.com/aws/Interconnect\" target=\"_blank\" rel=\"noopener noreferrer\"\u003eGitHub\u003c/a\u003e. For more information, see the AWS Interconnect - last mile \u003ca href=\"https://docs.aws.amazon.com/interconnect/\" target=\"_blank\" rel=\"noopener noreferrer\"\u003edocumentation\u003c/a\u003e\u0026nbsp;and \u003ca href=\"https://aws.amazon.com/interconnect/lastmile/pricing\" target=\"_blank\" rel=\"noopener noreferrer\"\u003epricing\u003c/a\u003e pages.\u003c/p\u003e\n\n\nCategory: networking-and-content-delivery\nProducts: aws-direct-connect","role":"assistant"}}}

data: {"id":0,"jsonrpc":"2.0","result":{"message":{"content":"📢 AWS News Update (2026-04-10T23:55:53+0000)\n\n**Amazon EC2 X8i instances are now available in Europe (Paris)**\n\n\u003cp\u003eAmazon Web Services (AWS) is announcing the general availability of Amazon EC2 X8i instances, next-generation memory optimized instances powered by custom Intel Xeon 6 processors available only on AWS. X8i instances are SAP-certified and deliver the highest performance and fastest memory bandwidth among comparable Intel processors in the cloud. They deliver up to 43% higher performance, 1.5x more memory capacity (up to 6TB), and 3.3x more memory bandwidth compared to previous generation X2i instances.\u003cbr\u003e\n\u003cbr\u003e\nX8i instances are designed for memory-intensive workloads like SAP HANA, large databases, data analytics, and Electronic Design Automation (EDA). Compared to X2i instances, X8i instances offer up to 50% higher SAPS performance, up to 47% faster PostgreSQL performance, 88% faster Memcached performance, and 46% faster AI inference performance. X8i instances come in 14 sizes, from large to 96xlarge, including two bare metal options.\u003cbr\u003e\n\u003cbr\u003e\nX8i instances are available in the following \u003ca href=\"https://aws.amazon.com/about-aws/global-infrastructure/regional-product-services/\" target=\"_blank\"\u003eAWS Regions\u003c/a\u003e: US East (N. Virginia), US East (Ohio), US West (Oregon), Europe (Frankfurt), Europe (Stockholm) and Europe (Paris).\u003cbr\u003e\n\u003cbr\u003e\nTo get started, visit the \u003ca href=\"https://aws.amazon.com/console/\" target=\"_blank\"\u003eAWS Management Console\u003c/a\u003e. X8i instances can be purchased via Savings Plans, On-Demand instances, and Spot instances. For more information visit \u003ca href=\"https://aws.amazon.com/ec2/instance-types/x8i/\" target=\"_blank\"\u003eX8i instances page\u003c/a\u003e.\u003c/p\u003e\n\n\nCategory: compute\nProducts: amazon-ec2","role":"assistant"}}}

data: {"id":0,"jsonrpc":"2.0","result":{"message":{"content":"📢 AWS News Update (2026-03-30T18:01:38+0000)\n\n**Amazon CloudWatch pipelines now supports drop and conditional processing**\n\n\u003cp\u003eAmazon CloudWatch pipelines now supports conditional processing and a new drop events processor, giving you more control over how your log data is transformed. CloudWatch pipelines is a fully managed service that ingests, transforms, and routes log data to CloudWatch without requiring you to manage infrastructure. Until now, processors applied to all log entries uniformly. With conditional processing, you can define rules that determine when a processor runs and which individual log entries it acts on, so you only transform the data that matters.\u003c/p\u003e\n\u003cp\u003eConditional processing is available across 21 processors including Add Entries, Delete Entries, Copy Values, Grok, Rename Key, and more. For each processor, you can set a \u0026quot;run when\u0026quot; condition to skip the entire processor if the condition is not met, or an entry-level condition to control whether each individual action within the processor is applied. The new Drop Events processor lets you filter out unwanted log entries from third-party pipeline connectors based on conditions you define, helping reduce noise and lower costs.\u003c/p\u003e\n\u003cp\u003eConditional processing and the Drop Events processor are available at no additional cost in all AWS Regions where CloudWatch pipelines is generally available. Standard CloudWatch Logs ingestion and storage rates still apply.\u003c/p\u003e\n\u003cp\u003eTo get started, visit the CloudWatch pipelines page in the Amazon CloudWatch console. To learn more, see the \u003ca href=\"https://docs.aws.amazon.com/AmazonCloudWatch/latest/monitoring/cloudwatch-pipelines.html\"\u003eCloudWatch pipelines documentation\u003c/a\u003e.\u003c/p\u003e\n\n\nCategory: management-and-governance\nProducts: amazon-cloudwatch","role":"assistant"}}}

data: {"id":0,"jsonrpc":"2.0","result":{"message":{"content":"📢 AWS News Update (2026-04-09T21:46:02+0000)\n\n**AWS Deadline Cloud supports monitor creation in multiple regions**\n\n\u003cp\u003eToday, AWS Deadline Cloud announces support for creating monitors in multiple AWS Regions without additional configuration of your IAM Identity Center instance. AWS Deadline Cloud is a fully managed service that helps creative teams manage and scale their rendering workloads in the cloud.\u003cbr\u003e\n\u003cbr\u003e\nYou can now deploy render farms with monitors across multiple Regions without needing to adjust your existing IAM Identity Center configuration. You can operate more efficiently by placing rendering resources in regions closest to your artists and studios worldwide, and can run and compare workloads across regions to help optimize your rendering strategy or diversify your instance types. Deadline Cloud automatically routes authentication requests to your IAM Identity Center instance in its primary Region, so your identity data remains in place without replication and requires no changes to your identity management setup.\u003cbr\u003e\n\u003cbr\u003e\nTo learn more, see Getting Started with Deadline Cloud in the \u003ca title=\"deadlineclouduserguide\" href=\"https://docs.aws.amazon.com/deadline-cloud/latest/userguide/monitors-additional-regions.html\"\u003eAWS Deadline Cloud User Guide\u003c/a\u003e.\u0026nbsp;\u003c/p\u003e\n\n\nCategory: media-services\nProducts: aws-deadline-cloud","role":"assistant"}}}

data: {"id":0,"jsonrpc":"2.0","result":{"message":{"content":"📢 AWS News Update (2026-03-25T19:49:58+0000)\n\n**Amazon CloudWatch pipelines introduces new compliance and governance capabilities**\n\n\u003cp\u003eAmazon CloudWatch pipelines now includes new compliance and governance capabilities to help you maintain data integrity and control access when processing logs. CloudWatch pipelines is a fully managed service that ingests, transforms, and routes log data to CloudWatch without requiring you to manage infrastructure. Because pipeline processors modify log events during transformation, organizations with audit or regulatory requirements need ways to preserve original data and track what has been changed. These new tools address those needs directly.\u003c/p\u003e\n\u003cp\u003eYou can now enable a \u0026quot;keep original\u0026quot; toggle to automatically store a copy of your raw logs before any transformation takes place, ensuring the unmodified data is always available when needed. Pipelines also adds new metadata to processed log entries indicating that the log has been transformed, making it easy to distinguish between original and processed data during audits or investigations. Additionally, new IAM condition keys let administrators restrict who can create pipelines based on log source name and type, giving operators fine-grained control over pipeline creation across their organization.\u003c/p\u003e\n\u003cp\u003eThese compliance and governance features are available at no additional cost. Standard CloudWatch Logs storage rates apply to both the original and transformed copies of your log data when the keep original log option is enabled. You can use these features in all AWS Regions where CloudWatch pipelines is generally available.\u003c/p\u003e\n\u003cp\u003eTo get started, visit the CloudWatch Ingestion page in the Amazon CloudWatch console. To learn more, see the \u003ca href=\"https://docs.aws.amazon.com/AmazonCloudWatch/latest/monitoring/cloudwatch-pipelines.html\"\u003eCloudWatch pipelines documentation\u003c/a\u003e.\u003c/p\u003e\n\n\nCategory: management-and-governance\nProducts: amazon-cloudwatch","role":"assistant"}}}

data: {"id":0,"jsonrpc":"2.0","result":{"message":{"content":"📢 AWS News Update (2026-04-06T14:11:33+0000)\n\n**Second-generation Amazon FSx for NetApp ONTAP is now available in four additional AWS commercial and AWS GovCloud (US) Regions**\n\n\u003cp\u003eAmazon FSx for NetApp ONTAP second-generation file systems are now available in 4 additional AWS Regions: Europe (London), Asia Pacific (Hyderabad), South America (Sao Paulo), and\u0026nbsp;AWS GovCloud (US-West).\u0026nbsp;\u003cbr\u003e\n\u003cbr\u003e\nAmazon FSx makes it easier and more cost effective to launch, run, and scale feature-rich, high-performance file systems in the cloud. Second-generation \u003ca href=\"https://aws.amazon.com/fsx/netapp-ontap/\"\u003eFSx for ONTAP\u003c/a\u003e file systems give you more performance scalability and flexibility over first-generation file systems by allowing you to create or expand file systems with up to 12 highly-available (HA) pairs of file servers, providing your workloads with up to 72 GBps of throughput and 1 PiB of provisioned SSD storage.\u003cbr\u003e\n\u003cbr\u003e\nWith this regional expansion, second-generation FSx for ONTAP file systems are available in the following AWS Regions: US East (N. Virginia, Ohio), US West (N. California, Oregon), Canada (Central), Europe (Frankfurt, Ireland, London, Spain, Stockholm, Zurich), South America (Sao Paulo), Asia Pacific (Hyderabad, Mumbai, Seoul, Singapore, Sydney, Tokyo), and AWS GovCloud (US-West). You can create second-generation Multi-AZ file systems with a single HA pair, and Single-AZ file systems with up to 12 HA pairs. To learn more, visit the \u003ca href=\"https://docs.aws.amazon.com/fsx/latest/ONTAPGuide/what-is-fsx-ontap.html\"\u003eFSx for ONTAP user guide\u003c/a\u003e.\u003c/p\u003e\n\u003ch2\u003e\u0026nbsp;\u003c/h2\u003e\n\n\nCategory: storage\nProducts: amazon-fsx-netapp-ontap, aws-govcloud-us","role":"assistant"}}}

data: {"id":0,"jsonrpc":"2.0","result":{"message":{"content":"📢 AWS News Update (2026-04-14T14:06:30+0000)\n\n**Amazon Quick now supports document-level access controls for ACL support for S3**\n\n\u003cp\u003eAmazon Quick now supports document-level access control lists (ACLs) for Amazon S3 knowledge bases, enabling you to manage granular permissions for documents stored in S3. With this feature, you can control which users and groups can access specific documents or folders within your knowledge base, ensuring that sensitive information is only available to authorized personnel.\u003cbr\u003e\n\u003cbr\u003e\nYou can configure document-level ACLs using two methods optimized for different use cases. The global ACL configuration file provides centralized permission management at the folder level, ideal for organizations with stable permission structures. Alternatively, document-level metadata files enable faster permission updates by allowing you to define access controls for individual documents, requiring reindexing only for affected documents rather than entire folder structures.\u003cbr\u003e\nDocument-level ACL configuration is permanent and must be set when creating a new knowledge base. For ACL-enabled knowledge bases, documents without an associated ACL entry are not ingested, ensuring comprehensive access control across your document repository.\u003c/p\u003e\n\u003cp\u003eThis Feature is available in all \u003ca href=\"https://docs.aws.amazon.com/quick/latest/userguide/regions.html#regions-qs\" target=\"_blank\" rel=\"noopener noreferrer\"\u003eAWS Regions where Amazon Quick is available\u003c/a\u003e.\u0026nbsp;\u003cbr\u003e\nTo get started with document-level ACLs for Amazon S3 knowledge bases, visit the \u003ca href=\"https://docs.aws.amazon.com/quicksuite/\" target=\"_blank\"\u003eAmazon Quick User Guide\u003c/a\u003e.\u003c/p\u003e\n\n\nCategory: artificial-intelligence\nProducts: amazon-quick-suite","role":"assistant"}}}

data: {"id":0,"jsonrpc":"2.0","result":{"message":{"content":"📢 AWS News Update (2026-04-10T15:31:07+0000)\n\n**AWS Billing and Cost Management Dashboards Now Supports Scheduled Email Delivery**\n\n\u003cp\u003eAWS Billing and Cost Management Dashboards now support scheduled email delivery for your reports. You can now automate report distribution on flexible recurring schedules, eliminating manual compilation\u0026nbsp;work\u0026nbsp;and ensuring financial insights reach decision-makers without requiring console access.\u0026quot;\u003c/p\u003e\n\u003cp\u003eScheduled email reports enable you to configure daily, weekly, or monthly delivery schedules for your dashboards.\u0026nbsp;Recipients receive emails containing secure links to password-protected PDF reports optimized for offline viewing.\u0026nbsp;Manage recipients through AWS User Notifications, and once configured, reports generate and distribute automatically on your chosen schedule. You can also access these capabilities programmatically through AWS SDKs and CLI tools.\u003c/p\u003e\n\u003cp\u003eThis feature is available at no\u0026nbsp;additional\u0026nbsp;cost in all commercial AWS Regions, excluding AWS China Regions. To get started, open the AWS Billing and Cost Management console, navigate to Dashboards, select a dashboard, and choose 'Manage email reports' from the Actions menu. For more information, see the\u0026nbsp;\u003ca href=\"https://docs.aws.amazon.com/cost-management/latest/userguide/dashboards.html\"\u003eDashboards user guide\u003c/a\u003e\u0026nbsp;and\u0026nbsp;\u003ca href=\"https://aws.amazon.com/blogs/aws-cloud-financial-management/automate-aws-cost-reporting-with-scheduled-dashboard-email-delivery/\"\u003eannouncement blog post\u003c/a\u003e.\u003c/p\u003e\n\n\nCategory: cost-management\nProducts: cloud-financial-management","role":"assistant"}}}

data: {"id":0,"jsonrpc":"2.0","result":{"message":{"content":"📢 AWS News Update (2026-04-06T20:08:28+0000)\n\n**AWS RTB Fabric supports health checks for real-time bidding workloads **\n\n\u003cp\u003eAWS RTB Fabric now supports health checks for real-time bidding workloads that use \u003ca href=\"https://docs.aws.amazon.com/autoscaling/ec2/userguide/auto-scaling-groups.html\"\u003eEC2 Auto Scaling groups\u003c/a\u003e (AGS). Health checks in AWS RTB Fabric continuously monitors and automatically routes traffic to healthy instances with \u003ca href=\"https://docs.aws.amazon.com/rtb-fabric/latest/userguide/health-checks-for-managed-endpoints.html\" target=\"_blank\" rel=\"noopener noreferrer\"\u003econfigurable settings in RTB responder gateways\u003c/a\u003e.\u0026nbsp;This helps eliminate failed real-time bidding transactions from bootstrapping, draining, or failed instances. With this launch, AWS RTB Fabric helps advertising technology (AdTech) companies improve uptime, reduce error rates, and prevent revenue loss from failed auctions.\u003c/p\u003e\n\u003cp\u003eAWS RTB Fabric helps you connect with your AdTech partners such as Amazon Ads, GumGum, Kargo, MobileFuse, Sovrn, TripleLift, Viant, Yieldmo, and more in three steps while delivering single-digit millisecond latency through a private, high-performance network environment. RTB Fabric reduces standard cloud networking costs by up to 80% and does not require upfront commitments. AWS RTB Fabric is generally available in the following \u003ca href=\"https://aws.amazon.com/about-aws/global-infrastructure/regional-product-services/\"\u003eAWS Regions\u003c/a\u003e: US East (N. Virginia), US West (Oregon), Asia Pacific (Singapore), Asia Pacific (Tokyo), Europe (Frankfurt), and Europe (Ireland). To learn more, visit the AWS RTB Fabric \u003ca href=\"https://docs.aws.amazon.com/rtb-fabric/latest/userguide/health-checks-for-managed-endpoints.html\" target=\"_blank\" rel=\"noopener noreferrer\"\u003edocumentation\u003c/a\u003e or \u003ca href=\"https://aws.amazon.com/rtb-fabric/\" target=\"_blank\" rel=\"noopener noreferrer\"\u003eproduct\u003c/a\u003e page.\u0026nbsp;\u003c/p\u003e\n\n\nCategory: networking\nProducts: aws-rtb-fabric","role":"assistant"}}}

data: {"id":0,"jsonrpc":"2.0","result":{"message":{"content":"📢 AWS News Update (2026-04-10T15:43:53+0000)\n\n**AWS Backup extends Amazon FSx support to 5 additional AWS Regions and expands cross-Region and cross-account copy to 14 AWS Regions**\n\n\u003cp\u003eAWS Backup is expanding support for Amazon FSx for Windows File Server, Amazon FSx for OpenZFS, and Amazon FSx for Lustre with two regional enhancements.\u003c/p\u003e\n\u003cp\u003eFirst, AWS Backup now supports backup and restore of these FSx file systems in 5 additional AWS Regions: Asia Pacific (Malaysia, Taipei, Thailand), Canada West (Calgary), and Mexico (Central). You can now centrally manage FSx backup policies, automate backup schedules, and monitor backup activity through AWS Backup in these Regions.\u003c/p\u003e\n\u003cp\u003eSecond, AWS Backup now supports cross-Region and cross-account copy of FSx backups in 14 Regions: Africa (Cape Town), Asia Pacific (Hong Kong, Hyderabad, Jakarta, Malaysia, Melbourne, Taipei, Thailand), Canada West (Calgary), Europe (Milan, Spain, Zurich), Israel (Tel Aviv), and Mexico (Central). This capability is available for both on-demand copies and scheduled backup plans with copy rules. With support in opt-in Regions, you can also store FSx backups in AWS Backup logically air-gapped vaults, providing additional defense against inadvertent or malicious deletions and helping you recover from ransomware events.\u003c/p\u003e\n\u003cp\u003eYou can now configure cross-Region and cross-account copy rules to meet your compliance and business continuity requirements. To learn more, visit the AWS Backup \u003ca href=\"https://docs.aws.amazon.com/aws-backup/latest/devguide/backup-feature-availability.html\"\u003efeature availability page\u003c/a\u003e.\u003c/p\u003e\n\n\nCategory: TBD\nProducts: aws-backup, amazon-fsx-for-lustre, amazon-fsx-for-openzfs, amazon-fsx-for-windows-file-server","role":"assistant"}}}

data: {"id":0,"jsonrpc":"2.0","result":{"message":{"content":"📢 AWS News Update (2026-04-10T20:43:00+0000)\n\n**Amazon RDS now supports the latest CU and GDR updates for Microsoft SQL Server**\n\n\u003cp\u003e\u003ca href=\"https://aws.amazon.com/rds/sqlserver/\" target=\"_blank\"\u003eAmazon Relational Database Service (Amazon RDS) for SQL Server \u003c/a\u003enow supports the latest Cumulative Updates (CU) and General Distribution Release (GDR) updates for Microsoft SQL Server. This release includes support for Microsoft SQL Server 2016 SP3+GDR \u003ca href=\"https://support.microsoft.com/en-us/topic/kb5077474-description-of-the-security-update-for-sql-server-2016-sp3-gdr-march-10-2026-3f455bec-1221-4962-b068-0b11bf96b66a\" target=\"_blank\"\u003eKB5077474\u003c/a\u003e (RDS version 13.00.6480.4.v1), SQL Server 2017 CU31+GDR \u003ca href=\"https://support.microsoft.com/en-us/topic/kb5077471-description-of-the-security-update-for-sql-server-2017-cu31-march-10-2026-f020d5eb-e356-42e8-a9ba-0ef061430b15\" target=\"_blank\"\u003eKB5077471\u003c/a\u003e (RDS version 14.00.3520.4.v1), SQL Server 2019 CU32+GDR \u003ca href=\"https://support.microsoft.com/en-us/topic/kb5077469-description-of-the-security-update-for-sql-server-2019-cu32-march-10-2026-5ec2c609-35cb-483d-aa80-5e66821e5c97\" target=\"_blank\"\u003eKB5077469\u003c/a\u003e (RDS version 15.00.4460.4.v1) and SQL Server 2022 CU24 \u003ca href=\"https://learn.microsoft.com/en-us/troubleshoot/sql/releases/sqlserver-2022/cumulativeupdate24\" target=\"_blank\"\u003eKB5080999\u003c/a\u003e (RDS version 16.00.4245.2.v1).\u003cbr\u003e\n\u003cbr\u003e\nThe GDR updates address vulnerabilities described in CVE-2026-21262 and CVE-2026-26115. For additional information on the improvements and fixes included in these updates, see Microsoft documentation for KB5077474, KB5077471, KB5077469, KB5080999. We recommend that you upgrade your Amazon RDS for SQL Server instances to apply these updates using Amazon RDS Management Console, or by using the AWS SDK or CLI. You can learn more about upgrading your database instance in the \u003ca href=\"https://docs.aws.amazon.com/AmazonRDS/latest/UserGuide/USER_UpgradeDBInstance.SQLServer.html\" target=\"_blank\"\u003eAmazon RDS SQL Server User Guide for upgrading your RDS Microsoft SQL Server DB engine\u003c/a\u003e.\u003c/p\u003e\n\n\nCategory: databases\nProducts: amazon-rds-for-sql-server, aws-govcloud-us","role":"assistant"}}}

data: {"id":0,"jsonrpc":"2.0","result":{"message":{"content":"📢 AWS News Update (2026-04-09T21:57:45+0000)\n\n**Amazon Timestream for InfluxDB Now Supports Customer-Defined Maintenance Windows**\n\n\u003cp\u003eAmazon Timestream for InfluxDB now supports customer-defined maintenance windows, giving you control over when routine maintenance is performed on your InfluxDB databases. This feature is available for both InfluxDB 2 instances and InfluxDB 3 clusters across all supported editions.\u003cbr\u003e\n\u003cbr\u003e\nWith this launch, you can specify a weekly maintenance window using a day-and-time format in your preferred timezone. Timestream for InfluxDB supports IANA timezone identifiers such as America/New_York, Europe/London, and Asia/Tokyo, and automatically handles Daylight Saving Time transitions so you don't need to manually adjust your schedule. If you don't specify a maintenance window, the service continues to manage maintenance timing automatically.\u003cbr\u003e\n\u003cbr\u003e\nYou can set or update your preferred maintenance window when creating or modifying a resource using the Amazon Timestream for InfluxDB console, AWS CLI, or AWS SDKs.\u0026nbsp;You can use Amazon Timestream for InfluxDB Customer-Defined Maintenance Windows in all \u003ca href=\"https://docs.aws.amazon.com/general/latest/gr/timestream.html\"\u003eRegions\u003c/a\u003e where Timestream for InfluxDB is offered.\u003cbr\u003e\n\u003cbr\u003e\nTo get started with Amazon Timestream for InfluxDB, visit the \u003ca href=\"https://console.aws.amazon.com/timestream\"\u003eAmazon Timestream for InfluxDB console\u003c/a\u003e. For more information, see the \u003ca href=\"https://docs.aws.amazon.com/timestream/\"\u003eAmazon Timestream for InfluxDB documentation\u003c/a\u003e and \u003ca href=\"https://aws.amazon.com/timestream/pricing/\"\u003epricing page\u003c/a\u003e.\u003c/p\u003e\n\n\nCategory: internet-of-things\nProducts: amazon-timestream","role":"assistant"}}}

data: {"id":0,"jsonrpc":"2.0","result":{"message":{"content":"📢 AWS News Update (2026-04-09T21:57:54+0000)\n\n**Amazon Bedrock now supports cost allocation by IAM user and role**\n\n\u003cp\u003eAmazon Bedrock now supports cost allocation by IAM principal, such as IAM users and IAM roles, in AWS Cost and Usage Report 2.0 (CUR 2.0) and Cost Explorer. This enables customers to understand and attribute Bedrock model inference costs across users, teams, projects, and applications.\u003cbr\u003e\n\u003cbr\u003e\nWith this launch, customers can tag their IAM users and roles with attributes like team, project, or cost center, activate them as cost allocation tags, and analyze Bedrock model inference costs by the tags in Cost Explorer or at the line-item level in CUR 2.0. To get started, tag your IAM users and roles and activate them as cost allocation tags in the Billing and Cost Management console. Then create a CUR 2.0 data export and select \u0026quot;Include caller identity (IAM principal) allocation data\u0026quot; or filter by tags in Cost Explorer.\u003cbr\u003e\n\u003cbr\u003e\nThis feature is available in all AWS commercial Regions where Amazon Bedrock is available. To learn more, see \u003ca href=\"https://docs.aws.amazon.com/awsaccountbilling/latest/aboutv2/iam-principal-cost-allocation.html\"\u003eUsing IAM principal for Cost Allocation documentation\u003c/a\u003e. To get started with Amazon Bedrock, visit \u003ca href=\"https://docs.aws.amazon.com/bedrock/latest/userguide/getting-started.html\"\u003eAmazon Bedrock documentation.\u003c/a\u003e\u003c/p\u003e\n\n\nCategory: cost-management\nProducts: amazon-bedrock","role":"assistant"}}}

data: {"id":0,"jsonrpc":"2.0","result":{"message":{"content":"📢 AWS News Update (2026-04-08T17:25:23+0000)\n\n**Amazon OpenSearch Service supports Managed Prometheus and agent tracing**\n\n\u003cp\u003eAmazon OpenSearch Service now provides a unified observability experience that brings together metrics, logs, traces, and AI agent tracing in a single interface. This release introduces native integration with Amazon Managed Service for Prometheus and comprehensive agent tracing capabilities, addressing the dual challenges of prohibitive costs from premium observability platforms and operational complexity from fragmented tooling. Site Reliability Engineers, DevOps Engineers, and Platform Engineering teams can now consolidate their observability stack without costly data duplication or constant context switching between multiple tools.\u003c/p\u003e\n\u003cp\u003eYou can now query Prometheus metrics directly using native PromQL syntax alongside logs and traces in OpenSearch UI's observability workspace—without duplicating data. Combined with new application monitoring workflows powered by RED metrics (Rate, Errors, Duration) and AI agent tracing using OpenTelemetry GenAI semantic conventions, operations teams can correlate slow traces to application logs, overlay Prometheus metrics on service dashboards, and trace LLM agent execution—all without switching tools. This live query architecture delivers significant cost reduction compared to premium platforms while maintaining operational excellence.\u003c/p\u003e\n\u003cp\u003eThe new unified observability experience is available on OpenSearch UI in 20 AWS regions: US East (N. Virginia, Ohio), US West (N. California, Oregon), Asia Pacific (Hong Kong, Mumbai, Osaka, Seoul, Singapore, Sydney, Tokyo), Europe (Frankfurt, Ireland, London, Milan, Paris, Spain, Stockholm), Canada (Central), and South America (São Paulo).\u003c/p\u003e\n\u003cp\u003eTo learn more, visit the\u0026nbsp;\u003ca href=\"https://docs.aws.amazon.com/opensearch-service/latest/developerguide/observability.html\"\u003eOpenSearch Service observability documentation\u003c/a\u003e\u0026nbsp;and\u0026nbsp;\u003ca href=\"https://docs.aws.amazon.com/opensearch-service/latest/developerguide/direct-query-prometheus-overview.html\"\u003edirect query documentation\u003c/a\u003e.\u003c/p\u003e\n\n\nCategory: analytics\nProducts: amazon-opensearch-service, Amazon Managed Service for Prometheus","role":"assistant"}}}

data: {"id":0,"jsonrpc":"2.0","result":{"message":{"content":"📢 AWS News Update (2026-03-30T17:41:51+0000)\n\n**Amazon S3 Lifecycle pauses actions on objects that are unable to replicate**\n\n\u003cp\u003eAmazon S3 Lifecycle now prevents expiration and transition actions on objects that failed replication, helping you to coordinate replication configuration or permissions changes with actions defined in your lifecycle rules.\u003c/p\u003e\n\u003cp\u003eIncorrect permissions or replication configuration can prevent objects from being replicated. With this change, S3 Lifecycle no longer expires or transitions objects that have failed replication, even if they match one of the lifecycle rules that you have defined. Once you have corrected your replication configuration or permissions, you can use S3 Batch Replication to replicate objects that previously failed. After successful replication, S3 Lifecycle will automatically process these objects according to your configured rules.\u003c/p\u003e\n\u003cp\u003eThis change applies automatically to all existing and new S3 Lifecycle configurations, across 37 AWS Regions, including the AWS China and AWS GovCloud (US) Regions. We are in the process of deploying this change and plan to complete the deployment in the coming days. To learn more, visit \u003ca href=\"https://docs.aws.amazon.com/AmazonS3/latest/userguide/object-lifecycle-mgmt.html\" target=\"_blank\" rel=\"noopener noreferrer\"\u003eS3 Lifecycle documentation\u003c/a\u003e and \u003ca href=\"https://docs.aws.amazon.com/AmazonS3/latest/userguide/replication-troubleshoot.html\" target=\"_blank\" rel=\"noopener noreferrer\"\u003eS3 Replication troubleshooting documentation\u003c/a\u003e.\u003c/p\u003e\n\n\nCategory: storage\nProducts: amazon-s3","role":"assistant"}}}

data: {"id":0,"jsonrpc":"2.0","result":{"message":{"content":"📢 AWS News Update (2026-04-09T20:27:02+0000)\n\n**Amazon RDS Blue/Green Deployments now supports Amazon RDS Proxy**\n\n\u003cp\u003eAmazon RDS Blue/Green Deployments now supports Amazon RDS Proxy, enabling faster application recovery during switchover by eliminating DNS propagation delays. Blue/Green Deployments create a fully managed staging environment (Green) that allows you to deploy and test production changes, keeping your current production database (Blue) safe. When ready, you can switchover to the new production environment and your applications begin accessing it immediately without any configuration changes.\u003cbr\u003e\n\u003cbr\u003e\nDuring a Blue/Green Deployment switchover for single-Region configurations, RDS Proxy actively monitors database instances and detects when the Green environment becomes the new production environment. This allows RDS Proxy to quickly redirect connections to the Green environment, enabling faster application recovery. You don't need to modify your drivers or change your existing application setup.\u003cbr\u003e\n\u003cbr\u003e\nAmazon RDS Blue/Green Deployments with Amazon RDS Proxy is available for Amazon Aurora with MySQL compatibility, Amazon Aurora with PostgreSQL compatibility, Amazon RDS for MySQL, Amazon RDS for PostgreSQL, and Amazon RDS for MariaDB in all commercial AWS Regions where \u003ca href=\"https://docs.aws.amazon.com/AmazonRDS/latest/UserGuide/Concepts.RDS_Fea_Regions_DB-eng.Feature.RDSProxy.html\" target=\"_blank\" rel=\"noopener noreferrer\"\u003eRDS Proxy is available\u003c/a\u003e.\u003cbr\u003e\n\u003cbr\u003e\nIn a few clicks, update your databases using RDS Blue/Green Deployments via the Amazon RDS Console or Amazon CLI. To learn more, see \u003ca href=\"https://docs.aws.amazon.com/AmazonRDS/latest/AuroraUserGuide/blue-green-deployments-overview.html\" target=\"_blank\" rel=\"noopener noreferrer\"\u003eBlue/Green Deployments overview\u003c/a\u003e in the Amazon RDS documentation.\u003c/p\u003e\n\n\nCategory: databases\nProducts: amazon-aurora","role":"assistant"}}}

data: {"id":0,"jsonrpc":"2.0","result":{"message":{"content":"📢 AWS News Update (2026-04-06T18:06:27+0000)\n\n**AWS Marketplace announces the Discovery API for programmatic access to catalog data**\n\n\u003cp\u003eToday, AWS Marketplace announces the Discovery API, giving you programmatic access to product and pricing information across the AWS Marketplace catalog — including SaaS, AI agents and tools, AMI, containers, and machine learning models.\u003cbr\u003e\n\u003cbr\u003e\nWith the Discovery API, buyers can embed catalog data into internal portals, enrich procurement tools with current pricing and offer terms, and streamline vendor evaluation workflows. Sellers and channel partners can surface product listings, public pricing, and private offer details directly within their own websites and storefronts — helping customers browse, compare, and move to purchase without leaving the partner experience.\u003cbr\u003e\n\u003cbr\u003e\nThe API provides access to product descriptions, categories, pricing across public and private offers, and offer terms, so you can build experiences tailored to how your organization discovers and procures software through AWS Marketplace.\u003cbr\u003e\n\u003cbr\u003e\nThe AWS Marketplace Discovery API is available in US East (N. Virginia), US West (Oregon), and Europe (Ireland).\u003cbr\u003e\n\u003cbr\u003e\nYou can get started by configuring IAM permissions for your AWS account and calling the API through the AWS SDK. For more information, see the \u003ca href=\"https://docs.aws.amazon.com/marketplace/latest/APIReference/discovery-apis.html\"\u003eAWS Marketplace Discovery API Reference\u003c/a\u003e.\u003c/p\u003e\n\n\nCategory: aws-marketplace-and-partners\nProducts: aws-marketplace","role":"assistant"}}}

data: {"id":0,"jsonrpc":"2.0","result":{"message":{"content":"📢 AWS News Update (2026-04-08T23:53:49+0000)\n\n**AWS Agent Registry for centralized agent discovery and governance is now available in Preview**\n\n\u003cp\u003eAWS Agent Registry, available through Amazon Bedrock AgentCore, is now in preview — a private, governed catalog and discovery layer for agents, tools, skills, MCP servers, and custom resources within the organization. It gives teams complete visibility into their AI landscape, enabling them to discover existing agents and tools instead of rebuilding capabilities that already exist. The registry can be accessed via the AgentCore Console UI, APIs (AWS CLI, AWS SDK), or as an MCP server that builders can query and invoke directly from their IDEs. Registry supports both IAM and OAuth (Custom JWT) based access.\u003c/p\u003e\n\u003cp\u003eTeams can register resources manually through the console or API, or use URL-based discovery, which automatically retrieves metadata such as tool schemas and capability descriptions from a live MCP server or agent endpoint. Records go through an approval workflow where administrators can approve records before they become discoverable, and they can plug the registry into their existing approval workflows to enforce governance policies. AWS CloudTrail provides complete audit trails of all registry access and administrative actions, ensuring compliance and security oversight. For discovery, the registry offers both semantic and keyword search, so developers can quickly find agents by describing their use case in natural language.\u0026nbsp;\u003c/p\u003e\n\u003cp\u003eAWS Agent Registry (preview) is available in five AWS Regions where AgentCore is available: US West (Oregon), Asia Pacific (Tokyo), Asia Pacific (Sydney), Europe (Ireland), and US East (N. Virginia). Learn more about the registry through the \u003ca href=\"https://aws.amazon.com/blogs/machine-learning/the-future-of-managing-agents-at-scale-aws-agent-registry-now-in-preview/\"\u003eblog\u003c/a\u003e, and deep dive using the \u003ca href=\"https://docs.aws.amazon.com/bedrock-agentcore/latest/devguide/registry.html\"\u003edocumentation\u003c/a\u003e.\u003c/p\u003e\n\n\nCategory: artificial-intelligence\nProducts: amazon-bedrock","role":"assistant"}}}

data: {"id":0,"jsonrpc":"2.0","result":{"message":{"content":"📢 AWS News Update (2026-03-13T21:02:10+0000)\n\n**Amazon OpenSearch Serverless now supports Zstandard (zstd) codec for index compression**\n\n\u003cp\u003eAmazon OpenSearch Serverless now supports Zstandard codecs for index storage, giving customers greater control over the trade-off between storage costs and query performance. With this launch, customers can configure\u0026nbsp;Zstandard compression to achieve up to 32% reduction in index size compared to the default LZ4 codec, helping lower managed storage costs for data-intensive workloads.\u003c/p\u003e\n\u003cp\u003eCustomers running large-scale log analytics, observability pipelines, and time-series workloads on Amazon OpenSearch Serverless can benefit most from\u0026nbsp;Zstandard compression where high data volumes make storage efficiency a significant cost driver. The Zstandard compression algorithm is available in two different modes in Amazon OpenSearch Serverless:\u0026nbsp;zstd\u0026nbsp;and\u0026nbsp;zstd_no_dict. Customers can tune the compression level to balance their specific needs: lower levels (e.g., level 1) deliver meaningful storage savings with minimal impact on indexing throughput and query latency, while higher levels (e.g., level 6) maximize compression ratios at the cost of slower indexing speeds.\u0026nbsp;\u003c/p\u003e\n\u003cp\u003eZstandard codec support is available today in all AWS Regions where Amazon OpenSearch Serverless is \u003ca href=\"https://docs.aws.amazon.com/general/latest/gr/opensearch-service.html\" target=\"_blank\" rel=\"noopener noreferrer\"\u003esupported\u003c/a\u003e. To get started, you can specify these codecs in your index settings at creation time. For more information, see the Amazon OpenSearch Serverless \u003ca href=\"https://docs.aws.amazon.com/opensearch-service/latest/developerguide/serverless-zstd-compression.html\" target=\"_blank\" rel=\"noopener noreferrer\"\u003edocumentation\u003c/a\u003e.\u003c/p\u003e\n\n\nCategory: analytics\nProducts: amazon-opensearch-service","role":"assistant"}}}

data: {"id":0,"jsonrpc":"2.0","result":{"message":{"content":"📢 AWS News Update (2026-04-08T23:43:20+0000)\n\n**AWS Private CA now supports customer managed permissions for cross-account sharing**\n\n\u003cp\u003eAWS Private Certificate Authority (AWS Private CA) now supports customer managed permissions in AWS Resource Access Manager (AWS RAM). AWS Private CA lets you share certificate authorities (CAs) across accounts using AWS RAM so you can centralize your PKI instead of creating separate CAs in every account. With customer managed permissions, you can now select exactly which AWS Private CA API operations to allow when sharing a CA, granting only the specific operations each consuming account needs.\u003cbr\u003e\n\u003cbr\u003e\nPreviously, you could only use AWS managed permissions, which provide predefined sets of actions and restrict cross-account issuers to specific certificate templates. Now you can select from read operations (e.g., DescribeCertificateAuthority, GetCertificate, and GetCertificateAuthorityCertificate) and write operations (e.g., IssueCertificate and RevokeCertificate) to tailor access for each consuming account or organizational unit. With customer managed permissions, cross-account issuers are not restricted to a specific certificate template.\u003cbr\u003e\n\u003cbr\u003e\nCustomer managed permissions for AWS Private CA are available in all AWS Regions where AWS Private CA and AWS RAM are available. To learn more, see \u003ca href=\"https://docs.aws.amazon.com/privateca/latest/userguide/pca-cmp.html\"\u003eCustomer managed permissions in RAM\u003c/a\u003e in the AWS Private CA User Guide and \u003ca href=\"https://docs.aws.amazon.com/ram/latest/userguide/create-customer-managed-permissions.html\"\u003eCreating and using customer managed permissions\u003c/a\u003e in the AWS RAM User Guide.\u003c/p\u003e\n\n\nCategory: security-identity-and-compliance\nProducts: aws-private-certificate-authority","role":"assistant"}}}

data: {"id":0,"jsonrpc":"2.0","result":{"message":{"content":"📢 AWS News Update (2026-04-07T23:26:57+0000)\n\n**Amazon EC2 Capacity Manager now supports tag-based dimensions**\n\n\u003cp\u003eStarting today, Amazon EC2 Capacity Manager supports tag-based dimensions, enabling you to use tags from your EC2 resources to group and filter capacity metrics. EC2 Capacity Manager helps you monitor and optimize capacity usage across On-Demand Instances, Spot Instances, and Capacity Reservations. This launch also introduces Account Name as a new built-in dimension.\u003c/p\u003e\n\u003cp\u003eYou can activate up to five custom tag keys — such as environment, team, or cost-center — and use them alongside built-in dimensions like Region, Instance Type, and Availability Zone to group and filter capacity metrics by tag values in the console and APIs, and include tag data as additional columns in newly created S3 data exports. Capacity Manager also includes four Capacity Manager-provided tags by default: EC2 Auto Scaling group name, EKS cluster name, EKS Kubernetes node pool, and Karpenter node pool. The new Account Name dimension makes it easier to identify accounts when analyzing cross-account capacity data across your organization.\u003c/p\u003e\n\u003cp\u003eThis feature is available in all AWS Regions where EC2 Capacity Manager is available. To get started, navigate to the Settings tab in Capacity Manager and choose Manage tag keys, or use the AWS CLI. To learn more, see \u003ca href=\"https://docs.aws.amazon.com/AWSEC2/latest/UserGuide/managing-monitored-tag-keys.html\"\u003eManaging monitored tag keys\u003c/a\u003e in the \u003ci\u003eAmazon EC2 User Guide\u003c/i\u003e. For more information about Amazon EC2 Capacity Manager, visit the \u003ca href=\"https://docs.aws.amazon.com/AWSEC2/latest/UserGuide/capacity-manager.html\"\u003eEC2 Capacity Manager documentation\u003c/a\u003e.\u003c/p\u003e\n\n\nCategory: TBD\nProducts: ","role":"assistant"}}}

data: {"id":0,"jsonrpc":"2.0","result":{"message":{"content":"📢 AWS News Update (2026-04-07T22:40:54+0000)\n\n**Amazon Route 53 Resolver endpoints now support DNS delegation for private hosted zones in AWS GovCloud (US) Regions**\n\n\u003cp\u003eStarting today, domain name system (DNS) delegation for private hosted zone subdomains can be used with Route 53 inbound and outbound Resolver endpoints in AWS GovCloud (US) Regions. This allows you to delegate the authority for a subdomain from your on-premises infrastructure to the Route 53 Resolver cloud service and vice versa, enabling a simplified cloud experience across namespaces in AWS and on your own local infrastructure.\u003cbr\u003e\n\u003cbr\u003e\nMany AWS customers delegate subdomain management to individual teams while maintaining central control of apex domains. Previously, we launched Route 53 Resolver delegation support for private hosted zones in commercial AWS Regions, enabling customers to use standard name server records to delegate subdomain authority between Route 53 and on-premises DNS — eliminating the need for conditional forwarding rules across their organization. With today's release, this delegation capability is now available for Route 53 Resolver endpoints in AWS GovCloud (US-East) and AWS GovCloud (US-West) Regions as well.\u003cbr\u003e\n\u003cbr\u003e\nInbound and outbound delegation is provided at no additional cost to Resolver endpoints usage. For more details on pricing, visit the Route 53 \u003ca href=\"https://aws.amazon.com/route53/pricing/\" target=\"_blank\" rel=\"noopener noreferrer\"\u003epricing page\u003c/a\u003e, and to learn more about this feature, visit the \u003ca href=\"https://docs.aws.amazon.com/Route53/latest/DeveloperGuide/outbound-delegation-tutorial.html\" target=\"_blank\" rel=\"noopener noreferrer\"\u003edeveloper guide\u003c/a\u003e.\u0026nbsp;\u003c/p\u003e\n\n\nCategory: networking-and-content-delivery\nProducts: amazon-route-53, aws-govcloud-us","role":"assistant"}}}

data: {"id":0,"jsonrpc":"2.0","result":{"message":{"content":"📢 AWS News Update (2026-04-08T19:29:28+0000)\n\n**SageMaker HyperPod now supports gang scheduling for distributed training workloads**\n\n\u003cp\u003eAmazon SageMaker HyperPod task governance now supports gang scheduling, which ensures all pods required for a distributed training job are ready before training begins. Administrators can configure gang scheduling to prevent wasted compute from partial job runs and avoid deadlocks from jobs waiting for resources.\u003cbr\u003e\n\u003cbr\u003e\nData scientists running distributed AI/ML training jobs on Amazon SageMaker HyperPod clusters using the EKS orchestrator require multiple pods to work together across nodes with pod-to-pod communication. When some pods start but others do not, jobs can hold onto resources without making progress, block other workloads, and increase costs. Gang scheduling resolves this by monitoring all pods in a workload and pulling the workload back if not all pods are ready within a set time. Pulled-back workloads are automatically requeued to prevent stalling. Administrators can adjust settings on the HyperPod Console, such as how long to wait for pods to be ready, how to handle node failures, whether to admit workloads one at a time to avoid deadlocks on busy clusters, and how retries are scheduled.\u003cbr\u003e\n\u003cbr\u003e\nThis capability is currently available for Amazon SageMaker HyperPod clusters using the EKS orchestrator across the following \u003ca href=\"https://aws.amazon.com/about-aws/global-infrastructure/regional-product-services/\"\u003eAWS Regions\u003c/a\u003e: US East (N. Virginia), US East (Ohio), US West (N. California), US West (Oregon), Asia Pacific (Mumbai), Asia Pacific (Singapore), Asia Pacific (Sydney), and Asia Pacific (Tokyo), Asia Pacific (Jakarta), Europe (Frankfurt), Europe (Ireland), Europe (London), Europe (Stockholm), Europe (Spain), and South America (São Paulo).\u003cbr\u003e\n\u003cbr\u003e\nTo learn more, visit \u003ca href=\"https://aws.amazon.com/sagemaker-ai/hyperpod/\"\u003eSageMaker HyperPod webpage\u003c/a\u003e, and \u003ca href=\"https://docs.aws.amazon.com/sagemaker/latest/dg/sagemaker-hyperpod-eks-operate-console-ui-governance-tasks-gang-scheduling.html\"\u003eHyperPod task governance documentation\u003c/a\u003e.\u003c/p\u003e\n\n\nCategory: artificial-intelligence\nProducts: amazon-sagemaker","role":"assistant"}}}

data: {"id":0,"jsonrpc":"2.0","result":{"message":{"content":"📢 AWS News Update (2026-04-03T20:50:04+0000)\n\n**Amazon IVS Real-Time Streaming now supports redundant ingest **\n\n\u003cp\u003eAmazon Interactive Video Service (Amazon IVS) Real-Time Streaming now supports redundant ingest, helping protect your live streams against source encoder failures and first-mile network issues. With redundant ingest, you can stream from two encoders simultaneously to a single stage with automated failover, ensuring uninterrupted delivery to your viewers.\u003c/p\u003e\n\u003cp\u003eRedundant ingest is ideal for live events, 24/7 live streams, or any scenario where uninterrupted delivery is essential. This capability helps you maintain viewer engagement during unexpected disruptions and enables continuous 24/7 streaming.\u0026nbsp;\u003c/p\u003e\n\u003cp\u003e\u003ca href=\"https://aws.amazon.com/ivs/\"\u003eAmazon IVS\u003c/a\u003e is a managed live streaming solution designed to make low-latency or real-time video available to viewers around the world. Visit the \u003ca href=\"https://aws.amazon.com/about-aws/global-infrastructure/regional-product-services/\"\u003eAWS region table\u003c/a\u003e for a full list of AWS Regions where the Amazon IVS console and APIs for control and creation of video streams are available.\u003c/p\u003e\n\u003cp\u003eTo learn more, please visit the \u003ca href=\"https://docs.aws.amazon.com/ivs/latest/RealTimeUserGuide/rt-rtmp-publishing.html\"\u003eAmazon IVS Real-Time Streaming RTMP ingest documentation\u003c/a\u003e page.\u003c/p\u003e\n\n\nCategory: TBD\nProducts: ","role":"assistant"}}}

data: {"id":0,"jsonrpc":"2.0","result":{"message":{"content":"📢 AWS News Update (2026-03-17T19:57:05+0000)\n\n**Amazon EKS managed node groups now support EC2 Auto Scaling warm pools**\n\n\u003cp\u003eAmazon Elastic Kubernetes Service (Amazon EKS) \u003ca href=\"https://docs.aws.amazon.com/eks/latest/userguide/managed-node-groups.html\"\u003emanaged node groups\u003c/a\u003e now support \u003ca href=\"https://docs.aws.amazon.com/autoscaling/ec2/userguide/ec2-auto-scaling-warm-pools.html\"\u003eAuto Scaling warm pools\u003c/a\u003e, enabling you to maintain pre-initialized EC2 instances ready for rapid scale-out. This reduces node provisioning latency for applications with burst traffic patterns, time-sensitive workloads, or long instance boot times due to complex initialization scripts and software dependencies.\u003cbr\u003e\n\u003cbr\u003e\nWith warm pools enabled, your EKS managed node group maintains a pool of instances that have already completed OS initialization, user data execution, and software configuration. When demand increases and the Auto Scaling group scales out, instances transition from the warm pool to active service without repeating the full cold-start sequence. You can configure instances in the warm pool as Stopped (lower cost, longer transition) or Running (higher cost, faster transition). You can also enable reuse on scale-in, which returns instances to the warm pool during scale-down instead of terminating them. Warm pools work with Cluster Autoscaler without requiring any additional configuration.\u003cbr\u003e\n\u003cbr\u003e\nYou can enable warm pools through the EKS API, AWS CLI, AWS Management Console, or AWS CloudFormation by adding a warmPoolConfig to your CreateNodegroup or UpdateNodegroupConfig requests. Existing managed node groups that do not enable warm pools are unaffected.\u003cbr\u003e\n\u003cbr\u003e\nThis feature is available in all AWS Regions where Amazon EKS is available, except for the China (Beijing) Region, operated by Sinnet and the China (Ningxia) Region, operated by NWCD. To get started, see the \u003ca href=\"https://docs.aws.amazon.com/eks/latest/userguide/managed-node-groups.html\"\u003eAmazon EKS managed node groups documentation\u003c/a\u003e.\u003c/p\u003e\n\n\nCategory: compute\nProducts: amazon-eks","role":"assistant"}}}

data: {"id":0,"jsonrpc":"2.0","result":{"message":{"content":"📢 AWS News Update (2026-04-08T19:46:39+0000)\n\n**Amazon Bedrock AgentCore Browser adds OS-level interaction capabilities**\n\n\u003cp\u003eAmazon Bedrock AgentCore Browser now supports OS-level interaction capabilities, enabling automation of browser workflows that require direct operating system control beyond Chrome DevTools Protocol (CDP) capabilities. This enhancement addresses automation scenarios where CDP alone is insufficient, such as mouse operations, print dialogs, native system alerts, and keyboard shortcuts. The feature serves AI agent developers, test automation engineers, and organizations building LLM-powered web interaction tools.\u003cbr\u003e\n\u003cbr\u003e\nThe new capabilities provide automation through mouse operations (click, move, drag, scroll), keyboard operations (type, press, shortcuts like ctrl+a and ctrl+p), and full desktop screenshots, all at OS-level coordinates extending beyond the browser viewport. Key use cases include automated testing with system dialog handling, document management workflows, complex UI interactions with right-click menus, and vision-based AI agents that require complete browser environment visibility.\u003cbr\u003e\n\u003cbr\u003e\nThis feature is available by default on all browser instances in all 14 \u003ca href=\"https://docs.aws.amazon.com/bedrock-agentcore/latest/devguide/agentcore-regions.html\"\u003eAWS Regions\u003c/a\u003e where Amazon Bedrock AgentCore Browser is available: US East (N. Virginia), US East (Ohio), US West (Oregon), Europe (Frankfurt), Europe (Ireland), Europe (London), Europe (Paris), Europe (Stockholm), Asia Pacific (Mumbai), Asia Pacific (Singapore), Asia Pacific (Sydney), Asia Pacific (Tokyo), Asia Pacific (Seoul), and Canada (Central).\u003cbr\u003e\n\u003cbr\u003e\nTo learn more, visit the \u003ca href=\"https://docs.aws.amazon.com/bedrock-agentcore/latest/devguide/browser-tool.html\"\u003eAgentCore Browser documentation\u003c/a\u003e.\u0026nbsp;\u003c/p\u003e\n\n\nCategory: artificial-intelligence\nProducts: amazon-bedrock","role":"assistant"}}}

data: {"id":0,"jsonrpc":"2.0","result":{"message":{"content":"📢 AWS News Update (2026-04-08T18:06:51+0000)\n\n**Amazon WorkSpaces Advisor now available for AI-powered troubleshooting**\n\n\u003cp\u003eAmazon WorkSpaces Advisor is a new AI-powered tool that helps administrators quickly troubleshoot and resolve issues with Amazon WorkSpaces Personal. Using generative AI capabilities, it analyzes WorkSpace configurations, identifies problems, and provides actionable recommendations to restore service and optimize performance.\u003cbr\u003e\n\u003c/p\u003e\n\u003cp\u003eWorkSpaces Advisor streamlines administrative workflows by reducing the time needed to investigate and fix common issues. Administrators can leverage AI-driven insights to proactively maintain their virtual desktop infrastructure, improve end-user experience, and minimize downtime across their WorkSpaces.\u003c/p\u003e\n\u003cp\u003eAmazon WorkSpaces Advisor is now available in all AWS commercial regions where Amazon WorkSpaces is offered. Visit the Amazon WorkSpaces console to access WorkSpaces Advisor and begin troubleshooting your environment. Learn more in the \u003ca href=\"https://aws.amazon.com/blogs/desktop-and-application-streaming/troubleshoot-amazon-workspaces-personal-issues-faster-with-workspaces-advisor/\"\u003efeature blog\u003c/a\u003e and \u003ca href=\"https://docs.aws.amazon.com/workspaces/latest/adminguide/workspaces-advisor.html\"\u003euser guide\u003c/a\u003e.\u003c/p\u003e\n\n\nCategory: desktop-and-app-streaming\nProducts: amazon-workspaces","role":"assistant"}}}

