• Open

    Clocking on at the Outrage Factory
    TL;DR Our online discourse is the victim of industrial scale pollution, and the incentives are being aligned in the wrong direction. Rather than polluters being penalised there’s now an entire industry that’s paid to pollute. Filter Failure at the Outrage Factory is no longer just the work of ‘amateur’ fringe trolls and state sponsored propaganda; […]  ( 14 min )
    Clocking on at the Outrage Factory
    TL;DR Our online discourse is the victim of industrial scale pollution, and the incentives are being aligned in the wrong direction. Rather than polluters being penalised there’s now an entire industry that’s paid to pollute. Filter Failure at the Outrage Factory is no longer just the work of ‘amateur’ fringe trolls and state sponsored propaganda; […]  ( 14 min )

  • Open

    How Myriad Genetics achieved fast, accurate, and cost-efficient document processing using the AWS open-source Generative AI Intelligent Document Processing Accelerator
    In this post, we explore how Myriad Genetics partnered with the AWS Generative AI Innovation Center to transform their healthcare document processing pipeline using Amazon Bedrock and Amazon Nova foundation models, achieving 98% classification accuracy while reducing costs by 77% and processing time by 80%. We detail the technical implementation using AWS's open-source GenAI Intelligent Document Processing Accelerator, the optimization strategies for document classification and key information extraction, and the measurable business impact on Myriad's prior authorization workflows.  ( 115 min )
    How CBRE powers unified property management search and digital assistant using Amazon Bedrock
    In this post, CBRE and AWS demonstrate how they transformed property management by building a unified search and digital assistant using Amazon Bedrock, enabling professionals to access millions of documents and multiple databases through natural language queries. The solution combines Amazon Nova Pro for SQL generation and Claude Haiku for document interactions, achieving a 67% reduction in processing time while maintaining enterprise-grade security across more than eight million documents.  ( 118 min )
    Managed Tiered KV Cache and Intelligent Routing for Amazon SageMaker HyperPod
    In this post, we introduce Managed Tiered KV Cache and Intelligent Routing for Amazon SageMaker HyperPod, new capabilities that can reduce time to first token by up to 40% and lower compute costs by up to 25% for long context prompts and multi-turn conversations. These features automatically manage distributed KV caching infrastructure and intelligent request routing, making it easier to deploy production-scale LLM inference workloads with enterprise-grade performance while significantly reducing operational overhead.  ( 113 min )

  • Open

    Apply fine-grained access control with Bedrock AgentCore Gateway interceptors
    We are launching a new feature: gateway interceptors for Amazon Bedrock AgentCore Gateway. This powerful new capability provides fine-grained security, dynamic access control, and flexible schema management.  ( 119 min )
    How Condé Nast accelerated contract processing and rights analysis with Amazon Bedrock
    In this post, we explore how Condé Nast used Amazon Bedrock and Anthropic’s Claude to accelerate their contract processing and rights analysis workstreams. The company’s extensive portfolio, spanning multiple brands and geographies, required managing an increasingly complex web of contracts, rights, and licensing agreements.  ( 113 min )
    Building AI-Powered Voice Applications: Amazon Nova Sonic Telephony Integration Guide
    Available through the Amazon Bedrock bidirectional streaming API, Amazon Nova Sonic can connect to your business data and external tools and can be integrated directly with telephony systems. This post will introduce sample implementations for the most common telephony scenarios.  ( 112 min )
    University of California Los Angeles delivers an immersive theater experience with AWS generative AI services
    In this post, we will walk through the performance constraints and design choices by OARC and REMAP teams at UCLA, including how AWS serverless infrastructure, AWS Managed Services, and generative AI services supported the rapid design and deployment of our solution. We will also describe our use of Amazon SageMaker AI and how it can be used reliably in immersive live experiences.  ( 114 min )
    Optimizing Mobileye’s REM™ with AWS Graviton: A focus on ML inference and Triton integration
    This post is written by Chaim Rand, Principal Engineer, Pini Reisman, Software Senior Principal Engineer, and Eliyah Weinberg, Performance and Technology Innovation Engineer, at Mobileye. The Mobileye team would like to thank Sunita Nadampalli and Guy Almog from AWS for their contributions to this solution and this post. Mobileye is driving the global evolution toward […]  ( 112 min )
    Evaluate models with the Amazon Nova evaluation container using Amazon SageMaker AI
    This blog post introduces the new Amazon Nova model evaluation features in Amazon SageMaker AI. This release adds custom metrics support, LLM-based preference testing, log probability capture, metadata analysis, and multi-node scaling for large evaluations.  ( 120 min )
    Beyond the technology: Workforce changes for AI
    In this post, we explore three essential strategies for successfully integrating AI into your organization: addressing organizational debt before it compounds, embracing distributed decision-making through the "octopus organization" model, and redefining management roles to align with AI-powered workflows. Organizations must invest in both technology and workforce preparation, focusing on streamlining processes, empowering teams with autonomous decision-making within defined parameters, and evolving each management layer from traditional oversight to mentorship, quality assurance, and strategic vision-setting.  ( 107 min )
    Enhanced performance for Amazon Bedrock Custom Model Import
    You can now achieve significant performance improvements when using Amazon Bedrock Custom Model Import, with reduced end-to-end latency, faster time-to-first-token, and improved throughput through advanced PyTorch compilation and CUDA graph optimizations. With Amazon Bedrock Custom Model Import you can to bring your own foundation models to Amazon Bedrock for deployment and inference at scale. In this post, we introduce how to use the improvements in Amazon Bedrock Custom Model Import.  ( 114 min )
    Amazon SageMaker AI introduces EAGLE based adaptive speculative decoding to accelerate generative AI inference
    Amazon SageMaker AI now supports EAGLE-based adaptive speculative decoding, a technique that accelerates large language model inference by up to 2.5x while maintaining output quality. In this post, we explain how to use EAGLE 2 and EAGLE 3 speculative decoding in Amazon SageMaker AI, covering the solution architecture, optimization workflows using your own datasets or SageMaker's built-in data, and benchmark results demonstrating significant improvements in throughput and latency.  ( 112 min )

  • Open

    Train custom computer vision defect detection model using Amazon SageMaker
    In this post, we demonstrate how to migrate computer vision workloads from Amazon Lookout for Vision to Amazon SageMaker AI by training custom defect detection models using pre-trained models available on AWS Marketplace. We provide step-by-step guidance on labeling datasets with SageMaker Ground Truth, training models with flexible hyperparameter configurations, and deploying them for real-time or batch inference—giving you greater control and flexibility for automated quality inspection use cases.  ( 116 min )
    Practical implementation considerations to close the AI value gap
    The AWS Customer Success Center of Excellence (CS COE) helps customers get tangible value from their AWS investments. We've seen a pattern: customers who build AI strategies that address people, process, and technology together succeed more often. In this post, we share practical considerations that can help close the AI value gap.  ( 111 min )
    Introducing bidirectional streaming for real-time inference on Amazon SageMaker AI
    We're introducing bidirectional streaming for Amazon SageMaker AI Inference, which transforms inference from a transactional exchange into a continuous conversation. This post shows you how to build and deploy a container with bidirectional streaming capability to a SageMaker AI endpoint. We also demonstrate how you can bring your own container or use our partner Deepgram's pre-built models and containers on SageMaker AI to enable bi-directional streaming feature for real-time inference.  ( 114 min )
    Warner Bros. Discovery achieves 60% cost savings and faster ML inference with AWS Graviton
    Warner Bros. Discovery (WBD) is a leading global media and entertainment company that creates and distributes the world’s most differentiated and complete portfolio of content and brands across television, film and streaming. In this post, we describe the scale of our offerings, artificial intelligence (AI)/machine learning (ML) inference infrastructure requirements for our real time recommender systems, and how we used AWS Graviton-based Amazon SageMaker AI instances for our ML inference workloads and achieved 60% cost savings and 7% to 60% latency improvements across different models.  ( 110 min )
    Physical AI in practice: Technical foundations that fuel human-machine interactions
    In this post, we explore the complete development lifecycle of physical AI—from data collection and model training to edge deployment—and examine how these intelligent systems learn to understand, reason, and interact with the physical world through continuous feedback loops. We illustrate this workflow through Diligent Robotics' Moxi, a mobile manipulation robot that has completed over 1.2 million deliveries in hospitals, saving nearly 600,000 hours for clinical staff while transforming healthcare logistics and returning valuable time to patient care.  ( 111 min )
    HyperPod now supports Multi-Instance GPU to maximize GPU utilization for generative AI tasks
    In this post, we explore how Amazon SageMaker HyperPod now supports NVIDIA Multi-Instance GPU (MIG) technology, enabling you to partition powerful GPUs into multiple isolated instances for running concurrent workloads like inference, research, and interactive development. By maximizing GPU utilization and reducing wasted resources, MIG helps organizations optimize costs while maintaining performance isolation and predictable quality of service across diverse machine learning tasks.  ( 128 min )

  • Open

    Accelerate generative AI innovation in Canada with Amazon Bedrock cross-Region inference
    We are excited to announce that customers in Canada can now access advanced foundation models including Anthropic's Claude Sonnet 4.5 and Claude Haiku 4.5 on Amazon Bedrock through cross-Region inference (CRIS). This post explores how Canadian organizations can use cross-Region inference profiles from the Canada (Central) Region to access the latest foundation models to accelerate AI initiatives. We will demonstrate how to get started with these new capabilities, provide guidance for migrating from older models, and share recommended practices for quota management.  ( 109 min )
    Power up your ML workflows with interactive IDEs on SageMaker HyperPod
    Amazon SageMaker HyperPod clusters with Amazon Elastic Kubernetes Service (EKS) orchestration now support creating and managing interactive development environments such as JupyterLab and open source Visual Studio Code, streamlining the ML development lifecycle by providing managed environments for familiar tools to data scientists. This post shows how HyperPod administrators can configure Spaces for their clusters, and how data scientists can create and connect to these Spaces.  ( 113 min )
    Claude Opus 4.5 now in Amazon Bedrock
    Anthropic's newest foundation model, Claude Opus 4.5, is now available in Amazon Bedrock, a fully managed service that offers a choice of high-performing foundation models from leading AI companies. In this post, I'll show you what makes this model different, walk through key business applications, and demonstrate how to use Opus 4.5's new tool use capabilities on Amazon Bedrock.  ( 110 min )
    Deploy GPT-OSS models with Amazon Bedrock Custom Model Import
    In this post, we show how to deploy the GPT-OSS-20B model on Amazon Bedrock using Custom Model Import while maintaining complete API compatibility with your current applications.  ( 112 min )
  • Open

    SRE Weekly Issue #498
    View on sreweekly.com A message from our sponsor, Costory: You didn’t sign up to do FinOps. Costory automatically explains why your cloud costs change, and reports it straight to Slack. Built for SREs who want to code, not wrestle with spreadsheets. Now on AWS & GCP Marketplaces. Start your free trial at costory.io Cloudflare outage […]  ( 4 min )

  • Open

    Streamline AI operations with the Multi-Provider Generative AI Gateway reference architecture
    In this post, we introduce the Multi-Provider Generative AI Gateway reference architecture, which provides guidance for deploying LiteLLM into an AWS environment to streamline the management and governance of production generative AI workloads across multiple model providers. This centralized gateway solution addresses common enterprise challenges including provider fragmentation, decentralized governance, operational complexity, and cost management by offering a unified interface that supports Amazon Bedrock, Amazon SageMaker AI, and external providers while maintaining comprehensive security, monitoring, and control capabilities.  ( 111 min )
    Deploy geospatial agents with Foursquare Spatial H3 Hub and Amazon SageMaker AI
    In this post, you'll learn how to deploy geospatial AI agents that can answer complex spatial questions in minutes instead of months. By combining Foursquare Spatial H3 Hub's analysis-ready geospatial data with reasoning models deployed on Amazon SageMaker AI, you can build agents that enable nontechnical domain experts to perform sophisticated spatial analysis through natural language queries—without requiring geographic information system (GIS) expertise or custom data engineering pipelines.  ( 114 min )
    How Wipro PARI accelerates PLC code generation using Amazon Bedrock
    In this post, we share how Wipro implemented advanced prompt engineering techniques, custom validation logic, and automated code rectification to streamline the development of industrial automation code at scale using Amazon Bedrock. We walk through the architecture along with the key use cases, explain core components and workflows, and share real-world results that show the transformative impact on manufacturing operations.  ( 117 min )

  • Open

    MSD explores applying generative Al to improve the deviation management process using AWS services
    This blog post has explores how MSD is harnessing the power of generative AI and databases to optimize and transform its manufacturing deviation management process. By creating an accurate and multifaceted knowledge base of past events, deviations, and findings, the company aims to significantly reduce the time and effort required for each new case while maintaining the highest standards of quality and compliance.  ( 109 min )
    Accelerating genomics variant interpretation with AWS HealthOmics and Amazon Bedrock AgentCore
    In this blog post, we show you how agentic workflows can accelerate the processing and interpretation of genomics pipelines at scale with a natural language interface. We demonstrate a comprehensive genomic variant interpreter agent that combines automated data processing with intelligent analysis to address the entire workflow from raw VCF file ingestion to conversational query interfaces.  ( 115 min )
    How Rufus scales conversational shopping experiences to millions of Amazon customers with Amazon Bedrock
    Our team at Amazon builds Rufus, an AI-powered shopping assistant which delivers intelligent, conversational experiences to delight our customers. More than 250 million customers have used Rufus this year. Monthly users are up 140% YoY and interactions are up 210% YoY. Additionally, customers that use Rufus during a shopping journey are 60% more likely to […]  ( 112 min )
    How Care Access achieved 86% data processing cost reductions and 66% faster data processing with Amazon Bedrock prompt caching
    In this post, we demonstrate how healthcare organizations can securely implement prompt caching technology to streamline medical record processing while maintaining compliance requirements.  ( 112 min )

  • Open

    Claude Code deployment patterns and best practices with Amazon Bedrock
    In this post, we explore deployment patterns and best practices for Claude Code with Amazon Bedrock, covering authentication methods, infrastructure decisions, and monitoring strategies to help enterprises deploy securely at scale. We recommend using Direct IdP integration for authentication, a dedicated AWS account for infrastructure, and OpenTelemetry with CloudWatch dashboards for comprehensive monitoring to ensure secure access, capacity management, and visibility into costs and developer productivity .  ( 116 min )
    Amazon Bedrock Guardrails expands support for code domain
    Amazon Bedrock Guardrails now extends its safety controls to protect code generation across twelve programming languages, addressing critical security challenges in AI-assisted software development. In this post, we explore how to configure content filters, prompt attack detection, denied topics, and sensitive information filters to safeguard against threats like prompt injection, data exfiltration, and malicious code generation while maintaining developer productivity .  ( 113 min )
    Announcing the AWS Well-Architected Responsible AI Lens
    Today, we're announcing the AWS Well-Architected Responsible AI Lens—a set of thoughtful questions and corresponding best practices that help builders address responsible AI concerns throughout development and operation.  ( 118 min )
    How Amazon uses AI agents to support compliance screening of billions of transactions per day
    Amazon's AI-powered Amazon Compliance Screening system tackles complex compliance challenges through autonomous agents that analyze, reason through, and resolve cases with precision. This blog post explores how Amazon’s Compliance team built its AI-powered investigation system through a series of AI agents built on AWS.  ( 123 min )
    Build an agentic solution with Amazon Nova, Snowflake, and LangGraph
    In this post, we cover how you can use tools from Snowflake AI Data Cloud and Amazon Web Services (AWS) to build generative AI solutions that organizations can use to make data-driven decisions, increase operational efficiency, and ultimately gain a competitive edge.  ( 129 min )
    Using Spectrum fine-tuning to improve FM training efficiency on Amazon SageMaker AI
    In this post you will learn how to use Spectrum to optimize resource use and shorten training times without sacrificing quality, as well as how to implement Spectrum fine-tuning with Amazon SageMaker AI training jobs. We will also discuss the tradeoff between QLoRA and Spectrum fine-tuning, showing that while QLoRA is more resource efficient, Spectrum results in higher performance overall.  ( 125 min )

  • Open

    Bringing tic-tac-toe to life with AWS AI services
    RoboTic-Tac-Toe is an interactive game where two physical robots move around a tic-tac-toe board, with both the gameplay and robots’ movements orchestrated by LLMs. Players can control the robots using natural language commands, directing them to place their markers on the game board. In this post, we explore the architecture and prompt engineering techniques used to reason about a tic-tac-toe game and decide the next best game strategy and movement plan for the current player.  ( 121 min )
    HyperPod enhances ML infrastructure with security and storage
    This blog post introduces two major enhancements to Amazon SageMaker HyperPod that strengthen security and storage capabilities for large-scale machine learning infrastructure. The new features include customer managed key (CMK) support for encrypting EBS volumes with organization-controlled encryption keys, and Amazon EBS CSI driver integration that enables dynamic storage management for Kubernetes volumes in AI workloads.  ( 126 min )
    Accelerating generative AI applications with a platform engineering approach
    In this post, I will illustrate how applying platform engineering principles to generative AI unlocks faster time-to-value, cost control, and scalable innovation.  ( 123 min )

  • Open

    Your complete guide to Amazon Quick Suite at AWS re:Invent 2025
    This year, re:Invent will be held in Las Vegas, Nevada, from December 1 to December 5, 2025, and this guide will help you navigate our comprehensive session catalog and plan your week. The sessions cater to business and technology leaders, product and engineering teams, and data and analytics teams interested in incorporating agentic AI capabilities across their teams and organization.  ( 122 min )
    Accelerate enterprise solutions with agentic AI-powered consulting: Introducing AWS Professional Service Agents
    I'm excited to announce AWS Professional Services now offers specialized AI agents including the AWS Professional Services Delivery Agent. This represents a transformation to the consulting experience that embeds intelligent agents throughout the consulting life cycle to deliver better value for customers.  ( 119 min )
    Amazon Bedrock AgentCore and Claude: Transforming business with agentic AI
    In this post, we explore how Amazon Bedrock AgentCore and Claude are enabling enterprises like Cox Automotive and Druva to deploy production-ready agentic AI systems that deliver measurable business value, with results including up to 63% autonomous issue resolution and 58% faster response times. We examine the technical foundation combining Claude's frontier AI capabilities with AgentCore's enterprise-grade infrastructure that allows organizations to focus on agent logic rather than building complex operational systems from scratch.  ( 124 min )
  • Open

    SRE Weekly Issue #497
    View on sreweekly.com A message from our sponsor, Costory: You didn’t sign up to do FinOps.Costory automatically explains why your cloud costs change, and reports it straight to Slack.Built for SREs who want to code, not wrestle with spreadsheets.Now on AWS & GCP Marketplaces. Start your free trial at costory.io To vibe or not to […]  ( 4 min )

  • Open

    Build a biomedical research agent with Biomni tools and Amazon Bedrock AgentCore Gateway
    In this post, we demonstrate how to build a production-ready biomedical research agent by integrating Biomni's specialized tools with Amazon Bedrock AgentCore Gateway, enabling researchers to access over 30 biomedical databases through a secure, scalable infrastructure. The implementation showcases how to transform research prototypes into enterprise-grade systems with persistent memory, semantic tool discovery, and comprehensive observability for scientific reproducibility .  ( 127 min )
    Make your web apps hands-free with Amazon Nova Sonic
    Graphical user interfaces have carried the torch for decades, but today’s users increasingly expect to talk to their applications. In this post we show how we added a true voice-first experience to a reference application—the Smart Todo App—turning routine task management into a fluid, hands-free conversation.  ( 120 min )
    Harnessing the power of generative AI: Druva’s multi-agent copilot for streamlined data protection
    Generative AI is transforming the way businesses interact with their customers and revolutionizing conversational interfaces for complex IT operations. Druva, a leading provider of data security solutions, is at the forefront of this transformation. In collaboration with Amazon Web Services (AWS), Druva is developing a cutting-edge generative AI-powered multi-agent copilot that aims to redefine the customer experience in data security and cyber resilience.  ( 123 min )

  • Open

    凤鸣山
    凤鸣山问丹 上虞的秋日,凤鸣山薄雾如纱。我沿石阶缓步而上,两旁古木参天,溪水泠泠作响,仿佛仍回荡着千年前炉火噼啪的余音。山腰处,一方“炼丹井”静卧于苔痕斑驳的石栏内,井水幽深,映不出魏伯阳的身影,却照见自己模糊的轮廓——这方寸水土,曾是他与弟子们埋首炼丹、叩问长生的地方。 山间传说里,魏伯阳炼成金丹后,以犬试药,犬倒地如死。两位弟子心生疑惧,弃丹而去;唯有一人笃信师道,愿同服共死。结果师徒与犬皆复苏飞升,而犹豫者终老山下。故事如山风拂过耳际,初听是仙迹,细思却如井水般冰凉:那“坚定”的信念,究竟该交付给谁?若丹本为毒,那“信”岂非成了引向深渊的绳索? 下山途中,偶遇一位老农在田埂上歇息。他笑谈:“如今谁还信什么金丹?但人总得信点什么吧,不然日子怎么过?”他信的是节气、是土地、是春种秋收的踏实。这朴素之“尚”,与魏伯阳的玄奥丹道相隔千年,却同样支撑着一种生活。吴晗先生曾言,社会风气之“尚”如潮汐涨落,或尚名节,或尚功利,或尚清谈——可无论何种“尚”,若只知随波逐流,不加省察,便如那两位弃丹弟子,既失了飞升之机,也未必真得了安稳。 归途车窗外,城市灯火渐次亮起,霓虹如新的“丹炉”,闪烁着财富、流量、速度的诱人光泽。我们这一代人,何尝不在各自的时代“炼丹”?只是炉中所炼,早已不是铅汞,而是对意义、价值与归属的渴求。然而,若只知追逐那最耀眼的光,却不肯停下脚步自问一句“此丹可服否?”,恐怕终将陷入另一种“假死”——身体活着,灵魂却已沉睡。 凤鸣山无凤鸣,唯有风过林梢。魏伯阳的丹炉早已冷透,但那口炼丹井却像一面古镜,照见所有时代人心深处的叩问:我们该信什么?又为何而信? 苏格拉底说:“未经省察的人生不值得过。”或许真正的“丹”,并非服食之物,而是那敢于审视自身信念的勇气——它不许诺飞升,却能让双脚在尘世站得更稳,让眼睛在迷雾中看得更清。  ( 1 min )

  • Open

    Introducing agent-to-agent protocol support in Amazon Bedrock AgentCore Runtime
    In this post, we demonstrate how you can use the A2A protocol for AI agents built with different frameworks to collaborate seamlessly. You'll learn how to deploy A2A servers on AgentCore Runtime, configure agent discovery and authentication, and build a real-world multi-agent system for incident response. We'll cover the complete A2A request lifecycle, from agent card discovery to task delegation, showing how standardized protocols eliminate the complexity of multi-agent coordination.  ( 126 min )
    Powering enterprise search with the Cohere Embed 4 multimodal embeddings model in Amazon Bedrock
    The Cohere Embed 4 multimodal embeddings model is now available as a fully managed, serverless option in Amazon Bedrock. In this post, we dive into the benefits and unique capabilities of Embed 4 for enterprise search use cases. We’ll show you how to quickly get started using Embed 4 on Amazon Bedrock, taking advantage of integrations with Strands Agents, S3 Vectors, and Amazon Bedrock AgentCore to build powerful agentic retrieval-augmented generation (RAG) workflows.  ( 123 min )
    A guide to building AI agents in GxP environments
    The regulatory landscape for GxP compliance is evolving to address the unique characteristics of AI. Traditional Computer System Validation (CSV) approaches, often with uniform validation strategies, are being supplemented by Computer Software Assurance (CSA) frameworks that emphasize flexible risk-based validation methods tailored to each system's actual impact and complexity (FDA latest guidance). In this post, we cover a risk-based implementation, practical implementation considerations across different risk levels, the AWS shared responsibility model for compliance, and concrete examples of risk mitigation strategies.  ( 125 min )
    Multi-Agent collaboration patterns with Strands Agents and Amazon Nova
    In this post, we explore four key collaboration patterns for multi-agent, multimodal AI systems – Agents as Tools, Swarms Agents, Agent Graphs, and Agent Workflows – and discuss when and how to apply each using the open-source AWS Strands Agents SDK with Amazon Nova models.  ( 131 min )
  • Open

    Checking that Docker image manifests are complete
    Background We build a bunch of stuff for RISC-V using the Dart official Docker image, but the RISC-V images can often arrive some time (days) after the more mainstream images[1]. That means that if we merge a Dependabot PR for an updated image it might well be missing RISC-V, causing the Continuous Delivery (CD) pipeline […]  ( 12 min )
    Checking that Docker image manifests are complete
    Background We build a bunch of stuff for RISC-V using the Dart official Docker image, but the RISC-V images can often arrive some time (days) after the more mainstream images[1]. That means that if we merge a Dependabot PR for an updated image it might well be missing RISC-V, causing the Continuous Delivery (CD) pipeline […]  ( 12 min )

  • Open

    Fine-tune VLMs for multipage document-to-JSON with SageMaker AI and SWIFT
    In this post, we demonstrate that fine-tuning VLMs provides a powerful and flexible approach to automate and significantly enhance document understanding capabilities. We also demonstrate that using focused fine-tuning allows smaller, multi-modal models to compete effectively with much larger counterparts (98% accuracy with Qwen2.5 VL 3B).  ( 132 min )
    How Clario automates clinical research analysis using generative AI on AWS
    In this post, we demonstrate how Clario has used Amazon Bedrock and other AWS services to build an AI-powered solution that automates and improves the analysis of COA interviews.  ( 121 min )
  • Open

    SRE Weekly Issue #496
    View on sreweekly.com A message from our sponsor, CodeRabbit: CodeRabbit is your AI co-pilot for code reviews. Get instant code review feedback, one-click fix suggestions and define custom rules with AST Grep to catch subtle issues static tools miss. Trusted across 1M repos and 70K open-source projects. ☞ Get Started Today The hidden trade-offs of […]  ( 4 min )

  • Open

    Connect Amazon Bedrock agents to cross-account knowledge bases
    Organizations need seamless access to their structured data repositories to power intelligent AI agents. However, when these resources span multiple AWS accounts integration challenges can arise. This post explores a practical solution for connecting Amazon Bedrock agents to knowledge bases in Amazon Redshift clusters residing in different AWS accounts.  ( 122 min )
    Democratizing AI: How Thomson Reuters Open Arena supports no-code AI for every professional with Amazon Bedrock
    In this blog post, we explore how TR addressed key business use cases with Open Arena, a highly scalable and flexible no-code AI solution powered by Amazon Bedrock and other AWS services such as Amazon OpenSearch Service, Amazon Simple Storage Service (Amazon S3), Amazon DynamoDB, and AWS Lambda. We'll explain how TR used AWS services to build this solution, including how the architecture was designed, the use cases it solves, and the business profiles that use it.  ( 123 min )
    Introducing structured output for Custom Model Import in Amazon Bedrock
    Today, we are excited to announce the addition of structured output to Custom Model Import. Structured output constrains a model's generation process in real time so that every token it produces conforms to a schema you define. Rather than relying on prompt-engineering tricks or brittle post-processing scripts, you can now generate structured outputs directly at inference time.  ( 121 min )

  • Open

    Transform your MCP architecture: Unite MCP servers through AgentCore Gateway
    Earlier this year, we introduced Amazon Bedrock AgentCore Gateway, a fully managed service that serves as a centralized MCP tool server, providing a unified interface where agents can discover, access, and invoke tools. Today, we're extending support for existing MCP servers as a new target type in AgentCore Gateway. With this capability, you can group multiple task-specific MCP servers aligned to agent goals behind a single, manageable MCP gateway interface. This reduces the operational complexity of maintaining separate gateways, while providing the same centralized tool and authentication management that existed for REST APIs and AWS Lambda functions.  ( 127 min )

  • Open

    How Amazon Search increased ML training twofold using AWS Batch for Amazon SageMaker Training jobs
    In this post, we show you how Amazon Search optimized GPU instance utilization by leveraging AWS Batch for SageMaker Training jobs. This managed solution enabled us to orchestrate machine learning (ML) training workloads on GPU-accelerated instance families like P5, P4, and others. We will also provide a step-by-step walkthrough of the use case implementation.  ( 124 min )
  • Open

    Don’t huff the fumes
    TL;DR Agentic systems are the latest thing being used to solve IT integration issues, becoming the glue squirted into the gaps between systems. But the use of natural language means that the distinction between ‘data’ and ‘code’ is almost impossible to make, which causes a whole raft of security concerns. This new glue may be […]  ( 13 min )
    Don’t huff the fumes
    TL;DR Agentic systems are the latest thing being used to solve IT integration issues, becoming the glue squirted into the gaps between systems. But the use of natural language means that the distinction between ‘data’ and ‘code’ is almost impossible to make, which causes a whole raft of security concerns. This new glue may be […]  ( 13 min )

  • Open

    Iterate faster with Amazon Bedrock AgentCore Runtime direct code deployment
    Amazon Bedrock AgentCore is an agentic platform for building, deploying, and operating effective agents securely at scale. Amazon Bedrock AgentCore Runtime is a fully managed service of Bedrock AgentCore, which provides low latency serverless environments to deploy agents and tools. It provides session isolation, supports multiple agent frameworks including popular open-source frameworks, and handles multimodal […]  ( 120 min )

  • Open

    How Switchboard, MD automates real-time call transcription in clinical contact centers with Amazon Nova Sonic
    In this post, we examine the specific challenges Switchboard, MD faced with scaling transcription accuracy and cost-effectiveness in clinical environments, their evaluation process for selecting the right transcription solution, and the technical architecture they implemented using Amazon Connect and Amazon Kinesis Video Streams. This post details the impressive results achieved and demonstrates how they were able to use this foundation to automate EMR matching and give healthcare staff more time to focus on patient care.  ( 119 min )
  • Open

    October 2025
    Pupdate The central heating went on a few days into the month, and it was also soon time for the boys to be wearing their coats out. Interactive Ball Toy Having learned my lesson about dodgy drop shippers last month I ordered from AliExpress when $wife found a fun looking toy in some Dachshund forum. […]  ( 14 min )
    October 2025
    Pupdate The central heating went on a few days into the month, and it was also soon time for the boys to be wearing their coats out. Interactive Ball Toy Having learned my lesson about dodgy drop shippers last month I ordered from AliExpress when $wife found a fun looking toy in some Dachshund forum. […]  ( 14 min )
  • Open

    SRE Weekly Issue #495
    View on sreweekly.com I’m back! Kidney donation was a fascinating and rewarding experience, and I encourage you to learn more. It’s amazing how it’s possible to fix one human with spare parts from another! I’ll share more about my experience later, but for now: thank you to the many of you that reached out with […]  ( 4 min )
2025-11-30T10:17:12.132Z osmosfeed 1.15.1