• Open

    Real-world reasoning: How Amazon Nova Lite 2.0 handles complex customer support scenarios
    This post evaluates the reasoning capabilities of our latest offering in the Nova family, Amazon Nova Lite 2.0, using practical scenarios that test these critical dimensions. We compare its performance against other models in the Nova family—Lite 1.0, Micro, Pro 1.0, and Premier—to elucidate how the latest version advances reasoning quality and consistency.  ( 116 min )
    Create AI-powered chat assistants for your enterprise with Amazon Quick Suite
    In this post, we show how to build chat agents in Amazon Quick Suite. We walk through a three-layer framework—identity, instructions, and knowledge—that transforms Quick Suite chat agents into intelligent enterprise AI assistants. In our example, we demonstrate how our chat agent guides feature discovery, use enterprise data to inform recommendations, and tailors solutions based on potential to impact and your team’s adoption readiness.  ( 116 min )

  • Open

    How AWS delivers generative AI to the public sector in weeks, not years
    Experts at the Generative AI Innovation Center share several strategies to help organizations excel with generative AI.  ( 110 min )
    S&P Global Data integration expands Amazon Quick Research capabilities
    Today, we are pleased to announce a new integration between Amazon Quick Research and S&P Global. This integration brings both S&P Global Energy news, research, and insights and S&P Global Market Intelligence data to Quick Research customers in one deep research agent. In this post, we explore S&P Global’s data sets and the solution architecture of the integration with Quick Research.  ( 109 min )
    Streamline AI agent tool interactions: Connect API Gateway to AgentCore Gateway with MCP
    AgentCore Gateway now supports API GatewayAs organizations explore the possibilities of agentic applications, they continue to navigate challenges of using enterprise data as context in invocation requests to large language models (LLMs) in a manner that is secure and aligned with enterprise policies. This post covers these new capabilities and shows how to implement them.  ( 111 min )
    Create an intelligent insurance underwriter agent powered by Amazon Nova 2 Lite and Amazon Quick Suite
    In this post, we demonstrate how to build an intelligent insurance underwriting agent that addresses three critical challenges: unifying siloed data across CRM systems and databases, providing explainable and auditable AI decisions for regulatory compliance, and enabling automated fraud detection with consistent underwriting rules. The solution combines Amazon Nova 2 Lite for transparent risk assessment, Amazon Bedrock AgentCore for managed MCP server infrastructure, and Amazon Quick Suite for natural language interactions—delivering a production-ready system that underwriters can deploy in under 30 minutes .  ( 111 min )
  • Open

    SRE Weekly Issue #500
    View on sreweekly.com A message from our sponsor, Depot: Stop hunting through GitHub Actions logs. Depot now offers powerful CI log search across all your repositories and workflows. With smart filtering by timeframe, runner type, and keywords, you’ll have all the information at your fingertips to debug faster. Wow, five hundred issues! I sent the […]  ( 4 min )

  • Open

    Milo cancer diary part 21 – CHOP #4
    Milo’s had a fantastic long remission – it’s been almost nine months since his last chemo. Long enough that we started hoping for a miracle, and that he might not relapse again. But… the good folk at North Downs Specialist Referrals (NDSR) were right to be concerned about his last scan, and get him back […]  ( 13 min )

  • Open

    November 2025
    Pupdate It’s been pretty cold and wet, so the boys are needing to wear their coats outside. Milo had a scan at the start of the month. Initially things were looking good, and the plan was to stretch out the next visit to three months time :) But then the technician noticed some lymph node […]  ( 14 min )
    November 2025
    Pupdate It’s been pretty cold and wet, so the boys are needing to wear their coats outside. Milo had a scan at the start of the month. Initially things were looking good, and the plan was to stretch out the next visit to three months time :) But then the technician noticed some lymph node […]  ( 14 min )
  • Open

    SRE Weekly Issue #499
    View on sreweekly.com Fix-mas Countdown The folks at Uptime Labs and Advanced Capacity Labs have announced an advent calendar for this December. Note: In order to take part, you’ll need to provide an email address to subscribe. I gave that some serious thought before including this here, but ultimately, I have a lot of trust […]  ( 4 min )

  • Open

    Clocking on at the Outrage Factory
    TL;DR Our online discourse is the victim of industrial scale pollution, and the incentives are being aligned in the wrong direction. Rather than polluters being penalised there’s now an entire industry that’s paid to pollute. Filter Failure at the Outrage Factory is no longer just the work of ‘amateur’ fringe trolls and state sponsored propaganda; […]  ( 14 min )
    Clocking on at the Outrage Factory
    TL;DR Our online discourse is the victim of industrial scale pollution, and the incentives are being aligned in the wrong direction. Rather than polluters being penalised there’s now an entire industry that’s paid to pollute. Filter Failure at the Outrage Factory is no longer just the work of ‘amateur’ fringe trolls and state sponsored propaganda; […]  ( 14 min )

  • Open

    How Myriad Genetics achieved fast, accurate, and cost-efficient document processing using the AWS open-source Generative AI Intelligent Document Processing Accelerator
    In this post, we explore how Myriad Genetics partnered with the AWS Generative AI Innovation Center to transform their healthcare document processing pipeline using Amazon Bedrock and Amazon Nova foundation models, achieving 98% classification accuracy while reducing costs by 77% and processing time by 80%. We detail the technical implementation using AWS's open-source GenAI Intelligent Document Processing Accelerator, the optimization strategies for document classification and key information extraction, and the measurable business impact on Myriad's prior authorization workflows.  ( 115 min )
    How CBRE powers unified property management search and digital assistant using Amazon Bedrock
    In this post, CBRE and AWS demonstrate how they transformed property management by building a unified search and digital assistant using Amazon Bedrock, enabling professionals to access millions of documents and multiple databases through natural language queries. The solution combines Amazon Nova Pro for SQL generation and Claude Haiku for document interactions, achieving a 67% reduction in processing time while maintaining enterprise-grade security across more than eight million documents.  ( 118 min )
    Managed Tiered KV Cache and Intelligent Routing for Amazon SageMaker HyperPod
    In this post, we introduce Managed Tiered KV Cache and Intelligent Routing for Amazon SageMaker HyperPod, new capabilities that can reduce time to first token by up to 40% and lower compute costs by up to 25% for long context prompts and multi-turn conversations. These features automatically manage distributed KV caching infrastructure and intelligent request routing, making it easier to deploy production-scale LLM inference workloads with enterprise-grade performance while significantly reducing operational overhead.  ( 113 min )

  • Open

    Apply fine-grained access control with Bedrock AgentCore Gateway interceptors
    We are launching a new feature: gateway interceptors for Amazon Bedrock AgentCore Gateway. This powerful new capability provides fine-grained security, dynamic access control, and flexible schema management.  ( 119 min )
    How Condé Nast accelerated contract processing and rights analysis with Amazon Bedrock
    In this post, we explore how Condé Nast used Amazon Bedrock and Anthropic’s Claude to accelerate their contract processing and rights analysis workstreams. The company’s extensive portfolio, spanning multiple brands and geographies, required managing an increasingly complex web of contracts, rights, and licensing agreements.  ( 113 min )
    Building AI-Powered Voice Applications: Amazon Nova Sonic Telephony Integration Guide
    Available through the Amazon Bedrock bidirectional streaming API, Amazon Nova Sonic can connect to your business data and external tools and can be integrated directly with telephony systems. This post will introduce sample implementations for the most common telephony scenarios.  ( 112 min )
    University of California Los Angeles delivers an immersive theater experience with AWS generative AI services
    In this post, we will walk through the performance constraints and design choices by OARC and REMAP teams at UCLA, including how AWS serverless infrastructure, AWS Managed Services, and generative AI services supported the rapid design and deployment of our solution. We will also describe our use of Amazon SageMaker AI and how it can be used reliably in immersive live experiences.  ( 114 min )
    Optimizing Mobileye’s REM™ with AWS Graviton: A focus on ML inference and Triton integration
    This post is written by Chaim Rand, Principal Engineer, Pini Reisman, Software Senior Principal Engineer, and Eliyah Weinberg, Performance and Technology Innovation Engineer, at Mobileye. The Mobileye team would like to thank Sunita Nadampalli and Guy Almog from AWS for their contributions to this solution and this post. Mobileye is driving the global evolution toward […]  ( 112 min )
    Evaluate models with the Amazon Nova evaluation container using Amazon SageMaker AI
    This blog post introduces the new Amazon Nova model evaluation features in Amazon SageMaker AI. This release adds custom metrics support, LLM-based preference testing, log probability capture, metadata analysis, and multi-node scaling for large evaluations.  ( 120 min )
    Beyond the technology: Workforce changes for AI
    In this post, we explore three essential strategies for successfully integrating AI into your organization: addressing organizational debt before it compounds, embracing distributed decision-making through the "octopus organization" model, and redefining management roles to align with AI-powered workflows. Organizations must invest in both technology and workforce preparation, focusing on streamlining processes, empowering teams with autonomous decision-making within defined parameters, and evolving each management layer from traditional oversight to mentorship, quality assurance, and strategic vision-setting.  ( 107 min )
    Enhanced performance for Amazon Bedrock Custom Model Import
    You can now achieve significant performance improvements when using Amazon Bedrock Custom Model Import, with reduced end-to-end latency, faster time-to-first-token, and improved throughput through advanced PyTorch compilation and CUDA graph optimizations. With Amazon Bedrock Custom Model Import you can to bring your own foundation models to Amazon Bedrock for deployment and inference at scale. In this post, we introduce how to use the improvements in Amazon Bedrock Custom Model Import.  ( 114 min )
    Amazon SageMaker AI introduces EAGLE based adaptive speculative decoding to accelerate generative AI inference
    Amazon SageMaker AI now supports EAGLE-based adaptive speculative decoding, a technique that accelerates large language model inference by up to 2.5x while maintaining output quality. In this post, we explain how to use EAGLE 2 and EAGLE 3 speculative decoding in Amazon SageMaker AI, covering the solution architecture, optimization workflows using your own datasets or SageMaker's built-in data, and benchmark results demonstrating significant improvements in throughput and latency.  ( 112 min )

  • Open

    Train custom computer vision defect detection model using Amazon SageMaker
    In this post, we demonstrate how to migrate computer vision workloads from Amazon Lookout for Vision to Amazon SageMaker AI by training custom defect detection models using pre-trained models available on AWS Marketplace. We provide step-by-step guidance on labeling datasets with SageMaker Ground Truth, training models with flexible hyperparameter configurations, and deploying them for real-time or batch inference—giving you greater control and flexibility for automated quality inspection use cases.  ( 116 min )
    Practical implementation considerations to close the AI value gap
    The AWS Customer Success Center of Excellence (CS COE) helps customers get tangible value from their AWS investments. We've seen a pattern: customers who build AI strategies that address people, process, and technology together succeed more often. In this post, we share practical considerations that can help close the AI value gap.  ( 111 min )
    Introducing bidirectional streaming for real-time inference on Amazon SageMaker AI
    We're introducing bidirectional streaming for Amazon SageMaker AI Inference, which transforms inference from a transactional exchange into a continuous conversation. This post shows you how to build and deploy a container with bidirectional streaming capability to a SageMaker AI endpoint. We also demonstrate how you can bring your own container or use our partner Deepgram's pre-built models and containers on SageMaker AI to enable bi-directional streaming feature for real-time inference.  ( 114 min )
    Warner Bros. Discovery achieves 60% cost savings and faster ML inference with AWS Graviton
    Warner Bros. Discovery (WBD) is a leading global media and entertainment company that creates and distributes the world’s most differentiated and complete portfolio of content and brands across television, film and streaming. In this post, we describe the scale of our offerings, artificial intelligence (AI)/machine learning (ML) inference infrastructure requirements for our real time recommender systems, and how we used AWS Graviton-based Amazon SageMaker AI instances for our ML inference workloads and achieved 60% cost savings and 7% to 60% latency improvements across different models.  ( 110 min )
    Physical AI in practice: Technical foundations that fuel human-machine interactions
    In this post, we explore the complete development lifecycle of physical AI—from data collection and model training to edge deployment—and examine how these intelligent systems learn to understand, reason, and interact with the physical world through continuous feedback loops. We illustrate this workflow through Diligent Robotics' Moxi, a mobile manipulation robot that has completed over 1.2 million deliveries in hospitals, saving nearly 600,000 hours for clinical staff while transforming healthcare logistics and returning valuable time to patient care.  ( 111 min )
    HyperPod now supports Multi-Instance GPU to maximize GPU utilization for generative AI tasks
    In this post, we explore how Amazon SageMaker HyperPod now supports NVIDIA Multi-Instance GPU (MIG) technology, enabling you to partition powerful GPUs into multiple isolated instances for running concurrent workloads like inference, research, and interactive development. By maximizing GPU utilization and reducing wasted resources, MIG helps organizations optimize costs while maintaining performance isolation and predictable quality of service across diverse machine learning tasks.  ( 128 min )

  • Open

    Accelerate generative AI innovation in Canada with Amazon Bedrock cross-Region inference
    We are excited to announce that customers in Canada can now access advanced foundation models including Anthropic's Claude Sonnet 4.5 and Claude Haiku 4.5 on Amazon Bedrock through cross-Region inference (CRIS). This post explores how Canadian organizations can use cross-Region inference profiles from the Canada (Central) Region to access the latest foundation models to accelerate AI initiatives. We will demonstrate how to get started with these new capabilities, provide guidance for migrating from older models, and share recommended practices for quota management.  ( 109 min )
    Power up your ML workflows with interactive IDEs on SageMaker HyperPod
    Amazon SageMaker HyperPod clusters with Amazon Elastic Kubernetes Service (EKS) orchestration now support creating and managing interactive development environments such as JupyterLab and open source Visual Studio Code, streamlining the ML development lifecycle by providing managed environments for familiar tools to data scientists. This post shows how HyperPod administrators can configure Spaces for their clusters, and how data scientists can create and connect to these Spaces.  ( 113 min )
    Claude Opus 4.5 now in Amazon Bedrock
    Anthropic's newest foundation model, Claude Opus 4.5, is now available in Amazon Bedrock, a fully managed service that offers a choice of high-performing foundation models from leading AI companies. In this post, I'll show you what makes this model different, walk through key business applications, and demonstrate how to use Opus 4.5's new tool use capabilities on Amazon Bedrock.  ( 110 min )
    Deploy GPT-OSS models with Amazon Bedrock Custom Model Import
    In this post, we show how to deploy the GPT-OSS-20B model on Amazon Bedrock using Custom Model Import while maintaining complete API compatibility with your current applications.  ( 112 min )
  • Open

    SRE Weekly Issue #498
    View on sreweekly.com A message from our sponsor, Costory: You didn’t sign up to do FinOps. Costory automatically explains why your cloud costs change, and reports it straight to Slack. Built for SREs who want to code, not wrestle with spreadsheets. Now on AWS & GCP Marketplaces. Start your free trial at costory.io Cloudflare outage […]  ( 4 min )

  • Open

    Streamline AI operations with the Multi-Provider Generative AI Gateway reference architecture
    In this post, we introduce the Multi-Provider Generative AI Gateway reference architecture, which provides guidance for deploying LiteLLM into an AWS environment to streamline the management and governance of production generative AI workloads across multiple model providers. This centralized gateway solution addresses common enterprise challenges including provider fragmentation, decentralized governance, operational complexity, and cost management by offering a unified interface that supports Amazon Bedrock, Amazon SageMaker AI, and external providers while maintaining comprehensive security, monitoring, and control capabilities.  ( 111 min )
    Deploy geospatial agents with Foursquare Spatial H3 Hub and Amazon SageMaker AI
    In this post, you'll learn how to deploy geospatial AI agents that can answer complex spatial questions in minutes instead of months. By combining Foursquare Spatial H3 Hub's analysis-ready geospatial data with reasoning models deployed on Amazon SageMaker AI, you can build agents that enable nontechnical domain experts to perform sophisticated spatial analysis through natural language queries—without requiring geographic information system (GIS) expertise or custom data engineering pipelines.  ( 114 min )
    How Wipro PARI accelerates PLC code generation using Amazon Bedrock
    In this post, we share how Wipro implemented advanced prompt engineering techniques, custom validation logic, and automated code rectification to streamline the development of industrial automation code at scale using Amazon Bedrock. We walk through the architecture along with the key use cases, explain core components and workflows, and share real-world results that show the transformative impact on manufacturing operations.  ( 117 min )

  • Open

    MSD explores applying generative Al to improve the deviation management process using AWS services
    This blog post has explores how MSD is harnessing the power of generative AI and databases to optimize and transform its manufacturing deviation management process. By creating an accurate and multifaceted knowledge base of past events, deviations, and findings, the company aims to significantly reduce the time and effort required for each new case while maintaining the highest standards of quality and compliance.  ( 109 min )
    Accelerating genomics variant interpretation with AWS HealthOmics and Amazon Bedrock AgentCore
    In this blog post, we show you how agentic workflows can accelerate the processing and interpretation of genomics pipelines at scale with a natural language interface. We demonstrate a comprehensive genomic variant interpreter agent that combines automated data processing with intelligent analysis to address the entire workflow from raw VCF file ingestion to conversational query interfaces.  ( 115 min )
    How Rufus scales conversational shopping experiences to millions of Amazon customers with Amazon Bedrock
    Our team at Amazon builds Rufus, an AI-powered shopping assistant which delivers intelligent, conversational experiences to delight our customers. More than 250 million customers have used Rufus this year. Monthly users are up 140% YoY and interactions are up 210% YoY. Additionally, customers that use Rufus during a shopping journey are 60% more likely to […]  ( 112 min )
    How Care Access achieved 86% data processing cost reductions and 66% faster data processing with Amazon Bedrock prompt caching
    In this post, we demonstrate how healthcare organizations can securely implement prompt caching technology to streamline medical record processing while maintaining compliance requirements.  ( 112 min )

  • Open

    Claude Code deployment patterns and best practices with Amazon Bedrock
    In this post, we explore deployment patterns and best practices for Claude Code with Amazon Bedrock, covering authentication methods, infrastructure decisions, and monitoring strategies to help enterprises deploy securely at scale. We recommend using Direct IdP integration for authentication, a dedicated AWS account for infrastructure, and OpenTelemetry with CloudWatch dashboards for comprehensive monitoring to ensure secure access, capacity management, and visibility into costs and developer productivity .  ( 116 min )
    Amazon Bedrock Guardrails expands support for code domain
    Amazon Bedrock Guardrails now extends its safety controls to protect code generation across twelve programming languages, addressing critical security challenges in AI-assisted software development. In this post, we explore how to configure content filters, prompt attack detection, denied topics, and sensitive information filters to safeguard against threats like prompt injection, data exfiltration, and malicious code generation while maintaining developer productivity .  ( 113 min )
    Announcing the AWS Well-Architected Responsible AI Lens
    Today, we're announcing the AWS Well-Architected Responsible AI Lens—a set of thoughtful questions and corresponding best practices that help builders address responsible AI concerns throughout development and operation.  ( 118 min )
    How Amazon uses AI agents to support compliance screening of billions of transactions per day
    Amazon's AI-powered Amazon Compliance Screening system tackles complex compliance challenges through autonomous agents that analyze, reason through, and resolve cases with precision. This blog post explores how Amazon’s Compliance team built its AI-powered investigation system through a series of AI agents built on AWS.  ( 123 min )
    Build an agentic solution with Amazon Nova, Snowflake, and LangGraph
    In this post, we cover how you can use tools from Snowflake AI Data Cloud and Amazon Web Services (AWS) to build generative AI solutions that organizations can use to make data-driven decisions, increase operational efficiency, and ultimately gain a competitive edge.  ( 129 min )
    Using Spectrum fine-tuning to improve FM training efficiency on Amazon SageMaker AI
    In this post you will learn how to use Spectrum to optimize resource use and shorten training times without sacrificing quality, as well as how to implement Spectrum fine-tuning with Amazon SageMaker AI training jobs. We will also discuss the tradeoff between QLoRA and Spectrum fine-tuning, showing that while QLoRA is more resource efficient, Spectrum results in higher performance overall.  ( 125 min )

  • Open

    Bringing tic-tac-toe to life with AWS AI services
    RoboTic-Tac-Toe is an interactive game where two physical robots move around a tic-tac-toe board, with both the gameplay and robots’ movements orchestrated by LLMs. Players can control the robots using natural language commands, directing them to place their markers on the game board. In this post, we explore the architecture and prompt engineering techniques used to reason about a tic-tac-toe game and decide the next best game strategy and movement plan for the current player.  ( 121 min )
    HyperPod enhances ML infrastructure with security and storage
    This blog post introduces two major enhancements to Amazon SageMaker HyperPod that strengthen security and storage capabilities for large-scale machine learning infrastructure. The new features include customer managed key (CMK) support for encrypting EBS volumes with organization-controlled encryption keys, and Amazon EBS CSI driver integration that enables dynamic storage management for Kubernetes volumes in AI workloads.  ( 126 min )
    Accelerating generative AI applications with a platform engineering approach
    In this post, I will illustrate how applying platform engineering principles to generative AI unlocks faster time-to-value, cost control, and scalable innovation.  ( 123 min )

  • Open

    Your complete guide to Amazon Quick Suite at AWS re:Invent 2025
    This year, re:Invent will be held in Las Vegas, Nevada, from December 1 to December 5, 2025, and this guide will help you navigate our comprehensive session catalog and plan your week. The sessions cater to business and technology leaders, product and engineering teams, and data and analytics teams interested in incorporating agentic AI capabilities across their teams and organization.  ( 122 min )
    Accelerate enterprise solutions with agentic AI-powered consulting: Introducing AWS Professional Service Agents
    I'm excited to announce AWS Professional Services now offers specialized AI agents including the AWS Professional Services Delivery Agent. This represents a transformation to the consulting experience that embeds intelligent agents throughout the consulting life cycle to deliver better value for customers.  ( 119 min )
    Amazon Bedrock AgentCore and Claude: Transforming business with agentic AI
    In this post, we explore how Amazon Bedrock AgentCore and Claude are enabling enterprises like Cox Automotive and Druva to deploy production-ready agentic AI systems that deliver measurable business value, with results including up to 63% autonomous issue resolution and 58% faster response times. We examine the technical foundation combining Claude's frontier AI capabilities with AgentCore's enterprise-grade infrastructure that allows organizations to focus on agent logic rather than building complex operational systems from scratch.  ( 124 min )
  • Open

    SRE Weekly Issue #497
    View on sreweekly.com A message from our sponsor, Costory: You didn’t sign up to do FinOps.Costory automatically explains why your cloud costs change, and reports it straight to Slack.Built for SREs who want to code, not wrestle with spreadsheets.Now on AWS & GCP Marketplaces. Start your free trial at costory.io To vibe or not to […]  ( 4 min )

  • Open

    Build a biomedical research agent with Biomni tools and Amazon Bedrock AgentCore Gateway
    In this post, we demonstrate how to build a production-ready biomedical research agent by integrating Biomni's specialized tools with Amazon Bedrock AgentCore Gateway, enabling researchers to access over 30 biomedical databases through a secure, scalable infrastructure. The implementation showcases how to transform research prototypes into enterprise-grade systems with persistent memory, semantic tool discovery, and comprehensive observability for scientific reproducibility .  ( 127 min )
    Make your web apps hands-free with Amazon Nova Sonic
    Graphical user interfaces have carried the torch for decades, but today’s users increasingly expect to talk to their applications. In this post we show how we added a true voice-first experience to a reference application—the Smart Todo App—turning routine task management into a fluid, hands-free conversation.  ( 120 min )
    Harnessing the power of generative AI: Druva’s multi-agent copilot for streamlined data protection
    Generative AI is transforming the way businesses interact with their customers and revolutionizing conversational interfaces for complex IT operations. Druva, a leading provider of data security solutions, is at the forefront of this transformation. In collaboration with Amazon Web Services (AWS), Druva is developing a cutting-edge generative AI-powered multi-agent copilot that aims to redefine the customer experience in data security and cyber resilience.  ( 123 min )

  • Open

    凤鸣山
    凤鸣山问丹 上虞的秋日,凤鸣山薄雾如纱。我沿石阶缓步而上,两旁古木参天,溪水泠泠作响,仿佛仍回荡着千年前炉火噼啪的余音。山腰处,一方“炼丹井”静卧于苔痕斑驳的石栏内,井水幽深,映不出魏伯阳的身影,却照见自己模糊的轮廓——这方寸水土,曾是他与弟子们埋首炼丹、叩问长生的地方。 山间传说里,魏伯阳炼成金丹后,以犬试药,犬倒地如死。两位弟子心生疑惧,弃丹而去;唯有一人笃信师道,愿同服共死。结果师徒与犬皆复苏飞升,而犹豫者终老山下。故事如山风拂过耳际,初听是仙迹,细思却如井水般冰凉:那“坚定”的信念,究竟该交付给谁?若丹本为毒,那“信”岂非成了引向深渊的绳索? 下山途中,偶遇一位老农在田埂上歇息。他笑谈:“如今谁还信什么金丹?但人总得信点什么吧,不然日子怎么过?”他信的是节气、是土地、是春种秋收的踏实。这朴素之“尚”,与魏伯阳的玄奥丹道相隔千年,却同样支撑着一种生活。吴晗先生曾言,社会风气之“尚”如潮汐涨落,或尚名节,或尚功利,或尚清谈——可无论何种“尚”,若只知随波逐流,不加省察,便如那两位弃丹弟子,既失了飞升之机,也未必真得了安稳。 归途车窗外,城市灯火渐次亮起,霓虹如新的“丹炉”,闪烁着财富、流量、速度的诱人光泽。我们这一代人,何尝不在各自的时代“炼丹”?只是炉中所炼,早已不是铅汞,而是对意义、价值与归属的渴求。然而,若只知追逐那最耀眼的光,却不肯停下脚步自问一句“此丹可服否?”,恐怕终将陷入另一种“假死”——身体活着,灵魂却已沉睡。 凤鸣山无凤鸣,唯有风过林梢。魏伯阳的丹炉早已冷透,但那口炼丹井却像一面古镜,照见所有时代人心深处的叩问:我们该信什么?又为何而信? 苏格拉底说:“未经省察的人生不值得过。”或许真正的“丹”,并非服食之物,而是那敢于审视自身信念的勇气——它不许诺飞升,却能让双脚在尘世站得更稳,让眼睛在迷雾中看得更清。  ( 1 min )

  • Open

    Introducing agent-to-agent protocol support in Amazon Bedrock AgentCore Runtime
    In this post, we demonstrate how you can use the A2A protocol for AI agents built with different frameworks to collaborate seamlessly. You'll learn how to deploy A2A servers on AgentCore Runtime, configure agent discovery and authentication, and build a real-world multi-agent system for incident response. We'll cover the complete A2A request lifecycle, from agent card discovery to task delegation, showing how standardized protocols eliminate the complexity of multi-agent coordination.  ( 126 min )
    Powering enterprise search with the Cohere Embed 4 multimodal embeddings model in Amazon Bedrock
    The Cohere Embed 4 multimodal embeddings model is now available as a fully managed, serverless option in Amazon Bedrock. In this post, we dive into the benefits and unique capabilities of Embed 4 for enterprise search use cases. We’ll show you how to quickly get started using Embed 4 on Amazon Bedrock, taking advantage of integrations with Strands Agents, S3 Vectors, and Amazon Bedrock AgentCore to build powerful agentic retrieval-augmented generation (RAG) workflows.  ( 123 min )
    A guide to building AI agents in GxP environments
    The regulatory landscape for GxP compliance is evolving to address the unique characteristics of AI. Traditional Computer System Validation (CSV) approaches, often with uniform validation strategies, are being supplemented by Computer Software Assurance (CSA) frameworks that emphasize flexible risk-based validation methods tailored to each system's actual impact and complexity (FDA latest guidance). In this post, we cover a risk-based implementation, practical implementation considerations across different risk levels, the AWS shared responsibility model for compliance, and concrete examples of risk mitigation strategies.  ( 125 min )
    Multi-Agent collaboration patterns with Strands Agents and Amazon Nova
    In this post, we explore four key collaboration patterns for multi-agent, multimodal AI systems – Agents as Tools, Swarms Agents, Agent Graphs, and Agent Workflows – and discuss when and how to apply each using the open-source AWS Strands Agents SDK with Amazon Nova models.  ( 131 min )
  • Open

    Checking that Docker image manifests are complete
    Background We build a bunch of stuff for RISC-V using the Dart official Docker image, but the RISC-V images can often arrive some time (days) after the more mainstream images[1]. That means that if we merge a Dependabot PR for an updated image it might well be missing RISC-V, causing the Continuous Delivery (CD) pipeline […]  ( 12 min )
    Checking that Docker image manifests are complete
    Background We build a bunch of stuff for RISC-V using the Dart official Docker image, but the RISC-V images can often arrive some time (days) after the more mainstream images[1]. That means that if we merge a Dependabot PR for an updated image it might well be missing RISC-V, causing the Continuous Delivery (CD) pipeline […]  ( 12 min )
2025-12-10T10:21:39.954Z osmosfeed 1.15.1