AWS Machine Learning Blog

Solution Architecture

Automate customer support with HAQM Bedrock, LangGraph, and Mistral models

In this post, we demonstrate how to use HAQM Bedrock and LangGraph to build a personalized customer support experience for an ecommerce retailer. By integrating the Mistral Large 2 and Pixtral Large models, we guide you through automating key customer support workflows such as ticket categorization, order details extraction, damage assessment, and generating contextual responses.

Mental model for choosing HAQM Bedrock options for cost optimization

Effective cost optimization strategies for HAQM Bedrock

With the increasing adoption of HAQM Bedrock, optimizing costs is a must to help keep the expenses associated with deploying and running generative AI applications manageable and aligned with your organization’s budget. In this post, you’ll learn about strategic cost optimization techniques while using HAQM Bedrock.

How E.ON saves £10 million annually with AI diagnostics for smart meters powered by HAQM Textract

E.ON’s story highlights how a creative application of HAQM Textract, combined with custom image analysis and pulse counting, can solve a real-world challenge at scale. By diagnosing smart meter errors through brief smartphone videos, E.ON aims to lower costs, improve customer satisfaction, and enhance overall energy service reliability. In this post, we dive into how this solution works and the impact it’s making.

Building intelligent AI voice agents with Pipecat and HAQM Bedrock

Building intelligent AI voice agents with Pipecat and HAQM Bedrock – Part 1

In this series of posts, you will learn how to build intelligent AI voice agents using Pipecat, an open-source framework for voice and multimodal conversational AI agents, with foundation models on HAQM Bedrock. It includes high-level reference architectures, best practices and code samples to guide your implementation.

Stream multi-channel audio to HAQM Transcribe using the Web Audio API

In this post, we explore the implementation details of a web application that uses the browser’s Web Audio API and HAQM Transcribe streaming to enable real-time dual-channel transcription. By using the combination of AudioContext, ChannelMergerNode, and AudioWorklet, we were able to seamlessly process and encode the audio data from two microphones before sending it to HAQM Transcribe for transcription.

How Kepler democratized AI access and enhanced client services with HAQM Q Business

At Kepler, a global full-service digital marketing agency serving Fortune 500 brands, we understand the delicate balance between creative marketing strategies and data-driven precision. In this post, we share how implementing HAQM Q Business transformed our operations by democratizing AI access across our organization while maintaining stringent security standards, resulting in an average savings of 2.7 hours per week per employee in manual work and improved client service delivery.

AWS Step Functions state machine for audio processing: Whisper transcription, speaker identification, and Bedrock summary tasks

Build a serverless audio summarization solution with HAQM Bedrock and Whisper

In this post, we demonstrate how to use the Open AI Whisper foundation model (FM) Whisper Large V3 Turbo, available in HAQM Bedrock Marketplace, which offers access to over 140 models through a dedicated offering, to produce near real-time transcription. These transcriptions are then processed by HAQM Bedrock for summarization and redaction of sensitive information.

Solution workflow

Implement semantic video search using open source large vision models on HAQM SageMaker and HAQM OpenSearch Serverless

In this post, we demonstrate how to use large vision models (LVMs) for semantic video search using natural language and image queries. We introduce some use case-specific methods, such as temporal frame smoothing and clustering, to enhance the video search performance. Furthermore, we demonstrate the end-to-end functionality of this approach by using both asynchronous and real-time hosting options on HAQM SageMaker AI to perform video, image, and text processing using publicly available LVMs on the Hugging Face Model Hub. Finally, we use HAQM OpenSearch Serverless with its vector engine for low-latency semantic video search.

Multi-account support for HAQM SageMaker HyperPod task governance

In this post, we discuss how an enterprise with multiple accounts can access a shared HAQM SageMaker HyperPod cluster for running their heterogenous workloads. We use SageMaker HyperPod task governance to enable this feature.