AWS Machine Learning Blog

Snowflake Arctic models are now available in HAQM SageMaker JumpStart

Today, we are excited to announce that the Snowflake Arctic Instruct model is available through HAQM SageMaker JumpStart to deploy and run inference. In this post, we walk through how to discover and deploy the Snowflake Arctic Instruct model using SageMaker JumpStart, and provide example use cases with specific prompts.

Enhance call center efficiency using batch inference for transcript summarization with HAQM Bedrock

Today, we are excited to announce general availability of batch inference for HAQM Bedrock. This new feature enables organizations to process large volumes of data when interacting with foundation models (FMs), addressing a critical need in various industries, including call center operations. In this post, we demonstrate the capabilities of batch inference using call center transcript summarization as an example.

Fine-tune Meta Llama 3.1 models for generative AI inference using HAQM SageMaker JumpStart

Fine-tuning Meta Llama 3.1 models with HAQM SageMaker JumpStart enables developers to customize these publicly available foundation models (FMs). The Meta Llama 3.1 collection represents a significant advancement in the field of generative artificial intelligence (AI), offering a range of capabilities to create innovative applications. The Meta Llama 3.1 models come in various sizes, with 8 billion, 70 billion, and 405 billion parameters, catering to diverse project needs. In this post, we demonstrate how to fine-tune Meta Llama 3-1 pre-trained text generation models using SageMaker JumpStart.

Reference architecture for summarizing customer reviews using HAQM Bedrock

Analyze customer reviews using HAQM Bedrock

This post explores an innovative application of large language models (LLMs) to automate the process of customer review analysis. LLMs are a type of foundation model (FM) that have been pre-trained on vast amounts of text data. This post discusses how LLMs can be accessed through HAQM Bedrock to build a generative AI solution that automatically summarizes key information, recognizes the customer sentiment, and generates actionable insights from customer reviews. This method shows significant promise in saving human analysts time while producing high-quality results. We examine the approach in detail, provide examples, highlight key benefits and limitations, and discuss future opportunities for more advanced product review summarization through generative AI.

solution-architecture-accuracy

Accuracy evaluation framework for HAQM Q Business

Generative artificial intelligence (AI), particularly Retrieval Augmented Generation (RAG) solutions, are rapidly demonstrating their vast potential to revolutionize enterprise operations. RAG models combine the strengths of information retrieval systems with advanced natural language generation, enabling more contextually accurate and informative outputs. From automating customer interactions to optimizing backend operation processes, these technologies are not just […]

Solution architecture

Elevate healthcare interaction and documentation with HAQM Bedrock and HAQM Transcribe using Live Meeting Assistant

Today, physicians spend about 49% of their workday documenting clinical visits, which impacts physician productivity and patient care. Did you know that for every eight hours that office-based physicians have scheduled with patients, they spend more than five hours in the EHR? As a consequence, healthcare practitioners exhibit a pronounced inclination towards conversational intelligence solutions, […]

Unlock the power of data governance and no-code machine learning with HAQM SageMaker Canvas and HAQM DataZone

HAQM DataZone is a data management service that makes it quick and convenient to catalog, discover, share, and govern data stored in AWS, on-premises, and third-party sources. HAQM DataZone allows you to create and manage data zones, which are virtual data lakes that store and process your data, without the need for extensive coding or […]

Accelerate performance using a custom chunking mechanism with HAQM Bedrock

This post explores how Accenture used the customization capabilities of Knowledge Bases for HAQM Bedrock to incorporate their data processing workflow and custom logic to create a custom chunking mechanism that enhances the performance of Retrieval Augmented Generation (RAG) and unlock the potential of your PDF data.