AWS Machine Learning Blog
Tag: AI/ML
Build an AI-powered document processing platform with open source NER model and LLM on HAQM SageMaker
In this post, we discuss how you can build an AI-powered document processing platform with open source NER and LLMs on SageMaker.
Optimizing Mixtral 8x7B on HAQM SageMaker with AWS Inferentia2
This post demonstrates how to deploy and serve the Mixtral 8x7B language model on AWS Inferentia2 instances for cost-effective, high-performance inference. We’ll walk through model compilation using Hugging Face Optimum Neuron, which provides a set of tools enabling straightforward model loading, training, and inference, and the Text Generation Inference (TGI) Container, which has the toolkit for deploying and serving LLMs with Hugging Face.
Build multi-agent systems with LangGraph and HAQM Bedrock
This post demonstrates how to integrate open-source multi-agent framework, LangGraph, with HAQM Bedrock. It explains how to use LangGraph and HAQM Bedrock to build powerful, interactive multi-agent applications that use graph-based orchestration.
Building an AIOps chatbot with HAQM Q Business custom plugins
In this post, we demonstrate how you can use custom plugins for HAQM Q Business to build a chatbot that can interact with multiple APIs using natural language prompts. We showcase how to build an AIOps chatbot that enables users to interact with their AWS infrastructure through natural language queries and commands. The chatbot is capable of handling tasks such as querying the data about HAQM Elastic Compute Cloud (HAQM EC2) ports and HAQM Simple Storage Service (HAQM S3) buckets access settings.
Automating regulatory compliance: A multi-agent solution using HAQM Bedrock and CrewAI
In this post, we explore how AI agents can streamline compliance and fulfill regulatory requirements for financial institutions using HAQM Bedrock and CrewAI. We demonstrate how to build a multi-agent system that can automatically summarize new regulations, assess their impact on operations, and provide prescriptive technical guidance. You’ll learn how to use HAQM Bedrock Knowledge Bases and HAQM Bedrock Agents with CrewAI to create a comprehensive, automated compliance solution.
Build an enterprise synthetic data strategy using HAQM Bedrock
In this post, we explore how to use HAQM Bedrock for synthetic data generation, considering these challenges alongside the potential benefits to develop effective strategies for various applications across multiple industries, including AI and machine learning (ML).
Generate training data and cost-effectively train categorical models with HAQM Bedrock
In this post, we explore how you can use HAQM Bedrock to generate high-quality categorical ground truth data, which is crucial for training machine learning (ML) models in a cost-sensitive environment. Generative AI solutions can play an invaluable role during the model development phase by simplifying training and test data creation for multiclass classification supervised learning use cases. We dive deep into this process on how to use XML tags to structure the prompt and guide HAQM Bedrock in generating a balanced label dataset with high accuracy. We also showcase a real-world example for predicting the root cause category for support cases. This use case, solvable through ML, can enable support teams to better understand customer needs and optimize response strategies.
Benchmarking customized models on HAQM Bedrock using LLMPerf and LiteLLM
This post begins a blog series exploring DeepSeek and open FMs on HAQM Bedrock Custom Model Import. It covers the process of performance benchmarking of custom models in HAQM Bedrock using popular open source tools: LLMPerf and LiteLLM. It includes a notebook that includes step-by-step instructions to deploy a DeepSeek-R1-Distill-Llama-8B model, but the same steps apply for any other model supported by HAQM Bedrock Custom Model Import.
Evaluate RAG responses with HAQM Bedrock, LlamaIndex and RAGAS
In this post, we’ll explore how to leverage HAQM Bedrock, LlamaIndex, and RAGAS to enhance your RAG implementations. You’ll learn practical techniques to evaluate and optimize your AI systems, enabling more accurate, context-aware responses that align with your organization’s specific needs.
Accelerate AWS Well-Architected reviews with Generative AI
In this post, we explore a generative AI solution leveraging HAQM Bedrock to streamline the WAFR process. We demonstrate how to harness the power of LLMs to build an intelligent, scalable system that analyzes architecture documents and generates insightful recommendations based on AWS Well-Architected best practices. This solution automates portions of the WAFR report creation, helping solutions architects improve the efficiency and thoroughness of architectural assessments while supporting their decision-making process.