AWS Machine Learning Blog
Category: Generative AI
Enterprise-grade natural language to SQL generation using LLMs: Balancing accuracy, latency, and scale
In this post, the AWS and Cisco teams unveil a new methodical approach that addresses the challenges of enterprise-grade SQL generation. The teams were able to reduce the complexity of the NL2SQL process while delivering higher accuracy and better overall performance.
AWS Field Experience reduced cost and delivered low latency and high performance with HAQM Nova Lite foundation model
The AFX team’s product migration to the Nova Lite model has delivered tangible enterprise value by enhancing sales workflows. By migrating to the HAQM Nova Lite model, the team has not only achieved significant cost savings and reduced latency, but has also empowered sellers with a leading intelligent and reliable solution.
Build an AI-powered document processing platform with open source NER model and LLM on HAQM SageMaker
In this post, we discuss how you can build an AI-powered document processing platform with open source NER and LLMs on SageMaker.
Use HAQM Bedrock Intelligent Prompt Routing for cost and latency benefits
Today, we’re happy to announce the general availability of HAQM Bedrock Intelligent Prompt Routing. In this blog post, we detail various highlights from our internal testing, how you can get started, and point out some caveats and best practices. We encourage you to incorporate HAQM Bedrock Intelligent Prompt Routing into your new and existing generative AI applications.
How Infosys improved accessibility for Event Knowledge using HAQM Nova Pro, HAQM Bedrock and HAQM Elemental Media Services
In this post, we explore how Infosys developed Infosys Event AI to unlock the insights generated from events and conferences. Through its suite of features—including real-time transcription, intelligent summaries, and an interactive chat assistant—Infosys Event AI makes event knowledge accessible and provides an immersive engagement solution for the attendees, during and after the event.
HAQM Bedrock Prompt Optimization Drives LLM Applications Innovation for Yuewen Group
Today, we are excited to announce the availability of Prompt Optimization on HAQM Bedrock. With this capability, you can now optimize your prompts for several use cases with a single API call or a click of a button on the HAQM Bedrock console. In this blog post, we discuss how Prompt Optimization improves the performance of large language models (LLMs) for intelligent text processing task in Yuewen Group.
Add Zoom as a data accessor to your HAQM Q index
This post demonstrates how Zoom users can access their HAQM Q Business enterprise data directly within their Zoom interface, alleviating the need to switch between applications while maintaining enterprise security boundaries. Organizations can now configure Zoom as a data accessor in HAQM Q Business, enabling seamless integration between their HAQM Q index and Zoom AI Companion. This integration allows users to access their enterprise knowledge in a controlled manner directly within the Zoom platform.
Host concurrent LLMs with LoRAX
In this post, we explore how Low-Rank Adaptation (LoRA) can be used to address these challenges effectively. Specifically, we discuss using LoRA serving with LoRA eXchange (LoRAX) and HAQM Elastic Compute Cloud (HAQM EC2) GPU instances, allowing organizations to efficiently manage and serve their growing portfolio of fine-tuned models, optimize costs, and provide seamless performance for their customers.
Build multi-agent systems with LangGraph and HAQM Bedrock
This post demonstrates how to integrate open-source multi-agent framework, LangGraph, with HAQM Bedrock. It explains how to use LangGraph and HAQM Bedrock to build powerful, interactive multi-agent applications that use graph-based orchestration.
Dynamic text-to-SQL for enterprise workloads with HAQM Bedrock Agents
This post demonstrates how enterprises can implement a scalable agentic text-to-SQL solution using HAQM Bedrock Agents, with advanced error-handling tools and automated schema discovery to enhance database query efficiency.