AWS Machine Learning Blog

Category: HAQM SageMaker Studio

Customize small language models on AWS with automotive terminology

In this post, we guide you through the phases of customizing SLMs on AWS, with a specific focus on automotive terminology for diagnostics as a Q&A task. We begin with the data analysis phase and progress through the end-to-end process, covering fine-tuning, deployment, and evaluation. We compare a customized SLM with a general purpose LLM, using various metrics to assess vocabulary richness and overall accuracy.

Build a reverse image search engine with HAQM Titan Multimodal Embeddings in HAQM Bedrock and AWS managed services

In this post, you will learn how to extract key objects from image queries using HAQM Rekognition and build a reverse image search engine using HAQM Titan Multimodal Embeddings from HAQM Bedrock in combination with HAQM OpenSearch Serverless Service.

Use HAQM SageMaker Studio with a custom file system in HAQM EFS

In this post, we explore three scenarios demonstrating the versatility of integrating HAQM EFS with SageMaker Studio. These scenarios highlight how HAQM EFS can provide a scalable, secure, and collaborative data storage solution for data science teams.

How Northpower used computer vision with AWS to automate safety inspection risk assessments

How Northpower used computer vision with AWS to automate safety inspection risk assessments

In this post, we share how Northpower has worked with their technology partner Sculpt to reduce the effort and carbon required to identify and remediate public safety risks. Specifically, we cover the computer vision and artificial intelligence (AI) techniques used to combine datasets into a list of prioritized tasks for field teams to investigate and mitigate.

Align Meta Llama 3 to human preferences with DPO, HAQM SageMaker Studio, and HAQM SageMaker Ground Truth

Align Meta Llama 3 to human preferences with DPO, HAQM SageMaker Studio, and HAQM SageMaker Ground Truth

In this post, we show you how to enhance the performance of Meta Llama 3 8B Instruct by fine-tuning it using direct preference optimization (DPO) on data collected with SageMaker Ground Truth.

Use LangChain with PySpark to process documents at massive scale with HAQM SageMaker Studio and HAQM EMR Serverless

Use LangChain with PySpark to process documents at massive scale with HAQM SageMaker Studio and HAQM EMR Serverless

In this post, we explore how to build a scalable and efficient Retrieval Augmented Generation (RAG) system using the new EMR Serverless integration, Spark’s distributed processing, and an HAQM OpenSearch Service vector database powered by the LangChain orchestration framework. This solution enables you to process massive volumes of textual data, generate relevant embeddings, and store them in a powerful vector database for seamless retrieval and generation.

LLM evaluation and selection journey

LLM experimentation at scale using HAQM SageMaker Pipelines and MLflow

Large language models (LLMs) have achieved remarkable success in various natural language processing (NLP) tasks, but they may not always generalize well to specific domains or tasks. You may need to customize an LLM to adapt to your unique use case, improving its performance on your specific dataset or task. You can customize the model […]