AWS Machine Learning Blog
Category: HAQM SageMaker Studio
Customize small language models on AWS with automotive terminology
In this post, we guide you through the phases of customizing SLMs on AWS, with a specific focus on automotive terminology for diagnostics as a Q&A task. We begin with the data analysis phase and progress through the end-to-end process, covering fine-tuning, deployment, and evaluation. We compare a customized SLM with a general purpose LLM, using various metrics to assess vocabulary richness and overall accuracy.
Build a reverse image search engine with HAQM Titan Multimodal Embeddings in HAQM Bedrock and AWS managed services
In this post, you will learn how to extract key objects from image queries using HAQM Rekognition and build a reverse image search engine using HAQM Titan Multimodal Embeddings from HAQM Bedrock in combination with HAQM OpenSearch Serverless Service.
Use HAQM SageMaker Studio with a custom file system in HAQM EFS
In this post, we explore three scenarios demonstrating the versatility of integrating HAQM EFS with SageMaker Studio. These scenarios highlight how HAQM EFS can provide a scalable, secure, and collaborative data storage solution for data science teams.
How Northpower used computer vision with AWS to automate safety inspection risk assessments
In this post, we share how Northpower has worked with their technology partner Sculpt to reduce the effort and carbon required to identify and remediate public safety risks. Specifically, we cover the computer vision and artificial intelligence (AI) techniques used to combine datasets into a list of prioritized tasks for field teams to investigate and mitigate.
Control data access to HAQM S3 from HAQM SageMaker Studio with HAQM S3 Access Grants
In this post, we demonstrate how to simplify data access to HAQM S3 from SageMaker Studio using S3 Access Grants, specifically for different user personas using IAM principals.
Making traffic lights more efficient with HAQM Rekognition
In this blog post, we show you how HAQM Rekognition can mitigate congestion at traffic intersections and reduce operations and maintenance costs.
Accelerate development of ML workflows with HAQM Q Developer in HAQM SageMaker Studio
In this post, we present a real-world use case analyzing the Diabetes 130-US hospitals dataset to develop an ML model that predicts the likelihood of readmission after discharge.
Align Meta Llama 3 to human preferences with DPO, HAQM SageMaker Studio, and HAQM SageMaker Ground Truth
In this post, we show you how to enhance the performance of Meta Llama 3 8B Instruct by fine-tuning it using direct preference optimization (DPO) on data collected with SageMaker Ground Truth.
Use LangChain with PySpark to process documents at massive scale with HAQM SageMaker Studio and HAQM EMR Serverless
In this post, we explore how to build a scalable and efficient Retrieval Augmented Generation (RAG) system using the new EMR Serverless integration, Spark’s distributed processing, and an HAQM OpenSearch Service vector database powered by the LangChain orchestration framework. This solution enables you to process massive volumes of textual data, generate relevant embeddings, and store them in a powerful vector database for seamless retrieval and generation.
LLM experimentation at scale using HAQM SageMaker Pipelines and MLflow
Large language models (LLMs) have achieved remarkable success in various natural language processing (NLP) tasks, but they may not always generalize well to specific domains or tasks. You may need to customize an LLM to adapt to your unique use case, improving its performance on your specific dataset or task. You can customize the model […]