AWS Machine Learning Blog
Category: HAQM SageMaker
Driving advanced analytics outcomes at scale using HAQM SageMaker powered PwC’s Machine Learning Ops Accelerator
This post was written in collaboration with Ankur Goyal and Karthikeyan Chokappa from PwC Australia’s Cloud & Digital business. Artificial intelligence (AI) and machine learning (ML) are becoming an integral part of systems and processes, enabling decisions in real time, thereby driving top and bottom-line improvements across organizations. However, putting an ML model into production […]
Accelerating time-to-insight with MongoDB time series collections and HAQM SageMaker Canvas
This is a guest post co-written with Babu Srinivasan from MongoDB. As industries evolve in today’s fast-paced business landscape, the inability to have real-time forecasts poses significant challenges for industries heavily reliant on accurate and timely insights. The absence of real-time forecasts in various industries presents pressing business challenges that can significantly impact decision-making and […]
Use HAQM DocumentDB to build no-code machine learning solutions in HAQM SageMaker Canvas
We are excited to announce the launch of HAQM DocumentDB (with MongoDB compatibility) integration with HAQM SageMaker Canvas, allowing HAQM DocumentDB customers to build and use generative AI and machine learning (ML) solutions without writing code. HAQM DocumentDB is a fully managed native JSON document database that makes it straightforward and cost-effective to operate critical […]
Boost productivity on HAQM SageMaker Studio: Introducing JupyterLab Spaces and generative AI tools
HAQM SageMaker Studio offers a broad set of fully managed integrated development environments (IDEs) for machine learning (ML) development, including JupyterLab, Code Editor based on Code-OSS (Visual Studio Code Open Source), and RStudio. It provides access to the most comprehensive set of tools for each step of ML development, from preparing data to building, training, […]
How AWS Prototyping enabled ICL-Group to build computer vision models on HAQM SageMaker
This is a customer post jointly authored by ICL and AWS employees. ICL is a multi-national manufacturing and mining corporation based in Israel that manufactures products based on unique minerals and fulfills humanity’s essential needs, primarily in three markets: agriculture, food, and engineered materials. Their mining sites use industrial equipment that has to be monitored […]
Automate PDF pre-labeling for HAQM Comprehend
HAQM Comprehend is a natural-language processing (NLP) service that provides pre-trained and custom APIs to derive insights from textual data. HAQM Comprehend customers can train custom named entity recognition (NER) models to extract entities of interest, such as location, person name, and date, that are unique to their business. To train a custom model, you […]
Improve your Stable Diffusion prompts with Retrieval Augmented Generation
Text-to-image generation is a rapidly growing field of artificial intelligence with applications in a variety of areas, such as media and entertainment, gaming, ecommerce product visualization, advertising and marketing, architectural design and visualization, artistic creations, and medical imaging. Stable Diffusion is a text-to-image model that empowers you to create high-quality images within seconds. In November […]
Streamlining ETL data processing at Talent.com with HAQM SageMaker
This post outlines the ETL pipeline we developed for feature processing for training and deploying a job recommender model at Talent.com. Our pipeline uses SageMaker Processing jobs for efficient data processing and feature extraction at a large scale. Feature extraction code is implemented in Python enabling the use of popular ML libraries to perform feature extraction at scale, without the need to port the code to use PySpark.
Fine-tune Llama 2 using QLoRA and Deploy it on HAQM SageMaker with AWS Inferentia2
In this post, we showcase fine-tuning a Llama 2 model using a Parameter-Efficient Fine-Tuning (PEFT) method and deploy the fine-tuned model on AWS Inferentia2. We use the AWS Neuron software development kit (SDK) to access the AWS Inferentia2 device and benefit from its high performance. We then use a large model inference container powered by […]
Build an end-to-end MLOps pipeline using HAQM SageMaker Pipelines, GitHub, and GitHub Actions
Machine learning (ML) models do not operate in isolation. To deliver value, they must integrate into existing production systems and infrastructure, which necessitates considering the entire ML lifecycle during design and development. ML operations, known as MLOps, focus on streamlining, automating, and monitoring ML models throughout their lifecycle. Building a robust MLOps pipeline demands cross-functional […]