AWS Machine Learning Blog
Category: HAQM SageMaker JumpStart
Build a secure enterprise application with Generative AI and RAG using HAQM SageMaker JumpStart
In this post, we build a secure enterprise application using AWS Amplify that invokes an HAQM SageMaker JumpStart foundation model, HAQM SageMaker endpoints, and HAQM OpenSearch Service to explain how to create text-to-text or text-to-image and Retrieval Augmented Generation (RAG). You can use this post as a reference to build secure enterprise applications in the Generative AI domain using AWS services.
Fine-tune Llama 2 for text generation on HAQM SageMaker JumpStart
Today, we are excited to announce the capability to fine-tune Llama 2 models by Meta using HAQM SageMaker JumpStart. The Llama 2 family of large language models (LLMs) is a collection of pre-trained and fine-tuned generative text models ranging in scale from 7 billion to 70 billion parameters. Fine-tuned LLMs, called Llama-2-chat, are optimized for dialogue use cases.
Build a generative AI-based content moderation solution on HAQM SageMaker JumpStart
In this post, we introduce a novel method to perform content moderation on image data with multi-modal pre-training and a large language model (LLM). With multi-modal pre-training, we can directly query the image content based on a set of questions of interest and the model will be able to answer these questions. This enables users to chat with the image to confirm if it contains any inappropriate content that violates the organization’s policies. We use the powerful generating capability of LLMs to generate the final decision including safe/unsafe labels and category type. In addition, by designing a prompt, we can make an LLM generate the defined output format, such as JSON format. The designed prompt template allows the LLM to determine if the image violates the moderation policy, identify the category of violation, explain why, and provide the output in a structured JSON format.
Optimize deployment cost of HAQM SageMaker JumpStart foundation models with HAQM SageMaker asynchronous endpoints
In this post, we target these situations and solve the problem of risking high costs by deploying large foundation models to HAQM SageMaker asynchronous endpoints from HAQM SageMaker JumpStart. This can help cut costs of the architecture, allowing the endpoint to run only when requests are in the queue and for a short time-to-live, while scaling down to zero when no requests are waiting to be serviced. This sounds great for a lot of use cases; however, an endpoint that has scaled down to zero will introduce a cold start time before being able to serve inferences.
Automatically generate impressions from findings in radiology reports using generative AI on AWS
This post demonstrates a strategy for fine-tuning publicly available LLMs for the task of radiology report summarization using AWS services. LLMs have demonstrated remarkable capabilities in natural language understanding and generation, serving as foundation models that can be adapted to various domains and tasks. There are significant benefits to using a pre-trained model. It reduces computation costs, reduces carbon footprints, and allows you to use state-of-the-art models without having to train one from scratch.
Intelligent video and audio Q&A with multilingual support using LLMs on HAQM SageMaker
Digital assets are vital visual representations of products, services, culture, and brand identity for businesses in an increasingly digital world. Digital assets, together with recorded user behavior, can facilitate customer engagement by offering interactive and personalized experiences, allowing companies to connect with their target audience on a deeper level. Efficiently discovering and searching for specific […]
Zero-shot and few-shot prompting for the BloomZ 176B foundation model with the simplified HAQM SageMaker JumpStart SDK
HAQM SageMaker JumpStart is a machine learning (ML) hub offering algorithms, models, and ML solutions. With SageMaker JumpStart, ML practitioners can choose from a growing list of best performing and publicly available foundation models (FMs) such as BLOOM, Llama 2, Falcon-40B, Stable Diffusion, OpenLLaMA, Flan-T5/UL2, or FMs from Cohere and LightOn. In this post and […]
Build production-ready generative AI applications for enterprise search using Haystack pipelines and HAQM SageMaker JumpStart with LLMs
In this post, we showcase how to build an end-to-end generative AI application for enterprise search with Retrieval Augmented Generation (RAG) by using Haystack pipelines and the Falcon-40b-instruct model from HAQM SageMaker JumpStart and HAQM OpenSearch Service.
Zero-shot text classification with HAQM SageMaker JumpStart
Natural language processing (NLP) is the field in machine learning (ML) concerned with giving computers the ability to understand text and spoken words in the same way as human beings can. Recently, state-of-the-art architectures like the transformer architecture are used to achieve near-human performance on NLP downstream tasks like text summarization, text classification, entity recognition, […]
Unlocking creativity: How generative AI and HAQM SageMaker help businesses produce ad creatives for marketing campaigns with AWS
Advertising agencies can use generative AI and text-to-image foundation models to create innovative ad creatives and content. In this post, we demonstrate how you can generate new images from existing base images using HAQM SageMaker, a fully managed service to build, train, and deploy ML models for at scale. With this solution, businesses large and […]