Remove 2019 Remove AWS Remove Natural Language Processing
article thumbnail

Fast and cost-effective LLaMA 2 fine-tuning with AWS Trainium

AWS Machine Learning Blog

In this post, we walk through how to fine-tune Llama 2 on AWS Trainium , a purpose-built accelerator for LLM training, to reduce training times and costs. We review the fine-tuning scripts provided by the AWS Neuron SDK (using NeMo Megatron-LM), the various configurations we used, and the throughput results we saw.

AWS 132
article thumbnail

AWS Inferentia2 builds on AWS Inferentia1 by delivering 4x higher throughput and 10x lower latency

AWS Machine Learning Blog

The size of the machine learning (ML) models––large language models ( LLMs ) and foundation models ( FMs )–– is growing fast year-over-year , and these models need faster and more powerful accelerators, especially for generative AI. With AWS Inferentia1, customers saw up to 2.3x With AWS Inferentia1, customers saw up to 2.3x

AWS 93
professionals

Sign Up for our Newsletter

This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.

article thumbnail

Build a contextual chatbot application using Knowledge Bases for Amazon Bedrock

AWS Machine Learning Blog

Note that you can also use Knowledge Bases for Amazon Bedrock service APIs and the AWS Command Line Interface (AWS CLI) to programmatically create a knowledge base. Create a Lambda function This Lambda function is deployed using an AWS CloudFormation template available in the GitHub repo under the /cfn folder.

AWS 127
article thumbnail

Amazon SageMaker unveils the Cohere Command R fine-tuning model

AWS Machine Learning Blog

AWS announced the availability of the Cohere Command R fine-tuning model on Amazon SageMaker. This latest addition to the SageMaker suite of machine learning (ML) capabilities empowers enterprises to harness the power of large language models (LLMs) and unlock their full potential for a wide range of applications.

AWS 132
article thumbnail

Advanced RAG patterns on Amazon SageMaker

AWS Machine Learning Blog

For more information on Mixtral-8x7B Instruct on AWS, refer to Mixtral-8x7B is now available in Amazon SageMaker JumpStart. Before you get started with the solution, create an AWS account. This identity is called the AWS account root user. The Mixtral-8x7B model is made available under the permissive Apache 2.0

AWS 138
article thumbnail

Talk to your slide deck using multimodal foundation models hosted on Amazon Bedrock and Amazon SageMaker – Part 2

AWS Machine Learning Blog

We stored the embeddings in a vector database and then used the Large Language-and-Vision Assistant (LLaVA 1.5-7b) We used AWS services including Amazon Bedrock , Amazon SageMaker , and Amazon OpenSearch Serverless in this solution. aws s3 cp {s3_img_path}. In this post, we demonstrate a different approach. I need numbers."

AWS 132
article thumbnail

Build a classification pipeline with Amazon Comprehend custom classification (Part I)

AWS Machine Learning Blog

“Data locked away in text, audio, social media, and other unstructured sources can be a competitive advantage for firms that figure out how to use it“ Only 18% of organizations in a 2019 survey by Deloitte reported being able to take advantage of unstructured data. The majority of data, between 80% and 90%, is unstructured data.

AWS 129