article thumbnail

Build and deploy a UI for your generative AI applications with AWS and Python

AWS Machine Learning Blog

AWS provides a powerful set of tools and services that simplify the process of building and deploying generative AI applications, even for those with limited experience in frontend and backend development. The AWS deployment architecture makes sure the Python application is hosted and accessible from the internet to authenticated users.

AWS 120
article thumbnail

WordFinder app: Harnessing generative AI on AWS for aphasia communication

AWS Machine Learning Blog

David Copland, from QARC, and Scott Harding, a person living with aphasia, used AWS services to develop WordFinder, a mobile, cloud-based solution that helps individuals with aphasia increase their independence through the use of AWS generative AI technology. The following diagram illustrates the solution architecture on AWS.

AWS 86
professionals

Sign Up for our Newsletter

This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.

article thumbnail

Apply Amazon SageMaker Studio lifecycle configurations using AWS CDK

AWS Machine Learning Blog

Solution overview The solution constitutes a best-practice Amazon SageMaker domain setup with a configurable list of domain user profiles and a shared SageMaker Studio space using the AWS Cloud Development Kit (AWS CDK). The AWS CDK is a framework for defining cloud infrastructure as code. The AWS CDK installed.

AWS 94
article thumbnail

Improving Retrieval Augmented Generation accuracy with GraphRAG

AWS Machine Learning Blog

Lettria , an AWS Partner, demonstrated that integrating graph-based structures into RAG workflows improves answer precision by up to 35% compared to vector-only retrieval methods. In this post, we explore why GraphRAG is more comprehensive and explainable than vector RAG alone, and how you can use this approach using AWS services and Lettria.

AWS 130
article thumbnail

Enterprise-grade natural language to SQL generation using LLMs: Balancing accuracy, latency, and scale

Flipboard

This post describes a pattern that AWS and Cisco teams have developed and deployed that is viable at scale and addresses a broad set of challenging enterprise use cases. Augmenting data with data definitions for prompt construction Several of the optimizations noted earlier require making some of the specifics of the data domain explicit.

SQL 148
article thumbnail

Fine-tune and host SDXL models cost-effectively with AWS Inferentia2

AWS Machine Learning Blog

We show how to then prepare the fine-tuned model to run on AWS Inferentia2 powered Amazon EC2 Inf2 instances , unlocking superior price performance for your inference workloads. After the model is fine-tuned, you can compile and host the fine-tuned SDXL on Inf2 instances using the AWS Neuron SDK. An Amazon Web Services (AWS) account.

AWS 85
article thumbnail

Get started quickly with AWS Trainium and AWS Inferentia using AWS Neuron DLAMI and AWS Neuron DLC

AWS Machine Learning Blog

Starting with the AWS Neuron 2.18 release , you can now launch Neuron DLAMIs (AWS Deep Learning AMIs) and Neuron DLCs (AWS Deep Learning Containers) with the latest released Neuron packages on the same day as the Neuron SDK release. AWS DLCs provide a set of Docker images that are pre-installed with deep learning frameworks.

AWS 124