Amazon’s AI stack, centered around Amazon Web Services (AWS), is vertically integrated. It spans from raw infrastructure all the way to end-user AI applications. The key idea: AWS doesn’t just give you models; it gives you every layer needed to build, train, deploy, and scale AI systems in production.
Let’s break it down cleanly, layer by layer.
Layer 1: Data Layer
Purpose: Enables data collection, storage, processing, and preparation for ML
- Amazon S3: Scalable object storage for datasets and model artifacts
- Amazon Redshift: Data warehousing for analytics and ML training
- AWS Glue: Serverless data integration and ETL (Extract, Transform, Load) service
- Amazon Kinesis: Real-time data streaming for ML applications
- AWS Lake Formation: Build, secure, and manage data lakes for ML
Layer 2: Machine Learning Platform Layer (Bedrock)
Purpose: Hosts Amazon’s in-house Foundation models as well as APIs to access partners’ models
· Amazon Foundation Models: Amazon has Nova family of Foundation models to cater to a variety of needs
o Nova Pro: Designed for complex, multi-step tasks, offering top-tier performance for reasoning and understanding
o Nova Lite: An efficient, fast, and cost-effective model suitable for text, image, and video tasks
o Nova Micro: An extremely fast and lightweight model focused on high-throughput, low-latency text tasks
o Nova 2 Omni (Preview): A multimodal reasoning model capable of processing text, images, video, and speech, while natively generating text and images
o Nova Reel: Dedicated to generating short video content, including the ability to take reference images to guide video creation
o Nova Canvas: Generates images and offers editing capabilities
o Nova Act: Foundation model that interacts with UIs: clicks buttons, fills forms, navigates apps - ideal for legacy system automation
- AWS IoT Greengrass: Run ML models locally on edge devices
- AWS Panorama: Computer vision at the edge
- AWS Outposts: Run AWS infrastructure on-premises for low-latency ML
Layer 3a: Tools for Builders’ Layer
Purpose: Tools for building custom ML models and workflows
- SageMaker Studio: Integrated development environment (IDE) for ML
- SageMaker Autopilot: Automated model building
- SageMaker JumpStart: Pre-built models and solutions
- SageMaker Pipelines: Orchestrate ML workflows
- SageMaker Feature Store: Central repository for ML features
- SageMaker Clarify: Detect bias and explain model predictions
- SageMaker Edge Manager: Deploy models to edge devices
- AWS Deep Learning AMIs: Pre-configured environments for deep learning frameworks (TensorFlow, PyTorch, etc.)
- AWS Deep Learning Containers: Docker images for deep learning
Layer 3b: AI Application Layer
Purpose: Provide pre-build plug-n-play AI APIs for ML workloads
- Amazon Rekognition: Image and video analysis (e.g., object detection, facial recognition)
- Amazon Polly: Text-to-speech service
- Amazon Lex: Build conversational interfaces (chatbots, voice assistants)
- Amazon Comprehend: Natural language processing (NLP) for text analysis
- Amazon Forecast: Time-series forecasting
- Amazon Personalize: Real-time personalized recommendations
- Amazon Textract: Extract text and data from documents
- Amazon Transcribe: Automatic speech recognition
- Amazon Translate: Language translation
- Amazon Forecast: Time-series forecasting for demand planning, inventory optimization, etc.
- Amazon Fraud Detector: Fraud prevention for transaction risk scoring, account takeover detection
- Q Developer: Assistant for developers, acts as a pair programmer, helping write, debug, and upgrade code (including complex tasks like migrating legacy Java code). Amongst its capabilities include coding suggestions and security scanning.
- Q Business: Ingests data from 40+ systems like Amazon S3 and Salesforce to help build “Q Apps” for sales, lawyers etc. to perform Q&A to help users get answers to their questions, provide summaries, generate content, and securely complete tasks based on data and information in their enterprise systems.
Users can also use Amazon Q Apps to generate apps in a single step from their conversation with or by describing their requirements.
- Q in QuickSight: Natural language BI for asking questions about data
Which Layer Should You Use?
|
Requirement |
Layer |
Key Service |
|
I want to build a custom LLM from scratch |
Tools for Builders’ Layer |
SageMaker |
|
I want to build an AI agent for my app |
Machine Learning Platform Layer |
Bedrock |
|
I need an AI to help my employees work faster |
AI Applications Layer |
Amazon Q |
