Status: ✅ COMPLETED
Date: 2025-11-23
Deploy the core AWS infrastructure required to support the Generative AI application. This includes storage for vector embeddings (if needed) or documents, and the compute layer for the AI logic.
- Use Terraform for Infrastructure as Code (IaC).
- Deploy to
us-west-2(or configured region). - Resources:
- S3 Bucket: For storing documents/assets.
- Lambda Function: For handling chat requests and streaming responses.
- CloudFront: For content delivery and lightweight validation (CloudFront Functions).
- Location:
/infra - State Management: Local state (for now).
infra/provider.tf: AWS Provider configuration.infra/variables.tf: Project variables.infra/main.tf: Resource definitions.infra/lambda-chat.tf: Chat Lambda configuration.infra/cloudfront.tf: CloudFront distribution.infra/ecr.tf: ECR Repository.
- Define Resources: Update
main.tfto define the S3 bucket and basic Lambda execution role. - Initialize: Run
terraform init. - Plan: Run
terraform planto preview changes. - Apply: Run
make deployto provision resources.
-
terraform validatepasses. -
make deploycompletes successfully.
- S3 Access: Use
infrastructure_test.ipynbto verify the created bucket is accessible and writable. - Create Embeddings: Verify that embeddings can be generated and stored (Pending).
- Lambda Invocation: Test invoking the Lambda function directly via AWS SDK.
Created by Warike technologies