This monorepo contains a collection of modular, end-to-end projects that demonstrate best practices for AI/ML engineering β from experimentation to production β with a focus on:
- MLOps & LLMOps patterns
- Real-world infrastructure integration
- CI/CD pipelines
- Evaluation, monitoring, and observability
These projects are designed to support a forthcoming book on AI/ML Engineering, and mirror the full lifecycle of modern ML and LLM systems.
Each project explores different slices of the AI/ML engineering stack, including:
- β Training & Experimentation: with Ray, Kedro, classic ML, notebooks
- π§ LLM Apps & Agents: Bedrock, CrewAI, LangGraph, RAG workflows
- π Deployment & Inference: via Ray Serve, FastAPI, Bedrock, Lambda
- π Monitoring & Observability: CloudWatch, Evidently AI, MLflow
- π§ͺ Evaluation: model metrics, agent output quality, prompt tests
- π CI/CD Workflows: GitHub Actions, reusable runners, CT/CI patterns
- π§° Infrastructure: Terraform/CDK, Docker, Python envs
projects/
βββ ray-ml-pipeline/ # Ray Datasets, Train, Serve
βββ kedro-ml-pipeline/ # Modular pipeline + optional MLflow
βββ bedrock-rag-agent/ # Full-stack LLM app (client + server)
βββ agent-frameworks/ # CrewAI, LangGraph, Pydantic-AI
βββ classic-ml/ # Traditional models + CI
βββ observability-pipelines/ # Lambda, CloudWatch, metrics
shared/ # Common utilities across projects
evaluation/ # Shared evaluation logic and runners
infra/ # Optional shared Terraform/CDK modules
.github/ # CI/CD workflows