GPU Sizing for ML Workloads
Learn to calculate VRAM requirements, select the right AWS instance, and optimize costs. Includes real benchmarks and a Python sizing calculator.
Deployment, latency, and scale.
Learn to calculate VRAM requirements, select the right AWS instance, and optimize costs. Includes real benchmarks and a Python sizing calculator.
Set up experiment tracking for ML models with MLflow and LLM observability with Langfuse. Includes hyperparameter sweeps, model registry, and cost tracking.
Build a complete ML pipeline with GitHub Actions: data validation, model training, automated testing, and staged deployment to production.
Deploy ML models to production with optimized inference: torch.compile vs ONNX benchmarks, FastAPI serving patterns, and AWS deployment options.
Monitor production ML models with data drift detection, performance tracking, and automated alerting. Includes working Python implementations.
Secure your ML infrastructure with IAM roles, secrets management, VPC configuration, and input validation. Practical patterns for production systems.
Compare Python's data modeling options for AI/ML applications. Learn when to use dataclasses, TypedDict, or Pydantic for API responses, embeddings metadata, and agent tool contracts.
Compare Mamba's selective state space architecture against LSTM and Transformer for hard drive failure prediction. Learn when SSMs beat attention.
Build a production-ready failure prediction system using real Backblaze data. Compare traditional ML vs deep learning approaches and learn when each shines.