Hey Devs! ๐
Iโm currently working on an exciting hands-on project that blends the power of LLMOps and DevOps: an AI-powered Anime Recommender System.
Though the project is still a work in progress, I wanted to share what Iโm exploring:
- The architecture and toolchain I'm using ๐งฑ
- The technologies Iโm learning along the way ๐งฐ
- And why this is such an exciting learning opportunity ๐ก
I believe in learning in public, so hereโs a peek behind the scenes!
๐ฏ Project Goal
The goal is to build a recommender system that
- Uses Groq LLM APIs and HuggingFace embeddings
- Stores vector data in ChromaDB
- Communicates with LLMs using LangChain
- Has an interactive UI built in Streamlit
- Is deployed on GCP VM via Docker + Kubernetes
- And monitored with Grafana Cloud
๐ ๏ธ Tech Stack
Category | Tools/Technologies |
---|---|
LLM & Embeddings | Groq API, HuggingFace |
GenAI Framework | LangChain |
Vector Store | ChromaDB |
Frontend/UI | Streamlit |
Containerization | Docker |
Orchestration | Kubernetes (via Minikube) |
Deployment Platform | GCP VM |
Monitoring | Grafana Cloud |
CLI | Kubectl |
Version Control & CI/CD | GitHub |
๐ Architecture (Work in Progress)
_Hereโs the architectural flow Iโm working
_
Project Setup
- Groq API, HuggingFace API
- Logging, Custom Exception Handling
- Virtual Environment, Project Structure
Core Engine
- Configuration and Data Loading
- Embedding storage using ChromaDB
- Prompt templating
- Recommender Class logic
- Training & Recommendation process
UI and Deployment
- Streamlit App as frontend
- Dockerizing the app
- Deploying on Minikube within a GCP VM
- Monitoring Kubernetes metrics with Grafana Cloud
๐ What Iโm Learning
โ
Integrating Groq and HuggingFace for LLM-based workflows
โ
Creating prompt pipelines with LangChain
โ
Managing embeddings using ChromaDB
โ
Deploying containerized apps using Docker on Kubernetes
โ
Running K8s inside GCP VM with Minikube
โ
Setting up Grafana Cloud dashboards for cluster observability
โ
CI/CD and GitHub-based DevOps flows for MLOps/LLMOps
๐ Whatโs Next
- Finalizing the embedding + recommender pipeline
- Adding GitHub Actions CI/CD
- Creating APIs for interaction
- Hosting a public demo
- Publishing full documentation and source code
๐ก Why Share This Now?
This isnโt a finished project. Iโm posting while I explore and build โ because
- Feedback makes me better
- Others might benefit from this stack
- It keeps me accountable and motivated
๐ฌ Stay Tuned for Part 2
In the next post, Iโll share
- Source code and repo walkthrough
- API endpoints for recommendations
- LLM prompt tuning strategies
- Live monitoring via Grafana Cloud
Thanks for reading! ๐
If youโre working on something similar or want to explore LLMOps, feel free to drop your thoughts and ideas in the comments!
Letโs connect ๐
Top comments (0)