Feedback

#2
by esha1111 - opened

The modular setup using Triton Inference Server, FastAPI, and Kubernetes is clean and scalable.
I’d recommend annotating auto-scaling and monitoring layers for full production-readiness. Great work!

Sign up or log in to comment