AI Serving Platform
Redis Queue ingestion, GPU worker execution, Kubernetes scheduling, and result upload workflows.
Seoul, South Korea · AI/ML Infrastructure
I design and operate Kubernetes-based AI/ML inference pipelines, GPU model serving systems, Redis queue workflows, CI/CD automation, and observability stacks for production medical AI workloads.
Redis Queue ingestion, GPU worker execution, Kubernetes scheduling, and result upload workflows.
KEDA-based scale-to-zero model workers to optimize GPU utilization and reduce idle cloud spend.
Prometheus, Grafana, cAdvisor, Node Exporter, Jenkins, ArgoCD, Helm, Docker, and Linux operations.
Experience
JLK · Publicly traded medical AI company specializing in stroke analysis
Selected Projects
Older school projects were removed so the portfolio focuses on current professional impact.
Built an infrastructure-agnostic model serving flow from request ingestion to Redis Queue, GPU worker execution, result upload, and monitoring for medical AI workloads.
Migrated 250 million DICOM objects totaling 60TB from Ncloud Object Storage to AWS S3. Used Terraform, AWS DataSync, boto3 orchestration, and greedy sharding to balance parallel workers.
Large CUDA-based images caused 5–10 minute cold starts. I automated cluster-wide image pre-pulling through a Jenkins-triggered DaemonSet to make model rollouts significantly faster.
Technical Skills
Python, Java
Kubernetes, Docker, KEDA, Helm, ArgoCD, Jenkins, AWS DataSync, EC2, S3, CloudFront
Linux, Redis, Nginx, GPU inference operations, CUDA-based containers
FastAPI, Celery, Gunicorn, Uvicorn
Prometheus, Grafana, Node Exporter, cAdvisor
GitLab, GitHub
Education & Certification
Bachelor of Science in Computer Science · Graduated Cum Laude
The Linux Foundation
Contact
I’m open to backend, infrastructure, platform engineering, and AI/ML infrastructure roles where production reliability and system design matter.