Cast AI automates Kubernetes cluster management across AWS, GCP, and Azure, with a tech stack centered on Go, Python, and container orchestration (Kubernetes, ArgoCD, Helm). The company is actively adopting large language models (LLAMA, Grok) and deploying inference optimization tools (vLLM, SGLang, TensorRT), signaling a pivot toward AI workload automation alongside traditional cloud cost optimization. Engineering dominates hiring (77 of 123 roles), with senior-level concentration (91 senior vs. 13 mid), indicating focus on scaling core platform capabilities rather than sales or customer success.
Cast AI builds an automation platform for Kubernetes and cloud-native infrastructure, targeting engineering teams managing multi-cloud deployments. The product addresses two operational pain points: manual tuning inefficiencies in Kubernetes cluster management and rising infrastructure costs across AWS, GCP, and Azure. Cast AI's active project roadmap emphasizes workload optimization, autoscaling, LLM deployment in Kubernetes, and infrastructure-as-code tooling. The company is headquartered in Miami and operates across 15 countries, with primary engineering centers in Eastern Europe and the United States.
Go, Python, Kubernetes, AWS, GCP, Azure, PostgreSQL, Prometheus, Grafana, GitLab CI/CD, ArgoCD, and Helm. Recent additions include vLLM, SGLang, and TensorRT for LLM optimization.
Workload optimization, Kubernetes autoscaling, LLM deployment and selection automation, authentication and authorization systems, billing/audit trails, and infrastructure-as-code platforms.
Other companies in the same industry, closest in size