AI inference accelerators and full-stack software for data center deployment
FuriosaAI builds custom silicon and software for AI inference at scale, combining hardware (NPU design), firmware, and Kubernetes-native orchestration. The tech stack is dominated by inference frameworks (TensorFlow, PyTorch, TensorRT-LLM, vLLM, Triton) and cloud infrastructure (Kubernetes, AWS, Azure, GCP), while active projects focus on chip-to-cloud integration—firmware, SDK development, Kubernetes scheduling, and partner hardware onboarding. The engineering-heavy org (30 of 41 hires) and datacenter-scale resource scheduling pain points indicate a company moving past chip design into full-stack systems.
Notable leadership hires: Art Director
FuriosaAI designs data center accelerators and software for AI inference workloads, founded in 2017 by engineers from AMD, Qualcomm, and Samsung. The company builds custom NPUs, accompanying SDKs, and cloud-native deployment tooling, partnering with TSMC, ASUS, SK Hynix, and Samsung on product integration. With over 140 employees across Seoul, Silicon Valley, and Europe, FuriosaAI serves AI teams in enterprise and cloud environments who need efficient, programmable inference hardware. Current hiring is concentrated in engineering roles across the US, South Korea, and Singapore, with a secondary focus on support and sales.
FuriosaAI uses Rust, C++, and Python for core development, alongside TensorFlow, PyTorch, TensorRT-LLM, and vLLM for inference optimization. Deployment runs on Kubernetes with support for AWS, Azure, and GCP cloud platforms.
Yes. FuriosaAI has 30 active engineering roles open across mid-level (26) and senior (13) positions, with recent postings in the US, South Korea, and Singapore. Hiring velocity is currently decelerating.
FuriosaAI is headquartered in Seoul, South Korea, with offices in Silicon Valley and Europe. The company has over 140 employees and 45 total active job openings.
Other companies in the same industry, closest in size