echoloc

Sciforium Tech Stack

Byte-native AI models and serverless LLM serving infrastructure

Technology, Information and Internet San Francisco 11–50 employees Founded 2024 Privately Held

Sciforium builds foundation models and LLM serving infrastructure, with a deep systems focus: the stack spans PyTorch, JAX, CUDA, ROCm, vLLM, and distributed training frameworks (Ray, Kubernetes). The research-heavy org (5 engineers, 1 researcher, 1 data scientist) is tackling inference latency, distributed training optimization, and GPU kernel development — pain points that reflect a build-from-metal approach rather than API-wrapping. Backing from AMD signals hardware partnership potential.

Tech Stack 32 technologies

Core StackPyTorch Python C++ Kubernetes TypeScript Stripe FastAPI Linux JAX TPU Flax XLA CUDA Ray ROCm NVIDIA Nsight vLLM AMD WebSockets Server-Sent Events Ubuntu CentOS Red Hat Enterprise Linux cuDNN NCCL NFS GPFS Lustre LDAP FreeIPA+2 more
AdoptingStripe

What Sciforium Is Building

Challenges

  • Scaling ai infrastructure
  • Scaling full stack for frontier ai models
  • Real-time inference performance
  • Optimizing distributed training performance
  • Debugging hardware-software interactions
  • Scaling large-scale ai training workloads
  • Performance optimization at scale
  • Optimizing inference latency
  • Building large etl pipelines for internet-scale data
  • Maintaining gpu compute cluster health

Active Projects

  • Post-training & alignment data
  • Large language models research
  • Generative media research
  • Scalable training systems research
  • Model serving platform
  • Distributed inference systems
  • Gpu kernel development
  • Distributed training system optimization
  • Ml software stack maintenance
  • Llm training and deployment

Hiring Activity

Minimal10 roles · 0 in 30d

Department

Engineering
5
Data
1
Research
1
Sales
1

Seniority

Senior
6
Mid
2
Company intelligence

Find more companies like Sciforium by tech stack, pain points and active projects

Get started free

About Sciforium

Sciforium is a San Francisco-based AI infrastructure company founded in 2024, developing byte-native multimodal foundation models and serverless LLM serving platforms. The company operates as a small, research-forward team focused on reducing the cost and complexity of large language model deployment. Core projects span large language model research, generative media, scalable training systems, and distributed inference optimization. They serve teams building or deploying frontier AI models at scale.

HeadquartersSan Francisco
Company Size11–50 employees
Founded2024
Hiring MarketsUnited States

Frequently Asked Questions

What is Sciforium's tech stack?

PyTorch, JAX, CUDA, ROCm, vLLM, Ray, Kubernetes, Flax, XLA, and C++. Infrastructure includes TPU, AMD, and NVIDIA Nsight. Distributed storage via Lustre, GPFS, NFS.

What is Sciforium working on?

Large language model research, generative media, distributed training system optimization, GPU kernel development, model serving platforms, and distributed inference systems. Core focus: scaling infrastructure and optimizing inference latency.

Similar Companies in Technology, Information and Internet

Other companies in the same industry, closest in size