In-memory AI accelerator hardware with custom compiler stack
EnCharge AI designs custom silicon for AI inference, built around in-memory computing architecture paired with a next-generation compiler stack (Torch-MLIR, XLA, LLVM). The hiring profile is heavily weighted toward principal and senior hardware engineers (13 principal roles), reflecting the maturity and depth required for post-silicon validation, timing closure on advanced nodes, and chiplet-based system integration. Active adoption of Llama alongside projects spanning SoC floorplanning, custom NPU kernel libraries, and AI serving infrastructure suggests a shift from pure accelerator design toward end-to-end inference deployment.
Notable leadership hires: Physical Design Lead
EnCharge AI builds in-memory computing hardware and software for AI workloads, founded in 2022 by semiconductor and AI systems veterans. The company operates as a hardware-software hybrid, developing custom silicon optimized for inference in power- and space-constrained environments, paired with a proprietary compiler and runtime stack. Projects span chip design (SoC implementation, floorplanning, post-silicon validation), compiler infrastructure (graph compilation, kernel library integration), and inference serving systems. The organization is structured around engineering depth, with hiring concentrated in hardware design, physical design, and silicon validation across US and India offices.
Hardware design: PCIe, RISC-V, SystemVerilog, UVM, Cadence, Synopsys TestMAX, Innovus, IC Compiler II. Software/compiler: Torch-MLIR, XLA, LLVM, MLIR, C++, Python. Infrastructure: Linux (Red Hat, Ubuntu), GitHub Actions, GitLab CI/CD, Jenkins.
High-speed SoC implementation, chiplet-based AI inference platforms, next-generation AI compiler stack, post-silicon validation, and serving infrastructure for video generation models. Current challenges center on timing closure, performance-power tradeoffs, and graph compilation optimization.
Other companies in the same industry, closest in size