AI-powered answer engine with real-time source verification
Perplexity operates a search and answer platform built on a heavy machine-learning stack: PyTorch, Triton, CUDA, TensorFlow, ONNX for model inference, plus Next.js and React on the frontend. The company is actively adopting RAG (retrieval-augmented generation), Kafka, and infrastructure-as-code (Terraform, dbt, Snowflake, Databricks), while simultaneously scaling inference bottlenecks and building agentic products—a set of challenges that explains their senior-heavy engineering hiring and focus on LLM deployment and CI/CD pipeline maturity.
Notable leadership hires: Site Lead
Perplexity is an answer engine that combines real-time web search with AI-powered reasoning to deliver sourced responses. Founded in 2022 and based in San Francisco, the company operates a 201–500-person organization with a backend and infrastructure focus: 53 engineers, 8 data specialists, and 7 security staff actively hiring. The platform relies on Python, Kubernetes, and a complex machine-learning inference stack, with active development on core search components, RAG pipelines for grounding answers, LLM-as-a-judge systems for quality control, and a companion browser product. The company also offers APIs for AI inference to external customers, signaling a move toward a platform business model alongside the direct consumer product.
Python, Next.js, TypeScript, PyTorch, Triton, CUDA, TensorFlow, Kubernetes, PostgreSQL, Redis, AWS, DynamoDB, Snowflake, and Databricks. Currently adopting Kafka, Terraform, dbt, and RAG frameworks.
Core search infrastructure, RAG pipelines for answer grounding, LLM-as-a-judge systems for quality control, a companion browser, APIs for external AI inference, and scalable legal infrastructure to support platform growth.
Other companies in the same industry, closest in size