San Francisco, Remote
Full time
Hybrid
Engineering
At Prime Intellect, we’re enabling the next generation of AI breakthroughs by helping our customers deploy and optimize massive GPU clusters. As our Solutions Architect for GPU Infrastructure, you’ll be the technical expert who transforms customer requirements into production-ready systems capable of training the world’s most advanced AI models.
We recently raised $15mm in funding (total of $20mm raised) led by Founders Fund, with participation from Menlo Ventures and prominent angels including Andrej Karpathy (Eureka AI, Tesla, OpenAI), Tri Dao (Chief Scientific Officer of Together AI), Dylan Patel (SemiAnalysis), Clem Delangue (Huggingface), Emad Mostaque (Stability AI) and many others.
Core Technical Responsibilities
This customer-facing role combines deep technical expertise with hands-on implementation. You’ll be instrumental in:
Customer Architecture & Design
Partner with clients to understand workload requirements and design optimal GPU cluster architectures
Create technical proposals and capacity planning for clusters ranging from 100 to 10,000+ GPUs
Develop deployment strategies for LLM training, inference, and HPC workloads
Present architectural recommendations to technical and executive stakeholders
Infrastructure Deployment & Optimization
Deploy and configure orchestration systems including SLURM and Kubernetes for distributed workloads
Implement high-performance networking with InfiniBand, RoCE, and NVLink interconnects
Optimize GPU utilization, memory management, and inter-node communication
Configure parallel filesystems (Lustre, BeeGFS, GPFS) for optimal I/O performance
Tune system performance from kernel parameters to CUDA configurations
Production Operations & Support
Serve as primary technical escalation point for customer infrastructure issues
Diagnose and resolve complex problems across the full stack – hardware, drivers, networking, and software
Implement monitoring, alerting, and automated remediation systems
Provide 24/7 on-call support for critical customer deployments
Create runbooks and documentation for customer operations teams
Technical Requirements
Required Experience
3+ years hands-on experience with GPU clusters and HPC environments
Deep expertise with SLURM and Kubernetes in production GPU settings
Proven experience with InfiniBand configuration and troubleshooting
Strong understanding of NVIDIA GPU architecture, CUDA ecosystem, and driver stack
Experience with infrastructure automation tools (Ansible, Terraform)
Proficiency in Python, Bash, and systems programming
Track record of customer-facing technical leadership
Infrastructure Skills
NVIDIA driver installation and troubleshooting (CUDA, Fabric Manager, DCGM)
Container runtime configuration for GPUs (Docker, Containerd, Enroot)
Linux kernel tuning and performance optimization
Network topology design for AI workloads
Power and cooling requirements for high-density GPU deployments
Nice to Have
Experience with 1000+ GPU deployments
NVIDIA DGX, HGX, or SuperPOD certification
Distributed training frameworks (PyTorch FSDP, DeepSpeed, Megatron-LM)
ML framework optimization and profiling
Experience with AMD MI300 or Intel Gaudi accelerators
Contributions to open-source HPC/AI infrastructure projects
Growth Opportunity
You’ll work directly with customers pushing the boundaries of AI, from startups training foundation models to enterprises deploying massive inference infrastructure. You’ll collaborate with our world-class engineering team while having direct impact on systems powering the next generation of AI breakthroughs.
We value expertise and customer obsession – if you’re passionate about building reliable, high-performance GPU infrastructure and have a track record of successful large-scale deployments, we want to talk to you.
Apply now and join us in our mission to democratize access to planetary scale computing.
Other similar jobs that might interest you