Paris
Concentrate provides one OpenAI-compatible API to access, route, and manage models across leading AI providers and open-source models through a single endpoint. We help teams save time, lower token spend with credits back from our bulk purchasing power, improve reliability, and avoid vendor lock-in. Supported by top-tier VCs. This is a remote role. You'll work directly with customers to solve LLM infrastructure and deployment problems, while also building the product and platform capabilities that make those solutions scalable. This is a highly hands-on role for someone who is technical, pragmatic, and excited to operate across customer work, engineering, and product at an early-stage AI API company. What You'll Do • Work closely with customers to understand LLM deployment needs and solve technical problems in production, • Debug issues end to end across application behavior, AI API integrations, infrastructure, and model and provider performance across OpenAI, Anthropic, Gemini, and open source models, • Build product features, internal tools, and platform improvements based on patterns you see in the field, • Improve multi-provider routing, LLM reliability, AI observability, latency, and token cost efficiency across multiple LLM providers, • Help customers reduce AI infrastructure costs, navigate rate limits, and architect for provider failover and redundancy, • Partner closely with founders on customer deployments, product direction, and technical strategy What We're Looking For • Strong technical ability and high ownership, • Strong debugging instincts across backend systems, AI APIs, infrastructure, and customer environments, • Experience working with or around LLM APIs, model routing, or AI spend management is a strong plus, • Comfort working directly with customers and operating in ambiguity, • Startup experience or experience in fast-moving, high-ownership environments, • Clear written and verbal communication skills, • Fluent English required Bonus • Experience with LLM gateways, AI gateway architecture, or enterprise AI infrastructure, • Familiarity with zero data retention, PII redaction, or AI compliance requirements, • Experience with LLM cost optimization, token spend analysis, or provider discount structures, • Experience in forward deployed, solutions, or customer-facing technical roles, • Early startup experience Salary Range: strong cash compensation + equity Recruitment Agency Notice Concentrate AI does not accept unsolicited resumes from recruitment agencies and is not responsible for any fees related to unsolicited submissions. FAQs What is Concentrate AI? Concentrate AI is an OpenAI-compatible API and fully managed multi-provider LLM platform that lets teams access, route, and manage models from leading AI providers and open-source models through one secure endpoint. How is Concentrate different from OpenRouter? Concentrate is similar to OpenRouter, but built for enterprise teams that need stronger security, rich usage analytics, centralized billing, production reliability, model fallbacks, and bulk-volume token pricing through a fully managed platform. It also supports integrations with Claude Code, Codex, and Cursor, and does not charge for SSO or a monthly platform fee. How is Concentrate different from LiteLLM? LiteLLM is a flexible developer tool for routing LLM calls, while Concentrate is a secure, fully managed, enterprise-ready platform with provider routing, model fallbacks, observability, usage analytics, centralized billing, and one bill for all token usage. Does Concentrate charge platform or SSO fees? No. Concentrate does not charge a monthly platform fee or extra for SSO. Teams get multi-provider routing, fallbacks, centralized billing, and enterprise-ready access controls without additional software fees. Why do teams use Concentrate? Teams use Concentrate to reduce token spend through bulk-volume pricing, improve reliability with model fallbacks, avoid vendor lock-in, simplify multi-provider LLM operations, and access leading AI models through one OpenAI-compatible API.