San Jose, California30+ days ago
You will help scale execution across initiatives spanning inference software, runtime enablement, model optimization, systems integration, performance, benchmark readiness, deployment workflows, ecosystem readiness, and product enablement deliverables—including engagement with public inference projects and ecosystems (e.g., SGLang, vLLM) where relevant, as well as benchmark platforms (e.g., MLPerf and InferenceX) where we drive submissions and readiness. In this role, you will be a key partner to engineering, product, and business leadership, ensuring that near-term execution strength is matched by clear long-term planning, rigorous prioritization, and proactive management of risks, dependencies, and decision points across a rapidly evolving AI ecosystem.