What You’ll Do:
As a Principal Product Manager for the Weights & Biases Inference Service, you will be responsible for defining, driving, and scaling one of the most critical components of our platform. You’ll own the vision, roadmap, and execution for the inference service end-to-end, ensuring it evolves to meet the performance, reliability, and usability needs of our customers. In this role, you will lead mission-critical initiatives that strengthen and extend the core infrastructure powering W&B’s machine learning workflows.
You’ll be responsible for driving the evolution of the W&B Inference Service through projects that demand deep coordination, technical clarity, and strong execution across multiple disciplines. Whether it’s improving the underlying systems that power inference performance, scaling internal tooling to accelerate developer productivity, or delivering backend changes that strengthen reliability and compliance, you will ensure these initiatives move from concept to launch with focus, urgency, and impact.
If you’re energized by shaping complex distributed systems, working closely with highly technical teams, creating clarity in ambiguous spaces, and delivering platform capabilities that unlock velocity for others, this role is for you.
About the role:
- Own the execution and evolution of the W&B Inference Service, delivering solutions that directly support the product vision and long-term platform strategy.
- Lead cross-team initiatives end-to-end, coordinating engineering, product, security, operations, and go-to-market stakeholders to ensure aligned priorities and seamless delivery across interdependent systems.
- Prioritize with intention, making informed trade-offs among performance, reliability, compliance, cost, and development velocity to ensure the inference service scales to meet customer and platform demands.
- Elevate developer and practitioner experiences by improving the operability, observability, and usability of the inference service and the tooling that surrounds it.
- Own execution from requirements through launch, defining success metrics, gathering customer and system insights, and ensuring every stage of development is anchored in measurable outcomes.
Who You Are:
-
Experience: A seasoned product manager with 7+ years working on high-scale platform or infrastructure products, with direct experience in model serving, inference systems, real-time APIs, or distributed compute services.
-
Inference systems versatility: You’ve worked across domains that commonly intersect with inference—including autoscaling, observability, GPU/accelerator utilization, routing/orchestration, developer tooling, IAM, and storage—and can reason about how changes ripple through a real-time serving stack.
-
Deep technical fluency: You can engage engineers on service architectures, performance bottlenecks, deployment topologies, model packaging formats, request/response patterns, and reliability trade-offs that impact low-latency inference. You’re comfortable interpreting architecture diagrams and discussing how design decisions influence throughput, cost, and SLAs.
-
Cross-functional leadership: Adept at coordinating teams across inference runtime, infrastructure, security, operations, and go-to-market, ensuring alignment on priorities that improve the performance, reliability, and usability of the inference service.
-
Customer empathy for practitioners: You understand the workflows of ML practitioners running production models and the needs of internal developers building on top of the inference platform. You’re motivated by uncovering friction in their serving pipelines and translating those insights into meaningful improvements.
- Execution mindset: You excel in ambiguous, fast-moving environments. You bring clarity to competing priorities, make thoughtful trade-offs among latency, reliability, cost, and velocity, and consistently drive inference-focused initiatives from concept to launch.
Preferred:
-
Direct experience as a PM for an inference or model-serving service, ideally involving real-time, low-latency, or high-throughput workloads. Experience with frameworks like TensorFlow, PyTorch, or model-serialization formats is a plus.
-
Background in adjacent platform domains such as identity & access management, billing and metering workflows, observability, or data infrastructure—especially where they intersect with running models in production.
-
Strong familiarity with cloud infrastructure (AWS, GCP, Azure), container orchestration, autoscaling, and deployment automation tools used to operate distributed inference systems.
- Exposure to W&B or similar MLOps tools, especially experiment tracking, model management, or deployment workflows.
Wondering if you’re a good fit? We believe in investing in our people, and value candidates who can bring their own diversified experiences to our teams – even if you aren't a 100% skill or experience match. Here are a few qualities we’ve found compatible with our team. If some of this describes you, we’d love to talk.
- You love to build frictionless products for developers
- You’re curious about AI and MLOps tooling
- You’re an expert in building inference systems that scale for production workloads
Why Us?
We work hard, have fun, and move fast! We’re in an exciting stage of hyper-growth that you will not want to miss out on. We’re not afraid of a little chaos, and we’re constantly learning. Our team cares deeply about how we build our product and how we work together, which is represented through our core values:
- Be Curious at Your Core
- Act Like an Owner
- Empower Employees
- Deliver Best-in-Class Client Experiences
- Achieve More Together
We support and encourage an entrepreneurial outlook and independent thinking. We foster an environment that encourages collaboration and provides the opportunity to develop innovative solutions to complex problems. As we get set for takeoff, the growth opportunities within the organization are constantly expanding. You will be surrounded by some of the best talent in the industry, who will want to learn from you, too. Come join us!
The base salary range for this role is $206,000 to $303,000. The starting salary will be determined based on job-related knowledge, skills, experience, and market location. We strive for both market alignment and internal equity when determining compensation. In addition to base salary, our total rewards package includes a discretionary bonus, equity awards, and a comprehensive benefits program (all based on eligibility).
What We Offer
The range we’ve posted represents the typical compensation range for this role. To determine actual compensation, we review the market rate for each candidate which can include a variety of factors. These include qualifications, experience, interview performance, and location.
In addition to a competitive salary, we offer a variety of benefits to support your needs, including:
- Medical, dental, and vision insurance - 100% paid for by CoreWeave
- Company-paid Life Insurance
- Voluntary supplemental life insurance
- Short and long-term disability insurance
- Flexible Spending Account
- Health Savings Account
- Tuition Reimbursement
- Ability to Participate in Employee Stock Purchase Program (ESPP)
- Mental Wellness Benefits through Spring Health
- Family-Forming support provided by Carrot
- Paid Parental Leave
- Flexible, full-service childcare support with Kinside
- 401(k) with a generous employer match
- Flexible PTO
- Catered lunch each day in our office and data center locations
- A casual work environment
- A work culture focused on innovative disruption
Our Workplace
While we prioritize a hybrid work environment, remote work may be considered for candidates located more than 30 miles from an office, based on role requirements for specialized skill sets. New hires will be invited to attend onboarding at one of our hubs within their first month. Teams also gather quarterly to support collaboration
California Consumer Privacy Act - California applicants only
CoreWeave is an equal opportunity employer, committed to fostering an inclusive and supportive workplace. All qualified applicants and candidates will receive consideration for employment without regard to race, color, religion, sex, disability, age, sexual orientation, gender identity, national origin, veteran status, or genetic information.
As part of this commitment and consistent with the Americans with Disabilities Act (ADA), CoreWeave will ensure that qualified applicants and candidates with disabilities are provided reasonable accommodations for the hiring process, unless such accommodation would cause an undue hardship. If reasonable accommodation is needed, please contact: careers@coreweave.com.
Export Control Compliance
This position requires access to export controlled information. To conform to U.S. Government export regulations applicable to that information, applicant must either be (A) a U.S. person, defined as a (i) U.S. citizen or national, (ii) U.S. lawful permanent resident (green card holder), (iii) refugee under 8 U.S.C. § 1157, or (iv) asylee under 8 U.S.C. § 1158, (B) eligible to access the export controlled information without a required export authorization, or (C) eligible and reasonably likely to obtain the required export authorization from the applicable U.S. government agency. CoreWeave may, for legitimate business reasons, decline to pursue any export licensing process.
Sponsored
Explore Product
Skills in this job
People also search for
Similar Jobs
More jobs at Weights & Biases
Fullstack Engineer, Billing- Weights & Biases
Weights & Biases
Customer Support Engineer - Tier 3 - Weights & Biases
Weights & Biases
Senior Customer Reliability Engineer - Weights & Biases
Weights & Biases
Senior AI Solutions Engineer, Post Sales- Weights & Biases
Weights & Biases
AI Engineer- ML/RL - Weights & Biases
Weights & Biases
Similar Jobs
More jobs at Weights & Biases
Fullstack Engineer, Billing- Weights & Biases
Weights & Biases
Customer Support Engineer - Tier 3 - Weights & Biases
Weights & Biases
Senior Customer Reliability Engineer - Weights & Biases
Weights & Biases
Senior AI Solutions Engineer, Post Sales- Weights & Biases
Weights & Biases
AI Engineer- ML/RL - Weights & Biases
Weights & Biases