Weights & Biases

Principal Product Manager, W&B Inference - Weights & Biases

$206k - $303k
Weights & Biases Livingston, NJ / New York; NY / San Francisco, CA / Sunnyvale; CA / Bellevue, WA 7 days ago
product
CoreWeave, the AI Hyperscaler™, acquired Weights & Biases to create the most powerful end-to-end platform to develop, deploy, and iterate AI faster. Since 2017, CoreWeave has operated a growing footprint of data centers covering every region of the US and across Europe, and was ranked as one of the TIME100 most influential companies of 2024. By bringing together CoreWeave’s industry-leading cloud infrastructure with the best-in-class tools AI practitioners know and love from Weights & Biases, we’re setting a new standard for how AI is built, trained, and scaled.

The integration of our teams and technologies is accelerating our shared mission: to empower developers with the tools and infrastructure they need to push the boundaries of what AI can do. From experiment tracking and model optimization to high-performance training clusters, agent building, and inference at scale, we’re combining forces to serve the full AI lifecycle — all in one seamless platform.

Weights & Biases has long been trusted by over 1,500 organizations — including AstraZeneca, Canva, Cohere, OpenAI, Meta, Snowflake, Square,Toyota, and Wayve — to build better models, AI agents and applications. Now, as part of CoreWeave, that impact is amplified across a broader ecosystem of AI innovators, researchers, and enterprises.

As we unite under one vision, we’re looking for bold thinkers and agile builders who are excited to shape the future of AI alongside us. If you're passionate about solving complex problems at the intersection of software, hardware, and AI, there's never been a more exciting time to join our team.

What You’ll Do:

As a Principal Product Manager for the Weights & Biases Inference Service, you will be responsible for defining, driving, and scaling one of the most critical components of our platform. You’ll own the vision, roadmap, and execution for the inference service end-to-end, ensuring it evolves to meet the performance, reliability, and usability needs of our customers. In this role, you will lead mission-critical initiatives that strengthen and extend the core infrastructure powering W&B’s machine learning workflows.

You’ll be responsible for driving the evolution of the W&B Inference Service through projects that demand deep coordination, technical clarity, and strong execution across multiple disciplines. Whether it’s improving the underlying systems that power inference performance, scaling internal tooling to accelerate developer productivity, or delivering backend changes that strengthen reliability and compliance, you will ensure these initiatives move from concept to launch with focus, urgency, and impact.

If you’re energized by shaping complex distributed systems, working closely with highly technical teams, creating clarity in ambiguous spaces, and delivering platform capabilities that unlock velocity for others, this role is for you.

About the role: 

  • Own the execution and evolution of the W&B Inference Service, delivering solutions that directly support the product vision and long-term platform strategy.
  • Lead cross-team initiatives end-to-end, coordinating engineering, product, security, operations, and go-to-market stakeholders to ensure aligned priorities and seamless delivery across interdependent systems.
  • Prioritize with intention, making informed trade-offs among performance, reliability, compliance, cost, and development velocity to ensure the inference service scales to meet customer and platform demands.
  • Elevate developer and practitioner experiences by improving the operability, observability, and usability of the inference service and the tooling that surrounds it.
  • Own execution from requirements through launch, defining success metrics, gathering customer and system insights, and ensuring every stage of development is anchored in measurable outcomes.

Who You Are: 

  • Experience: A seasoned product manager with 7+ years working on high-scale platform or infrastructure products, with direct experience in model serving, inference systems, real-time APIs, or distributed compute services.

  • Inference systems versatility: You’ve worked across domains that commonly intersect with inference—including autoscaling, observability, GPU/accelerator utilization, routing/orchestration, developer tooling, IAM, and storage—and can reason about how changes ripple through a real-time serving stack.

  • Deep technical fluency: You can engage engineers on service architectures, performance bottlenecks, deployment topologies, model packaging formats, request/response patterns, and reliability trade-offs that impact low-latency inference. You’re comfortable interpreting architecture diagrams and discussing how design decisions influence throughput, cost, and SLAs.

  • Cross-functional leadership: Adept at coordinating teams across inference runtime, infrastructure, security, operations, and go-to-market, ensuring alignment on priorities that improve the performance, reliability, and usability of the inference service.

  • Customer empathy for practitioners: You understand the workflows of ML practitioners running production models and the needs of internal developers building on top of the inference platform. You’re motivated by uncovering friction in their serving pipelines and translating those insights into meaningful improvements.

  • Execution mindset: You excel in ambiguous, fast-moving environments. You bring clarity to competing priorities, make thoughtful trade-offs among latency, reliability, cost, and velocity, and consistently drive inference-focused initiatives from concept to launch.

Preferred: 

  • Direct experience as a PM for an inference or model-serving service, ideally involving real-time, low-latency, or high-throughput workloads. Experience with frameworks like TensorFlow, PyTorch, or model-serialization formats is a plus.

  • Background in adjacent platform domains such as identity & access management, billing and metering workflows, observability, or data infrastructure—especially where they intersect with running models in production.

  • Strong familiarity with cloud infrastructure (AWS, GCP, Azure), container orchestration, autoscaling, and deployment automation tools used to operate distributed inference systems.

  • Exposure to W&B or similar MLOps tools, especially experiment tracking, model management, or deployment workflows.

Wondering if you’re a good fit? We believe in investing in our people, and value candidates who can bring their own diversified experiences to our teams – even if you aren't a 100% skill or experience match. Here are a few qualities we’ve found compatible with our team. If some of this describes you, we’d love to talk. 

  • You love to build frictionless products for developers 
  • You’re curious about AI and MLOps tooling
  • You’re an expert in building inference systems that scale for production workloads

Why Us?

We work hard, have fun, and move fast! We’re in an exciting stage of hyper-growth that you will not want to miss out on. We’re not afraid of a little chaos, and we’re constantly learning. Our team cares deeply about how we build our product and how we work together, which is represented through our core values:

  • Be Curious at Your Core
  • Act Like an Owner
  • Empower Employees
  • Deliver Best-in-Class Client Experiences
  • Achieve More Together

We support and encourage an entrepreneurial outlook and independent thinking. We foster an environment that encourages collaboration and provides the opportunity to develop innovative solutions to complex problems. As we get set for takeoff, the growth opportunities within the organization are constantly expanding. You will be surrounded by some of the best talent in the industry, who will want to learn from you, too. Come join us!

The base salary range for this role is $206,000 to $303,000. The starting salary will be determined based on job-related knowledge, skills, experience, and market location. We strive for both market alignment and internal equity when determining compensation. In addition to base salary, our total rewards package includes a discretionary bonus, equity awards, and a comprehensive benefits program (all based on eligibility). 

What We Offer

The range we’ve posted represents the typical compensation range for this role. To determine actual compensation, we review the market rate for each candidate which can include a variety of factors. These include qualifications, experience, interview performance, and location.

In addition to a competitive salary, we offer a variety of benefits to support your needs, including:

  • Medical, dental, and vision insurance - 100% paid for by CoreWeave
  • Company-paid Life Insurance 
  • Voluntary supplemental life insurance 
  • Short and long-term disability insurance 
  • Flexible Spending Account
  • Health Savings Account
  • Tuition Reimbursement 
  • Ability to Participate in Employee Stock Purchase Program (ESPP)
  • Mental Wellness Benefits through Spring Health 
  • Family-Forming support provided by Carrot
  • Paid Parental Leave 
  • Flexible, full-service childcare support with Kinside
  • 401(k) with a generous employer match
  • Flexible PTO
  • Catered lunch each day in our office and data center locations
  • A casual work environment
  • A work culture focused on innovative disruption

Our Workplace

While we prioritize a hybrid work environment, remote work may be considered for candidates located more than 30 miles from an office, based on role requirements for specialized skill sets. New hires will be invited to attend onboarding at one of our hubs within their first month. Teams also gather quarterly to support collaboration

California Consumer Privacy Act - California applicants only

CoreWeave is an equal opportunity employer, committed to fostering an inclusive and supportive workplace. All qualified applicants and candidates will receive consideration for employment without regard to race, color, religion, sex, disability, age, sexual orientation, gender identity, national origin, veteran status, or genetic information.

As part of this commitment and consistent with the Americans with Disabilities Act (ADA), CoreWeave will ensure that qualified applicants and candidates with disabilities are provided reasonable accommodations for the hiring process, unless such accommodation would cause an undue hardship. If reasonable accommodation is needed, please contact: careers@coreweave.com

 Export Control Compliance

This position requires access to export controlled information.  To conform to U.S. Government export regulations applicable to that information, applicant must either be (A) a U.S. person, defined as a (i) U.S. citizen or national, (ii) U.S. lawful permanent resident (green card holder), (iii) refugee under 8 U.S.C. § 1157, or (iv) asylee under 8 U.S.C. § 1158, (B) eligible to access the export controlled information without a required export authorization, or (C) eligible and reasonably likely to obtain the required export authorization from the applicable U.S. government agency.  CoreWeave may, for legitimate business reasons, decline to pursue any export licensing process.

Sponsored

Explore Product

Skills in this job

People also search for