Posted 3mo ago

Senior/Staff Applied Scientist, Multimodal Representation Learning (Oncology)

@ Pathos AI
New York City, New York, United States
$150k-$200k/yrHybridFull Time
Responsibilities:Develop models, Evaluate benchmarks, Collaborate with partners
Requirements Summary:PhD strongly preferred in ML/AI, CS, statistics, computational biology, or related field; deep learning with PyTorch; foundation-model/representation learning; ability to work in ambiguous problems.
Technical Tools Mentioned:PyTorch
Save
Mark Applied
Hide Job
Report & Hide
Job Description

About the role:

Where Frontier AI Meets Frontier Biology to Deliver Frontier Medicine

We are hiring specialized scientists to accelerate development of our Oncology Foundation Model (OFM) stack. This is not a generic “model tinkering” role. The person in this seat will help define and build the modeling strategy that turns multimodal oncology data (clinical text/EHR, genomics, transcriptomics, pathology imaging, and derived features) into useful representations and predictive capabilities that directly support drug discovery and development.

You’ll operate at the intersection of:

  • Frontier AI (representation learning, multimodal learning, alignment, evaluation)
  • Messy biomedical reality (clinical endpoints, censoring, confounding, missingness, batch effects)
  • Mechanism + translation (models that can be interrogated, stress-tested, and connected to biology and outcomes)

This role complements (not duplicates) the computational biology roles that focus on our program-facing biomarker analyses and trial decisions.  

What You Will Do

Foundation model development

  • Design and implement multimodal pretraining and fine-tuning strategies for oncology data (e.g., contrastive objectives, masked modeling, multitask learning, retrieval-augmented training, late/early fusion variants).
  • Build model components that improve cross-modality grounding (e.g., aligning clinical narratives with molecular state and pathology signals).
  • Develop robust approaches for missing-modality settings (train-time and inference-time), ensuring the OFM remains useful when only subsets of modalities exist.

Clinical + molecular fluency 

  • Work with domain partners to define prediction targets and representation tests that matter: response, durability, toxicity, survival, progression, resistance, subtype stability, etc.
  • Incorporate oncology-specific realities into modeling and evaluation (censoring, treatment lines, temporal leakage, cohort shift, annotation noise).

Evaluation, benchmarking, and scientific rigor

  • Create evaluation harnesses that go beyond leaderboard metrics: ablations, cohort-shift tests, missingness stress tests, temporal generalization, calibration, and failure-mode analysis.
  • Define and maintain benchmark suites that reflect Pathos priorities and are reproducible across model iterations.
  • Partner with engineering to support scalable training/inference (multi-node GPU training, data pipelines, throughput optimization), while keeping scientific intent front-and-center 

Translation enablement 

  • Package model outputs so they can be consumed by internal science teams: embeddings, uncertainty estimates, interpretable signals, retrieval tools, and model cards that clearly state what’s reliable vs. not.
  • Collaborate with computational biologists, translational scientists, and clinicians to ensure the OFM supports mechanism discovery and patient stratification workflows 

Who You Are

Minimum Qualifications

  • Advanced degree (PhD strongly preferred) in ML/AI, CS, Statistics, Computational Biology, Bioinformatics, or a related field, or equivalent industry experience with a strong publication/impact record.
  • Deep hands-on experience with modern deep learning (PyTorch), including training large models and debugging optimization issues.
  • Demonstrated ability to design representation learning / foundation model approaches and evaluate them rigorously (not just “train and report AUCs”).
  • Comfort operating in ambiguous problem spaces with a bias toward execution and iteration.

Strongly Preferred 

  • Multimodal foundation model experience (any of: clinical + omics, imaging + text, multimodal retrieval, alignment, late fusion/mixture-of-experts).
  • Real experience with at least one of the following domains (enough to reason about the data-generating process and pitfalls):
    • Clinical text / EHR (notes, longitudinal events, coding systems, leakage traps)
    • Molecular/omics modeling (RNA/DNA/variant features, batch effects, multi-cohort generalization)
    • Pathology imaging (WSI feature learning, weak supervision, MIL, slide-level endpoints)

Nice to Have

  • Distributed training and systems experience (FSDP/DeepSpeed, multi-node performance profiling)
  • Experience with alignment methods (preference learning, instruction tuning, evaluation frameworks for reliability/robustness).
  • Publications in relevant venues (NeurIPS/ICML/ICLR/ACL/MLHC) and/or impactful open-source work.

Location

This is a hybrid role, requiring up to 3 days per week onsite, in our NYC Headquarters.