Posted 5mo ago

Machine Learning Researcher / Engineer (Foundational Models)

@ Pathway
Wrocław, Lower Silesian Voivodeship, Poland
$100k-$1000k/yrHybridFull Time
Responsibilities:train models, design experiments, improve architectures
Requirements Summary:Deep learning researcher with LM/RL focus, hands-on ML training (PyTorch/Jax/TensorFlow), GPU architecture knowledge, memory design, CI/CD, Git, English fluency, and collaboration in hybrid/work-from-office settings.
Technical Tools Mentioned:PyTorch, Jax, TensorFlow, Git, CI/CD, Build systems
Save
Mark Applied
Hide Job
Report & Hide
Job Description

About Pathway

At Pathway we are shaking the foundations of artificial intelligence by introducing the world’s first post-transformer model that adapts and thinks just like humans. 

Our breakthrough architecture outperforms Transformer and provides the enterprise with full visibility into how the model works. Combining the foundational model with the fastest data processing engine on the market, Pathway enables enterprises to move beyond incremental optimization and toward truly contextualized, experience-driven intelligence. We are trusted by organizations such as NATO, La Poste, and Formula 1 racing teams.

Pathway is led by co-founder & CEO Zuzanna Stamirowska, a complexity scientist who created a team consisting of AI pioneers, including CTO Jan Chorowski who was the first person to apply Attention to speech and worked with Nobel laureate Geoff Hinton at Google Brain, as well as CSO Adrian Kosowski, a leading computer scientist and quantum physicist who obtained his PhD at the age of 20. 

The company is backed by leading investors and advisors, including Lukasz Kaiser, co-author of the Transformer (“the T” in ChatGPT) and a key researcher behind OpenAI’s reasoning models. Pathway is headquartered in Palo Alto, California.


The Opportunity

This is an R&D position in attention-based models.

We are currently searching for 1 or 2 R&D Engineers with a strong track record in machine learning models research.

This is an extremely ambitious foundational project. There is a flexible GPU budget associated with this specific project, guaranteed to be in the 7-digit range minimum.

You Will

  • perform (distributed) model training.
  • help improve/adapt model architectures based on experiment results.
  • design new tasks and experiments.
  • optionally: oversee activities of team members involved in data preparation.

The results of your work will play a crucial role in the success of the project.