Cognition Logo

Cognition

Research Engineer, Infrastructure

Posted 3 Days Ago
In-Office or Remote
Hiring Remotely in CA
Mid level
In-Office or Remote
Hiring Remotely in CA
Mid level
The Research Engineer, Infrastructure will build distributed training systems, optimize performance, manage data pipelines, and enhance research workflows, ensuring the infrastructure scales with AI advancements.
The summary above was generated by AI
Who We Are

We are an applied AI lab building end-to-end software agents. We're the team behind Devin, the first AI software engineer, and Windsurf, an AI-native IDE. These products represent our vision for AI that doesn't just assist engineers, but works alongside them as a genuine teammate.

Our team is small and talent-dense: world-class competitive programmers, former founders, and researchers from the frontier of AI, including Scale AI, Palantir, Cursor, Google DeepMind, and others.

Role Mission

Research moves at the speed of the infrastructure underneath it. Every training run, evaluation loop, and experimental iteration depends on systems that are fast, reliable, and built to scale. This role exists to make sure nothing in the stack becomes the bottleneck that slows down the frontier.

You will own the core systems that researchers depend on daily: distributed training infrastructure, experiment orchestration, data pipelines, and the tooling that turns raw compute into usable research velocity. This is not a support role. You will work directly alongside researchers, understand the science deeply enough to anticipate what they need next, and build systems that hold up under the pressure of training jobs running across thousands of GPUs. We don't distinguish between research and engineering; the best infrastructure engineers here are also the ones who understand why the research works.

What You'll Accomplish
  • Distributed Training Infrastructure: Build and own the systems that run large-scale training jobs reliably across GPU clusters. This includes job launchers, checkpointing and recovery, fault tolerance, and the monitoring that keeps researchers informed and unblocked.

  • Scaling Agent Rollouts: Own the infrastructure that runs hundreds of thousands of concurrent coding agent rollouts in VM sandboxes, from high-fidelity environment design to the distributed systems that hold up at our largest RL training scales.

  • Performance Optimization: Profile and improve training throughput end to end. Identify bottlenecks across data loading, communication overhead, memory utilization, and compute efficiency. Implement solutions that meaningfully improve step time and MFU at scale.

  • Experiment Orchestration and Tooling: Design and maintain the systems researchers use to launch, track, and analyze experiments. Reduce friction in the research loop so that more time is spent on ideas and less on waiting.

  • Data Pipeline Engineering: Build high-throughput, reliable data pipelines for training and evaluation. Ensure data quality, reproducibility, and efficiency at the scale our training runs demand.

  • Debugging and Reliability: Diagnose and resolve training failures across GPUs, networking, numerics, and data. Maintain detailed understanding of failure modes and build systems that fail gracefully and recover fast.

  • Parallelism and Systems Research: Implement and optimize parallelism strategies: data, tensor, pipeline, and sequence parallelism. Understand the tradeoffs deeply and apply them to get the most out of available hardware.

  • Scaling Infrastructure Ahead of Research: Anticipate what the research team will need next and build it before it becomes a constraint. The best infrastructure engineers here are proactive, not reactive.

Exceptional Candidates Have Demonstrated
  • Deep experience building and operating distributed training systems for large models; comfortable owning infrastructure end to end from the cluster level down to the training loop

  • Strong systems engineering fundamentals: distributed systems, networking, storage, and the ability to reason about performance across the full hardware-software stack

  • Proficiency in Python and C++; experience with PyTorch or equivalent deep learning frameworks at a systems level, not just API usage

  • Hands-on experience with GPU performance profiling, memory optimization, and compute efficiency; able to diagnose why a training run is underperforming and fix it

  • Experience implementing or optimizing parallelism strategies (data, tensor, pipeline, sequence) for large model training

  • Track record of building tooling and abstractions that meaningfully accelerate research workflows

  • Strong debugging instincts across complex, distributed systems where failures are non-deterministic and hard to reproduce

  • Enough ML knowledge to engage substantively with researchers: understand what they are training, why the architecture choices matter, and what the infrastructure needs to support

  • We care more about demonstrated capability than credentials. A PhD is one signal among many.

Resources & Environment
  • Small, highly selective team where research and product move together; prototypes reach real deployment quickly

  • You'll own and operate infrastructure running across thousands of GPUs; compute is not a constraint and neither is access to the systems you need to do the work well

  • The environment rewards speed, autonomy, and technical depth with minimal process overhead; this is one of the most competitive and fast-moving problems in AI

Equal Opportunity

Cognition is an equal opportunity employer. We do not discriminate on the basis of race, color, religion, sex, sexual orientation, gender identity, national origin, age, disability, veteran status, or any other protected characteristic under applicable law. We are committed to providing reasonable accommodations for candidates with disabilities throughout the hiring process - please let us know if you need any.

Top Skills

C++
Gpu
Python
PyTorch

Similar Jobs

15 Days Ago
Remote or Hybrid
CA
Mid level
Mid level
Artificial Intelligence • Information Technology • Software
The role involves designing scalable data pipelines for 3D, video, and sensor data, optimizing infrastructure, and productionizing ML models with researchers.
Top Skills: SparkAWSAzureDaskDvcFlyteGCPKubernetesMlflowPythonPyTorchRay
16 Hours Ago
In-Office or Remote
CA
Senior level
Senior level
Blockchain • eCommerce • Fintech • Payments • Software • Financial Services • Cryptocurrency
The Senior ML/AI Modeler will automate risk investigations using AI, improve productivity through technology, and deploy ML systems at scale, while collaborating across teams to enhance performance and decision-making.
Top Skills: AirflowAWSGCPLlmsPythonPyTorchSnowflake
16 Hours Ago
Remote or Hybrid
CA
Senior level
Senior level
Blockchain • Fintech • Mobile • Payments • Software • Financial Services
The role involves developing and deploying AI systems to automate risk and compliance processes, and driving strategic initiatives across teams to enhance productivity and decision-making.
Top Skills: AirflowAWSGCPLlmsMlPythonPyTorchSnowflake

What you need to know about the Vancouver Tech Scene

Raincouver, Vancity, The Big Smoke — Vancouver is known by many names, and in recent years, it has gained a reputation as a growing hub for both tech and sustainability. Renowned for its natural beauty, the city has become a magnet for professionals eager to create environmental solutions, and with an emphasis on clean technology, renewable energy and environmental innovation, it's attracted companies across various industries, all working toward a shared goal: advancing clean technology.

Sign up now Access later

Create Free Account

Please log in or sign up to report this job.

Create Free Account