Cupertino, CA, US
127 days ago
SoC Device Driver Engineer, Machine Learning Accelerators
Custom silicon chips live at the heart of AWS Machine Learning servers, and our team builds the backend software to run these servers. We’re looking for someone to lead our system-on-chip (SoC) driver software team and help us deliver at scale, as we build the next generation of driver software.

As the lead for the SoC driver team, you will:
- Build and manage a small, strong team of developers
- Work with hardware designers to write drivers for newly developed hardware modules
- Refactor and maintain existing codebases throughout the device lifecycle
- Continuously test and deploy your software stack to multiple internal customers
- Innovate on the tooling you provide to customers, making it easier for them to use and debug our SoCs

Annapurna Labs, our organization within AWS, designs and deploys some of the largest custom silicon in the world, with many subsystems that must all be managed, tested, and monitored. The SoC drivers are a critical piece of the AWS management software stack that ensure the chip is functional, performant, and secure.

You will thrive in this role if you:
- Enjoy building and managing small teams
- Are familiar with modular driver architectures (such as the Linux or Windows driver stacks)
- Are proficient in C++ and familiar with Python
- Know how to build effective abstractions over low-level SoC details
- Have strong opinions about software architecture, and are able to apply them effectively
- Enjoy learning new technologies, building software at scale, moving fast, and working closely with colleagues as part of a small team within a large organization

Although we build and deploy machine learning chips, no machine learning background is needed for this role. Your team (and your software) won’t be doing machine learning. Our driver stack lives at the lowest level of the backend AWS infrastructure responsible for managing our ML servers. You and your team will develop drivers for components used by machine learning (example: PCIe, HBM, etc.), but won’t need to deeply understand ML yourselves.

This role can be based in either Cupertino, CA or Austin, TX. The team is split between the two sites, with no preference for one over the other.

This is a fast-paced role where you'll work with thought-leaders in multiple technology areas. You'll have high standards for yourself and everyone you work with, and you'll be constantly looking for ways to improve your software, as well as our products' overall performance, quality, and cost.

We're changing an industry. We're searching for individuals who are ready for this challenge, who want to reach beyond what is possible today. Come join us and build the future of machine learning!

We are open to hiring candidates to work out of one of the following locations:

Austin, TX, USA | Cupertino, CA, USA
Confirm your E-mail: Send Email