Start Date
Immediate
Expiry Date
11 Dec, 25
Salary
0.0
Posted On
12 Sep, 25
Experience
0 year(s) or above
Remote Job
Yes
Telecommute
Yes
Sponsor Visa
No
Skills
Computer Science, Data Structures, Llvm, Linear Programming, Code Generation, Computer Architecture, Computer Engineering, Optimization, Amazon, Machine Learning, Programming Language Theory
Industry
Information Technology/IT
DESCRIPTION
At AWS, our mission is to make deep learning accessible to every developer by democratizing access to cutting-edge infrastructure. To achieve this, we’ve built custom silicon (AWS Inferentia and Trainium) and the AWS Neuron SDK that together deliver high-performance, cost-effective machine learning in the cloud.
The AWS Neuron SDK includes a compiler, runtime, debugger, and libraries integrated with popular frameworks such as PyTorch and TensorFlow. It is preinstalled in AWS Deep Learning AMIs and Containers so customers can quickly get started with training and inference on AWS ML accelerators.
The Neuron Toronto team focuses on performance, kernels, and tooling—analyzing and optimizing end-to-end ML workloads, developing and maintaining highly optimized kernels, and building performance modeling, profiling, and accuracy debugging tools. Together, these efforts ensure that Neuron delivers best-in-class performance, flexibility, and usability for customers deploying large-scale machine learning models.
As a student intern, you will contribute to the efforts that make Neuron best-in-class for ML workloads. You’ll gain hands-on experience building business-critical features, analyzing performance, developing compiler or kernel optimizations, and building tools that provide deep insights into model execution. You’ll be mentored by experienced engineers while working on technology that directly accelerates customer workloads at scale.
BASIC QUALIFICATIONS
Are enrolled in a Bachelor’s degree or higher in Computer Science, Engineering Science, Computer Engineering, Electrical Engineering, or majors relating to these fields with an anticipated graduation date between May 2027 - May 2028
Strong interests and academic qualifications/research focus in two of the following: 1. Knowledge of code generation, compute graph optimization, resource scheduling 2. Compiler - Optimizing compilers (internals of LLVM, clang, etc) 3. Machine Learning frameworks (PyTorch, JAX) and Machine Learning (Experience with XLA, TVM, MLIR, LLVM) 4. Kernel development—experience writing CUDA kernels, OpenCL kernels, or ML-specific kernels
Available for a 12-16-month internship starting May 2026
PREFERRED QUALIFICATIONS
How To Apply:
Incase you would like to apply to this job directly from the source, please click here
Please refer the Job description for details