- Amazon (Cupertino, CA)
- …and the Trn1 and Inf2 servers that use them. This role is for a software engineer in the Neuron Inference ( ML Apps) team for AWS Neuron. This role is ... Stable Diffusion, Vision Transformers and many more. The Neuron Inference team works side by side with compiler engineers...running on the AWS Trainium and Inferentia silicon. Strong software development using C++/Python and ML knowledge… more
- NVIDIA (Santa Clara, CA)
- …upon which every new AI-powered application is built. We are seeking a Principal Engineer for NVIDIA Inference Microservices (NIMs). The right person for this ... You will apply your deep technical expertise in AI Inference and containerized software to design an...traceability. What you'll be doing: + You are the engineer 's engineer demonstrating good engineering practice and… more
- NVIDIA (Santa Clara, CA)
- As a System Software Engineer (LLM Inference & Performance Optimization) you will be at the heart of our AI advancements. Our team is dedicated to pushing ... on GPU. + Experience with low-latency optimizations and real-time system constraints for ML inference . The base salary range is 180,000 USD - 339,250 USD.… more
- NVIDIA (Santa Clara, CA)
- …alike developing best-in-class AI models. We are now looking for a Senior Deep Learning Software Engineer to develop and scale up our automated inference and ... on the inference performance to ensure NVIDIA's inference software solutions (TRT, TRT-LLM, TRT Model...design. + Strong proficiency in Python, PyTorch, and related ML tools (eg HuggingFace). + Strong algorithms and programming… more
- NVIDIA (Santa Clara, CA)
- We are now looking for a Senior Software Engineer for Deep Learning Inference ! Would you like to make a big impact in Deep Learning by helping build a ... apply to Senior Engineering positions in the Deep Learning Inference TensorRT software team. What you'll be...working with TensorRT, PyTorch, TensorFlow, ONNX Runtime or other ML frameworks. NVIDIA is widely considered to be one… more
- Amazon (Cupertino, CA)
- …the Trn1 and Inf1 servers that use them. This role is for a software engineer in the Machine Learning Applications ( ML Apps) team for AWS Neuron. This role ... and runtime engineers to create, build and tune distributed inference solutions with Trn1. Experience optimizing inference ...Inferentia silicon and the TRn1 , Inf1 servers. Strong software development using C++/Python and ML knowledge… more
- Meta (Sunnyvale, CA)
- …generation hardware software codesign for AI domain specific problems. **Required Skills:** Software Engineer , Systems ML - Frameworks / Compilers / ... be a member of the MTIA (Meta Training & Inference Accelerator) Software team and part of...a highly flexible platform to train & serve new DL/ ML model architectures, combined with auto-tuned high performance for… more
- Google (Sunnyvale, CA)
- …UI design and mobile; the list goes on and is growing every day. As a software engineer , you will work on a specific project critical to Google Cloud's needs ... provides ML infrastructure customers with large-scale, cloud-based access to Google's ML supercomputers to run training and inference workloads using PyTorch… more
- Amazon (Cupertino, CA)
- …customer AWS Trainium and Inferentia silicon and the TRn1 , Inf1 servers. Strong software development using C++/Python and ML knowledge are both critical to this ... Description AWS Neuron is the complete software stack for the AWS Inferentia and Trainium...You will develop and extend support for the leading ML frameworks, delivering an outstanding user experience for PyTorch… more
- Amazon (Cupertino, CA)
- …and Inf1 servers that use them. This role is for a senior software engineer in the Machine Learning Applications ( ML Apps) team for AWS Neuron. This role ... will help lead the efforts building distributed training and inference support into Pytorch, Tensorflow using XLA and the...Inferentia silicon and the TRn1 , Inf1 servers. Strong software development and ML knowledge are both… more
- Meta (Menlo Park, CA)
- …do deep technical work. Our work is open-source, cutting-edge, and industry-leading. **Required Skills:** Software Engineer , Systems ML - PyTorch Compiler / ... GPU performance and writing high-performance CUDA kernels. 14. Research and software engineer experience demonstrated via fellowships, patents, internships, or… more
- Amazon (Cupertino, CA)
- Description We are seeking an experienced software engineer with low-level latency networking or interconnect expertise to optimize customer experience by ... and TPUs. This role is on the forefront of AI/ ML , we spend a good deal of the day...customers that uses AWS for large models training and inference will be using your software , and… more
- Snap Inc. (Palo Alto, CA)
- …precision, and always execute with privacy at the forefront. We're looking for a Staff Software Engineer at Snap Inc! What you'll do: + Design, implement, and ... operate our most critical ML Inference / Feature Store services that...or equivalent years of experience + 11+ years of software development experience Preferred Qualifications: + Experience with Java,… more
- Amazon (Palo Alto, CA)
- Description The Amazon Search team owns the software that powers Search - a critical customer-focused feature of Amazon.com. Whenever you visit an Amazon site ... Engine Infrastructure team is responsible for the large-scale distributed software systems that power those results. We design, build...with low latency. We work alongside applied scientists and ML engineers to make this happen. Joining this team,… more
- NVIDIA (Santa Clara, CA)
- At NVIDIA in Santa Clara, CA, USA, we are currently seeking a skilled AI/ ML Infrastructure Engineer to join our team. As an Engineer , you will have a unique ... of AI/ ML workflows, including data processing, model training, and inference pipelines. + Excellent problem-solving skills, with the ability to analyze complex… more
- Amazon (Cupertino, CA)
- …coding services to our most strategic customer opportunities to launch their training and inference workloads on AWS purpose built ML silicon offerings. Key job ... Description Annapurna ML pathfinding team is a new function within...customers to implement their models for both training and inference using the Neuron SDK associated libraries and APIs.… more
- NVIDIA (Santa Clara, CA)
- …services for ML applications in distributed environments! As a solutions engineer , you will work with the profiler team to understand performance monitoring ... APIs, and interfaces used by our customers. A strong understanding of large-scale ML performance analysis is essential. In this role, you will build strong customer… more
- Google (Sunnyvale, CA)
- …techniques + Experience with FlumeC++. Like Google's own ambitions, the work of a Software Engineer goes beyond just Search. Software Engineering Managers ... equivalent practical experience. + 8 years of experience with software development in one or more programming languages (eg,...deploy techniques that improve the performance and efficiency of ML training and ML inference … more
- Amazon (Cupertino, CA)
- …you want to build the future of the cloud for AI training and inference ? Want to do industry leading work delivering continuous price performance improvements in the ... operating AWS cloud offerings that enable high performance and scalability in AI/ ML and HPC workloads. AWS Infrastructure Services owns the design, planning,… more
- Amazon (Cupertino, CA)
- …you want to build the future of the cloud for AI training and inference ? Want to do industry leading work delivering continuous price performance improvements in the ... cloud offerings that enable high performance and scalability in AI/ ML and HPC workloads. You are intrigued by the...- vertically from baremetal server hardware up to the software in userland, and everything in the middle. You… more