- NVIDIA (Santa Clara, CA)
- …how you can make a lasting impact on the world. We are seeking a highly-skilled Senior On- Device Model Inference Optimization Engineer to join our team ... you'll be doing: + Develop and implement strategies to optimize AI model inference for on- device deployment. + Employ techniques like pruning, quantization,… more
- LinkedIn (Mountain View, CA)
- …large models together. The team is responsible for scaling LinkedIn's AI model training, feature engineering and serving with hundreds of billions of parameters ... and Serving performance optimizations across billions of user queries Model Training Infrastructure: As an engineer on the AI...infra on top of native cloud, enable GPU based inference for a large variety of use cases, cuda… more
- Qualcomm (San Diego, CA)
- …optimization + Generative AI system prototyping + Apply solutions toward system innovations for model efficiency advancement on device as well as in the cloud + ... building advanced machine learning technology, best-in-class solutions, and user friendly model optimization tools such as Qualcomm Innovation Center's AI Model… more
- NVIDIA (Santa Clara, CA)
- …platform upon which every new AI-powered application is built. We are seeking a senior engineer to design and build factory automation for NVIDIA Inference ... NVIDIA optimizes and serves performant inferencing for every AI model . Our NIM offerings are easy to use, highly...working with system software and platform layers including Kernel, device driver, memory, storage, networking and PCIe devices +… more