Analyze the performance of LLMs on NVIDIA GPUs by employing advanced profiling and projection tools.
Find opportunities for performance improvements in the IR-based compiler middle end optimizer and/or in precompiled kernel optimizations driven by Graph IR transformations.
Build and develop new compiler passes and optimization techniques to deliver outstanding, robust, and maintainable compiler infrastructure and tools.
Collaborate closely with architecture teams to influence and co-design future hardware features that improve compiler and runtime efficiency.
Work with geographically distributed teams across compiler, hardware, kernel, and framework domains to drive performance improvements and resolve complex issues.
Contribute to a core team at the forefront of deep learning and LLM inference technology, spanning hardware architecture development, kernel optimization, and integration with higher-level deep learning frameworks.
Requirements
Master’s or PhD in Computer Science, Computer Engineering, or a related field, or equivalent experience.
5+ years relevant experience.
Strong hands-on programming expertise in C++ and Python, with solid software engineering fundamentals.
Skilled in innovative LLM architectures, covering inference optimization, profiling, and compiler-level performance tuning.
Significant background in optimizing kernels through information retrieval techniques and generating code, including graph transformations, fusion, scheduling, and developing custom kernel generation frameworks like OpenAI Triton or other compiler-based code generation pipelines.
Hands-on experience with deep learning frameworks like TensorRT-LLM, vLLM, SGLang, Jax/XLA, or related compiler/runtime environments.
Proven ability to analyze and optimize LLM performance bottlenecks across model development, kernel execution, and runtime systems.
Excellent communication and collaboration skills, with the ability to work independently and effectively across distributed teams in a fast-paced environment.
Display a robust determination to continuously improve software and hardware performance by engaging in profiling, analysis, and optimization.
Proficiency in CUDA programming and familiarity with GPU-accelerated deep learning frameworks and performance tuning techniques.
Benefits
equity
benefits
Job title
Senior AI Software Engineer, LLM Inference Performance Analysis
Software Engineer responsible for full - stack development using modern frameworks and cloud solutions in U.S. locations. Collaborative role focusing on delivering quality software and technical mentorship.
Senior Engineer developing and maintaining .NET applications at a leading technology firm in India. Collaborating on system architecture, writing APIs, and utilizing cloud technologies.
Experienced Engineer supporting all phases of major projects in power generation, focusing on project development, compliance, and oversight within Duke Energy.
Software Engineer designing, building, and maintaining software products at MiTek. Join a collaborative team focused on innovation and delivering meaningful solutions.
Principal Software Developer leading design and development of account management applications at LPL Financial. Collaborating with cross - functional teams and implementing AI solutions for enhanced client onboarding.
Senior Full Stack Developer for a tech advisory and software development company. Responsibilities include mentoring, creating web apps, and enhancing client interactions in an agile environment.
Principal Software Engineer at Echodyne developing radar equipment subsystems with an experienced team. Engaging in high - performance software solutions for industry - leading radar technology.