Onsite AI Computing Performance Architect Intern, Performance Analysis, Kernel Development

Posted 3 hours ago

Apply now

About the role

  • Performance Architect Intern optimizing major layers and kernels for NVIDIA architectures. Collaborating with teams on in-depth performance analysis and resource optimization.

Responsibilities

  • Design, develop, and optimize major layers in LLM (e.g attention, GEMM, inter-GPU communication) for NVIDIA's new architectures.
  • Implement and fine-tune kernels to achieve optimal performance on NVIDIA GPUs.
  • Conduct in-depth performance analysis of GPU kernels, including Attention and other critical operations.
  • Identify bottlenecks, optimize resource utilization, and improve throughput, and power efficiency
  • Create and maintain workloads and micro-benchmark suites to evaluate kernel performance across various hardware and software configurations.
  • Generate performance projections, comparisons, and detailed analysis reports for internal and external stakeholders.
  • Collaborate with architecture, software, and product teams to guide the development of next-generation deep learning hardware and software.

Requirements

  • Pursuing BS, MS or PhD in relevant discipline (CS, EE, CE).
  • Strong software skills with C/C++, Python, MPI, OpenMP etc.
  • Solid computer science (CS) SW & HW arch background.
  • Experience of DL workload and operator performance will be a plus.
  • Familiarity with GPU computing and parallel programming models will be a plus.
  • Excellent oral and written communication skills.
  • Good organizational, time management and task prioritization skills.

Job title

AI Computing Performance Architect Intern, Performance Analysis, Kernel Development

Job type

Experience level

Entry level

Salary

Not specified

Degree requirement

Bachelor's Degree

Tech skills

Location requirements

Report this job

See something inaccurate? Let us know and we'll update the listing.

Report job