Hybrid Lead AI Engineer, FM Hosting, LLM Inference

Posted 4 hours ago

Apply now

About the role

  • Lead AI Engineer responsible for developing AI software components and delivering AI-powered products at Capital One. Collaborating with cross-functional teams to enhance customer interaction with innovative AI solutions.

Responsibilities

  • Partner with a cross-functional team of engineers, research scientists, technical program managers, and product managers to deliver AI-powered products that change how our associates work and how our customers interact with Capital One
  • Design, develop, test, deploy, and support AI software components including foundation model training, large language model inference, similarity search, guardrails, model evaluation, experimentation, governance, and observability, etc
  • Leverage a broad stack of Open Source and SaaS AI technologies such as AWS Ultraclusters, Huggingface, VectorDBs, Nemo Guardrails, PyTorch, and more
  • Invent and introduce state-of-the-art LLM optimization techniques to improve the performance — scalability, cost, latency, throughput — of large scale production AI systems
  • Contribute to the technical vision and the long term roadmap of foundational AI systems at Capital One

Requirements

  • Bachelor's degree in Computer Science, AI, Electrical Engineering, Computer Engineering, or related fields plus at least 4 years of experience developing AI and ML algorithms or technologies, or a Master's degree in the same fields plus at least 2 years of experience
  • At least 4 years of experience programming with Python, Go, Scala, or Java
  • Experience deploying scalable and responsible AI solutions on cloud platforms (e.g. AWS, Google Cloud, Azure, or equivalent private cloud)
  • Experience designing, developing, delivering, and supporting AI services
  • Experience developing AI and ML algorithms or technologies (e.g. LLM Inference, Similarity Search and VectorDBs, Guardrails, Memory) using Python, C++, C#, Java, or Golang
  • Experience developing and applying state-of-the-art techniques for optimizing training and inference software to improve hardware utilization, latency, throughput, and cost
  • Passion for staying abreast of the latest AI research and AI systems, and judiciously apply novel techniques in production

Benefits

  • Comprehensive, competitive, and inclusive set of health, financial and other benefits that support your total well-being
  • Performance based incentive compensation, which may include cash bonus(es) and/or long term incentives (LTI)

Job title

Lead AI Engineer, FM Hosting, LLM Inference

Job type

Experience level

Senior

Salary

$197,300 - $245,600 per year

Degree requirement

Bachelor's Degree

Location requirements

Report this job

See something inaccurate? Let us know and we'll update the listing.

Report job