Senior AI Accelerator Architect

Tencent • United State
Relocation
Apply
AI Summary

Design and optimize AI accelerator architectures for large-scale cloud computing environments. Conduct in-depth research on hardware logic and evaluate power-efficiency ratio. Define interconnect architecture and drive virtualization and resource pooling.

Key Highlights
Conduct in-depth research on AI accelerator architectures
Design and optimize AI accelerator architectures for large-scale cloud computing environments
Define interconnect architecture and drive virtualization and resource pooling
Key Responsibilities
Conduct in-depth research into the underlying hardware logic of various AI accelerators
Evaluate the power-efficiency ratio and suitability of different heterogeneous architectures in the context of Large Language Model (LLM) inference and training
Design and optimize high-performance operator libraries for large-scale cloud computing environments
Resolve long-tail latency issues in hardware scheduling, memory management, and distributed communication
Define the interconnect architecture and drive the virtualization, standardized access, and efficient pooling of heterogeneous computing resources in the cloud
Monitor global trends in semiconductors and accelerators and perform feasibility studies and experimental validation for the implementation of emerging technologies within cloud infrastructure
Technical Skills Required
GPGPU architectures CUDA Triton Parallel computing frameworks Low-level operator development languages Large-scale distributed systems Cluster topologies High-performance network protocols
Benefits & Perks
Base pay range: $145,100.00 to $273,200.00 per year
Sign on payment
Relocation package
Restricted stock units
Medical, dental, vision, life, and disability benefits
401(k) plan
15-25 days of vacation per year
13 days of holidays throughout the calendar year
10 days of paid sick leave per year

Job Description


Business Unit

What The Role Entails

  • Architecture Research: Conduct in-depth research into the underlying hardware logic of various AI accelerators; evaluate the power-efficiency ratio and suitability of different heterogeneous architectures in the context of Large Language Model (LLM) inference and training.
  • Operator & Performance Optimization: Design and optimize high-performance operator libraries for large-scale cloud computing environments; resolve long-tail latency issues in hardware scheduling, memory management, and distributed communication.
  • Interconnect Architecture Definition: Define the interconnect architecture ; drive the virtualization, standardized access, and efficient pooling of heterogeneous computing resources in the cloud.
  • Technology Trend Analysis: Monitor global trends in semiconductors and accelerators; perform feasibility studies and experimental validation for the implementation of emerging technologies within cloud infrastructure.

Who We Look For

  • Education: Master’s or Ph.D. degree in Computer Engineering, Electronic Engineering, Microelectronics, or a related field.
  • Core Expertise: Expertise in GPGPU architectures or other mainstream AI accelerator architectures.
  • Programming & Frameworks: Proficient in parallel computing frameworks; deep understanding of low-level operator development languages (e.g., CUDA, Triton).
  • Network & Distributed Systems: Solid understanding of large-scale distributed systems, cluster topologies (e.g., Fat-tree, Torus), and high-performance network protocols.
  • Industry Insight: Familiar with the architectural evolution of global leading computing enterprises; ability to objectively analyze the technical pros/cons and engineering challenges of different architectural paths.
  • Experience: Experience in the application, optimization, or architectural design of ultra-large-scale accelerator clusters is preferred.
  • Framework Optimization: Experience in the low-level adaptation and performance tuning of mainstream deep learning frameworks (e.g., PyTorch, TensorFlow) is preferred.

Location State(s)

US-California-Palo Alto

The expected base pay range for this position in the location(s) listed above is $145,100.00 to $273,200.00 per year. Actual pay may vary depending on job-related knowledge, skills, and experience. Employees hired for this position may be eligible for a sign on payment, relocation package, and restricted stock units, which will be evaluated on a case-by-case basis. Subject to the terms and conditions of the plans in effect, hired applicants are also eligible for medical, dental, vision, life and disability benefits, and participation in the Company’s 401(k) plan. The Employee is also eligible for up to 15 to 25 days of vacation per year (depending on the employee’s tenure), up to 13 days of holidays throughout the calendar year, and up to 10 days of paid sick leave per year. Your benefits may be adjusted to reflect your location, employment status, duration of employment with the company, and position level. Benefits may also be pro-rated for those who start working during the calendar year.

Equal Employment Opportunity at Tencent

As an equal opportunity employer, we firmly believe that diverse voices fuel our innovation and allow us to better serve our users and the community. We foster an environment where every employee of Tencent feels supported and inspired to achieve individual and common goals.

Similar Jobs

Explore other opportunities that match your interests

Visa Sponsorship Relocation Remote
Job Type Full-time
Experience Level Mid-Senior level

Veridian Tech Solutions, Inc.

United State
Visa Sponsorship Relocation Remote
Job Type Full-time
Experience Level Mid-Senior level

mogel

United State
Visa Sponsorship Relocation Remote
Job Type Full-time
Experience Level Director

acm global laboratories

United State

Subscribe our newsletter

New Things Will Always Update Regularly