
About this role
Full Time Senior Machine Learning Engineer - Orchestration in enterprise at ByteDance in San Jose, California, United States. Apply directly through the link below.
At a glance
- Work mode
- Office
- Employment
- Full Time
- Location
- San Jose, California, United States
- Salary
- 213k - 450k USD
- Experience
- Senior
Core stack
- Machine Learning
- Architecture
- Distributed
- Kubernetes
- Efficiency
- Design
- NLP
Quick answers
What is the salary range?
The salary range is 213k - 450k USD annually.
What skills are required?
Machine Learning, Architecture, Distributed, Kubernetes, Efficiency, Design, NLP.
ByteDance is hiring for this role. Visit career page
San Jose, United States
About the Team:
Data AML is ByteDance's Machine Learning mid-platform, providing training and inference systems for recommendation, advertising, CV, speech, and NLP for businesses such as Douyin, Jinri Toutiao, and Xigua Video. It provides powerful Machine Learning computing power to internal business units within the company and conducts research on some general and innovative algorithms for issues in these businesses. At the same time, it also provides some core capabilities of Machine Learning and Recommender systems to external enterprise customers through Volcano Engine. In addition, AML also conducts some cutting-edge research in fields such as Al for Science and scientific computing.
Responsibilities:
1) Optimizing resource efficiency in distributed orchestration and scheduling, through engineering means, enhances the scale of business/models supported per unit of computing power:
a) Use/secondarily develop distributed scheduling frameworks around the Kubernetes/Godel ecosystem, make reasonable selections in different business scenarios, and optimize scheduling strategies for cluster utilization/uniformity based on the characteristics of different scenarios;
b) Connect/extend AutoScaling for various models and business operations, as well as automatic parallelization tasks. Through the method of load modeling and analysis of different models, automatically optimize resource requests for models, optimize resource utilization efficiency at scale, and achieve global optimality;
c) Responsible for the preemption/eviction function of services with different priorities; responsible for the borrowing/mixed deployment docking work among different types of resources in different clusters; responsible for the scheduling/load adaptation in scenarios of multiple data centers, multiple regions, and multiple clouds;
2) Build a training system architecture for next-generation ultra-large and ultra-deep recommendation models:
a) Build a flexible and robust distributed training runtime around ultra-large-scale embedding and ultra-large-scale GPU synchronization training;
Design and optimize distributed computing APis and runtime for future-oriented research paradigms of recommended advertising models (e.g., RL/finetune/distillation);
c) Interface with the platform to optimize the diagnosability and usability of distributed training.
3) Construct an online orchestration architecture for the next-generation Recommender system:
a) Build a robust and stable distributed model inference architecture around the online training scenario of ultra-large-scale embeddings;
b) Optimize the usability of the online architecture of the recommended advertising model and the MLops process by integrating the research and experimental model of the business.
The base salary range for this position in the selected city is $212800 - $450000 annually.
Data AML is ByteDance's Machine Learning mid-platform, providing training and inference systems for recommendation, advertising, CV, speech, and NLP for businesses such as Douyin, Jinri Toutiao, and Xigua Video. It provides powerful Machine Learning computing power to internal business units within the company and conducts research on some general and innovative algorithms for issues in these businesses. At the same time, it also provides some core capabilities of Machine Learning and Recommender systems to external enterprise customers through Volcano Engine. In addition, AML also conducts some cutting-edge research in fields such as Al for Science and scientific computing.
Responsibilities:
1) Optimizing resource efficiency in distributed orchestration and scheduling, through engineering means, enhances the scale of business/models supported per unit of computing power:
a) Use/secondarily develop distributed scheduling frameworks around the Kubernetes/Godel ecosystem, make reasonable selections in different business scenarios, and optimize scheduling strategies for cluster utilization/uniformity based on the characteristics of different scenarios;
b) Connect/extend AutoScaling for various models and business operations, as well as automatic parallelization tasks. Through the method of load modeling and analysis of different models, automatically optimize resource requests for models, optimize resource utilization efficiency at scale, and achieve global optimality;
c) Responsible for the preemption/eviction function of services with different priorities; responsible for the borrowing/mixed deployment docking work among different types of resources in different clusters; responsible for the scheduling/load adaptation in scenarios of multiple data centers, multiple regions, and multiple clouds;
2) Build a training system architecture for next-generation ultra-large and ultra-deep recommendation models:
a) Build a flexible and robust distributed training runtime around ultra-large-scale embedding and ultra-large-scale GPU synchronization training;
Design and optimize distributed computing APis and runtime for future-oriented research paradigms of recommended advertising models (e.g., RL/finetune/distillation);
c) Interface with the platform to optimize the diagnosability and usability of distributed training.
3) Construct an online orchestration architecture for the next-generation Recommender system:
a) Build a robust and stable distributed model inference architecture around the online training scenario of ultra-large-scale embeddings;
b) Optimize the usability of the online architecture of the recommended advertising model and the MLops process by integrating the research and experimental model of the business.
The base salary range for this position in the selected city is $212800 - $450000 annually.
Job details
Workplace
Office
Location
San Jose, California, United States
Job type
Full Time
Experience
Senior
Salary
213k - 450k USD
per year
Jobr Assistant extension
Get the extension →