All active PyTorch roles based in Toronto.
Pick a job to read the details
Tap any role on the left — its description and apply link will open here.
Stripe is a financial infrastructure platform for businesses. Millions of companies—from the world’s largest enterprises to the most ambitious startups—use Stripe to accept payments, grow their revenue, and accelerate new business opportunities. Our mission is to increase the GDP of the internet, and we have a staggering amount of work ahead. That means you have an unprecedented opportunity to put the global economy within everyone’s reach while doing the most important work of your career.
The Support Experience engineering organization builds and improves Stripe’s user support from end to end: how users get help within our products, how they get in touch with us when they have questions, and how our teams use internal tools to answer those questions. We’re accountable for the quality and reliability of this support stack and we use data and firsthand user research to continuously improve it.
Providing great support to users of all sizes is culturally important to everyone at Stripe. We are a group of friendly, user-oriented engineers that partner closely with Stripe’s world-class design, product, and operational teams. This includes the external-facing support interfaces (support.stripe.com), content, entry points, internal tooling, case routing, and helping product teams across the company reduce support volume by improving our products. We are also using the latest generative AI technologies to re-imagine support experiences, and are developing AI assistants for Stripe’s users and internally to help our operations teams be more productive.
As a Machine Learning Engineer on the Support Experience team, you'll play a crucial role in enhancing our self-serve support experiences. You will be responsible for designing, building, training, evaluating, deploying, and owning ML models in production. For example, we apply LLMs to answer user questions with conversational agents and personalize product documentation, and are building automated systems to solve complex user problems. You will work closely with software engineers, machine learning engineers, product managers, and data scientists to operate Stripe’s ML powered systems, features, and products. You will also have the opportunity to contribute to and influence ML architecture at Stripe and be a part of a larger ML community.
We are looking for ML Engineers who are passionate about building ML systems that touch the lives of millions. You have experience developing efficient feature pipelines, building advanced ML models, and deploying them to production. You are comfortable with ambiguity, love to take initiative, have a bias towards action, and thrive in a collaborative environment.
Ready to apply?
Apply to Stripe
Stripe is a financial infrastructure platform for businesses. Millions of companies—from the world’s largest enterprises to the most ambitious startups—use Stripe to accept payments, grow their revenue, and accelerate new business opportunities. Our mission is to increase the GDP of the internet, and we have a staggering amount of work ahead. That means you have an unprecedented opportunity to put the global economy within everyone’s reach while doing the most important work of your career.
The Supportability Evaluation team acts as stewards of the financial ecosystem. Our mission is to protect Stripe’s reputation with our global financial partners by architecting highly precise, automated supportability controls. We develop the AI/ML models and systems that detect and action supportability violations in real-time. We're responsible for building high-fidelity detection engines that ensure our merchants remain compliant across the globe, balancing the scale of millions of users with the surgical precision required by the world’s largest financial institutions.
As a Machine Learning Engineer in Supportability, you will be responsible for designing, building, training, evaluating, deploying, and owning AI/ML models in production. You will work closely with software engineers, machine learning engineers, product managers, and data scientists to operate Stripe’s ML powered systems, features, and products. You will also have the opportunity to contribute to and influence AI/ML architecture at Stripe and be a part of a larger community.
We are looking for ML Engineers who are passionate about building AI/ML and AI systems that touch the lives of millions. You have experience building and evaluating advanced AI/ML models, and deploying them to production. You are comfortable with ambiguity, love to take initiative, have a bias towards action, and thrive in a collaborative environment.
Ready to apply?
Apply to Stripe
Share this job
About Faire
Faire is a technology wholesale platform built on the belief that the future is local. Independent retailers around the globe collectively represent a multi-hundred-billion-dollar wholesale market that has historically been fragmented and offline. At Faire, we're using the power of tech, data, and machine learning to connect this thriving community of entrepreneurs across the globe. Picture your favorite boutique in town — we help them discover the best products from around the world to sell in their stores. With the right tools and insights, we believe that we can level the playing field so businesses can grow and local communities can thrive.
We’re looking for smart, resourceful and passionate people to join us as we power the shop local movement. If you believe in community, come join ours.
About this role
As a Staff Machine Learning Platform Engineer, you will help design, improve, and operate a scalable ML platform to accelerate model training, deployment, and governance. You are the technical bridge between data science and production engineering. You’ll be joining a small but deeply critical team that scales Faire’s ability to support tens of thousands of local businesses in a constantly narrowing retail landscape.
What You Will Do
What it takes
Tech Stack
Faire uses a modern cloud based tech stack. For this role, you’ll want to be proficient with the following:
|
Category |
Technologies |
|
Languages |
Python, SQL, Kotlin |
|
ML Frameworks |
PyTorch, MLFlow |
|
Big Data & Processing |
Spark, Kafka, Databricks, Snowflake, Fivetran, Iceberg, Unity Catalog, Datadog, Airflow, Cockroach DB, MySQL |
|
Cloud & Infrastructure |
AWS, S3, SageMaker, Kubernetes, Docker, GitHub Actions, Terraform |
|
Generative AI |
Claude Sonnet 4.5, ChatGPT 5.2 |
Salary Range
Canada: the pay range for this role is $216,000 to $297,000 per year.
This role will also be eligible for equity and benefits. Actual base pay will be determined based on permissible factors such as transferable skills, work experience, market demands, and primary work location. The base pay range provided is subject to change and may be modified in the future.
Faire uses Artificial Intelligence (AI) to screen and select applicants for this position.
This job posting is for an existing vacancy.
Hybrid Faire employees currently go into the office 3 days per week on Tuesdays, Thursdays, and a third flex day of their choosing (Monday, Wednesday, or Friday). Additionally, hybrid in-office roles will have the flexibility to work remotely up to 4 weeks per year. Specific Workplace and Information Technology positions may require onsite attendance 5 days per week as will be indicated in the job posting.
Why you’ll love working at Faire
Faire was founded in 2017 by a team of early product and engineering leads from Square. We’re backed by some of the top investors in retail and tech including: Y Combinator, Lightspeed Venture Partners, Forerunner Ventures, Khosla Ventures, Sequoia Capital, Founders Fund, and DST Global. We have headquarters in San Francisco and Kitchener-Waterloo, and a global employee presence across offices in Toronto, London, and New York. To learn more about Faire and our customers, you can read more on our blog.
Faire provides equal employment opportunities (EEO) to all employees and applicants for employment without regard to race, color, religion, sex, national origin, age, disability, genetics, sexual orientation, gender identity or gender expression.
Faire is committed to providing access, equal opportunity and reasonable accommodation for individuals with disabilities in employment, its services, programs, and activities. Accommodations are available throughout the recruitment process and applicants with a disability may request to be accommodated throughout the recruitment process. We will work with all applicants to accommodate their individual accessibility needs. To request reasonable accommodation, please fill out our Accommodation Request Form (https://bit.ly/faire-form)
Privacy
For information about the type of personal data Faire collects from applicants, as well as your choices regarding the data collected about you, please visit Faire’s Privacy Notice (https://www.faire.com/privacy)
Ready to apply?
Apply to Faire
Are you looking to thrive in a stimulating work environment?
Join Levio, a leader in digital transformation, and take your career to the next level. You will work alongside high-caliber professionals on ambitious, large-scale technology projects, directly embedded in our clients’ environments. At Levio, we value expertise, curiosity, and continuous improvement — and we give you the space to grow.
The salary range provided reflects a good faith estimate based on factors such as experience, technical expertise, location, and relevant certifications. Final compensation will be determined according to the specific circumstances of each candidate.
Estimated salary range: $100,000 to $140,000 per year.
This posting is a current hiring need.
Levio offers a comprehensive and flexible benefits package designed to support your professional growth and personal wellbeing, including:
Position Details
Notice on the Use of Artificial Intelligence in Recruitment
We use AI enabled tools to help sort and review applications based on job related criteria. Final decisions regarding candidate progression are always made by a human recruiter.
Employment Equity
Levio subscribes to the principle of employment equity and applies an equal access employment program for women, Indigenous peoples, visible minorities, ethnic minorities, and persons with disabilities.
We value diversity and inclusion and are committed to creating a healthy, accessible, and rewarding work environment that highlights the unique contributions of our employees. Accommodations are available upon request for candidates participating in all aspects of the selection process.
Ready to apply?
Apply to Levio
Join Levio, a leader in digital transformation, and take your career to the next level. You will work alongside high-caliber professionals on ambitious, large-scale technology projects, directly embedded in our clients’ environments. At Levio, we value expertise, curiosity, and continuous improvement — and we give you the space to grow.
The ML / AI Engineer design, build, deploy, and operate production-grade machine learning and generative AI systems. This role owns the end-to-end ML lifecycle, ensuring models and AI services are scalable, reliable, secure, and deliver measurable business value. The role will be remote.
The salary range provided reflects a good faith estimate based on factors such as experience, technical expertise, location, and relevant certifications. Final compensation will be determined according to the specific circumstances of each candidate.
Estimated salary range: $110,000 to $150,000 per year.
This posting is a current hiring need.
Levio offers a comprehensive and flexible benefits package designed to support your professional growth and personal wellbeing, including:
Position Details
Notice on the Use of Artificial Intelligence in Recruitment
We use AI enabled tools to help sort and review applications based on job related criteria. Final decisions regarding candidate progression are always made by a human recruiter.
Employment Equity
Levio subscribes to the principle of employment equity and applies an equal access employment program for women, Indigenous peoples, visible minorities, ethnic minorities, and persons with disabilities.
We value diversity and inclusion and are committed to creating a healthy, accessible, and rewarding work environment that highlights the unique contributions of our employees. Accommodations are available upon request for candidates participating in all aspects of the selection process.
Ready to apply?
Apply to Levio
Share this job
We are seeking a Senior Machine Learning Engineer to join the Growth Tech Alliance. In this role, you will architect and deploy the robust infrastructure behind our intelligent marketing systems. You will be responsible for maturing algorithmic prototypes into high-performance production systems, ensuring our AI-driven marketing optimization is served reliably and autonomously at a global scale.
S'more about the team
We are hiring a Senior Machine Learning Engineer to take our AI tooling to the next level by architecting and deploying the robust infrastructure behind our intelligent marketing optimization systems. You will provide critical engineering execution for our AI initiatives. You will develop scalable microservices for predictive scoring, orchestrate complex LLM-based agents for creative intelligence. As the ML engineering expert for the team, you will drive the maturation of algorithmic prototypes into high-performance production systems with maximum Speed & Agility, shaping the future of how HelloFresh automates marketing at an unprecedented scale.
Lettuce share what this role will be responsible for
As a core member of the engineering team, you will focus on productionizing ML infrastructure across several domains:
Sound a-peeling? Here's what we're looking for
Let’s cut to the cheese, this is why you'll love it here
Flexible Hybrid Approach
At HelloFresh, we know that flexible work arrangements are essential in enabling you to do your best work, while balancing your personal and life needs. Offering remote work flexibility, along with the opportunity to interact and collaborate in the office are all a part of creating a great employee experience.
To meet these needs, we are pleased to provide Flexible Hybrid work. Flexible Hybrid is a people-first approach that is based on choice, trust, personalization, and empowers teams to choose when and how often they work from the office and work from home, in addition to team days and company days. This means a minimum of 2 days in office per week, with most teams in office between 2-3 days a week.
#LI-HYBRID
#Engineering
HelloFresh Canada uses AI-integrated technology to help us process and evaluate applications more efficiently. This includes tools that screen and assess candidate qualifications based on the requirements for this role. While these tools assist our workflow, all final selection decisions are made by our hiring team.
This is a posting for an existing vacancy. We are actively seeking to fill this position.
Ready to apply?
Apply to HelloFresh
Share this job
Tenstorrent is leading the industry on cutting-edge AI technology, revolutionizing performance expectations, ease of use, and cost efficiency. With AI redefining the computing paradigm, solutions must evolve to unify innovations in software models, compilers, platforms, networking, and semiconductors. Our diverse team of technologists have developed a high performance RISC-V CPU from scratch, and share a passion for AI and a deep desire to build the best AI platform possible. We value collaboration, curiosity, and a commitment to solving hard problems. We are growing our team and looking for contributors of all seniorities.
Join the team revolutionizing AI computing at Tenstorrent. You'll work on TT-Forge, our MLIR-based compiler that enables developers to run AI on all configurations of Tenstorrent hardware using an open-source, performant, and general-purpose compiler. You will be at the forefront of the AI hardware revolution, building compiler technologies that redefine what’s possible.
This role is hybrid and based out of Toronto, ON.
We welcome candidates at various experience levels for this role. During the interview process, candidates will be assessed for the appropriate level, and offers will align with that level, which may differ from the one in this posting.
Who You Are
What We Need
What You Will Learn
Compensation for all engineers at Tenstorrent ranges from $100k - $500k including base and variable compensation targets. Experience, skills, education, background and location all impact the actual offer made.
Tenstorrent offers a highly competitive compensation package and benefits, and we are an equal opportunity employer.
This offer of employment is contingent upon the applicant being eligible to access U.S. export-controlled technology. Due to U.S. export laws, including those codified in the U.S. Export Administration Regulations (EAR), the Company is required to ensure compliance with these laws when transferring technology to nationals of certain countries (such as EAR Country Groups D:1, E1, and E2). These requirements apply to persons located in the U.S. and all countries outside the U.S. As the position offered will have direct and/or indirect access to information, systems, or technologies subject to these laws, the offer may be contingent upon your citizenship/permanent residency status or ability to obtain prior license approval from the U.S. Commerce Department or applicable federal agency. If employment is not possible due to U.S. export laws, any offer of employment will be rescinded.
Ready to apply?
Apply to Tenstorrent
Share this job
Tenstorrent is leading the industry on cutting-edge AI technology, revolutionizing performance expectations, ease of use, and cost efficiency. With AI redefining the computing paradigm, solutions must evolve to unify innovations in software models, compilers, platforms, networking, and semiconductors. Our diverse team of technologists have developed a high performance RISC-V CPU from scratch, and share a passion for AI and a deep desire to build the best AI platform possible. We value collaboration, curiosity, and a commitment to solving hard problems. We are growing our team and looking for contributors of all seniorities.
Join the team revolutionizing AI computing at Tenstorrent. You'll work on TT-Forge, our MLIR-based compiler that enables developers to run AI on all configurations of Tenstorrent hardware using an open-source, performant, and general-purpose compiler. You will be at the forefront of the AI hardware revolution, building compiler technologies that redefine what’s possible.
This role is hybrid, and can be based out of Santa Clara, CA; Austin, TX; or Toronto; ON.
We welcome candidates at various experience levels for this role. During the interview process, candidates will be assessed for the appropriate level, and offers will align with that level, which may differ from the one in this posting.
Who You Are
What We Need
What You Will Learn
Compensation for all engineers at Tenstorrent ranges from $100k - $500k including base and variable compensation targets. Experience, skills, education, background and location all impact the actual offer made.
Tenstorrent offers a highly competitive compensation package and benefits, and we are an equal opportunity employer.
This offer of employment is contingent upon the applicant being eligible to access U.S. export-controlled technology. Due to U.S. export laws, including those codified in the U.S. Export Administration Regulations (EAR), the Company is required to ensure compliance with these laws when transferring technology to nationals of certain countries (such as EAR Country Groups D:1, E1, and E2). These requirements apply to persons located in the U.S. and all countries outside the U.S. As the position offered will have direct and/or indirect access to information, systems, or technologies subject to these laws, the offer may be contingent upon your citizenship/permanent residency status or ability to obtain prior license approval from the U.S. Commerce Department or applicable federal agency. If employment is not possible due to U.S. export laws, any offer of employment will be rescinded.
Ready to apply?
Apply to Tenstorrent
At Lyft, our purpose is to serve and connect. We aim to achieve this by cultivating a work environment where all team members belong and have the opportunity to thrive.
As a Data Scientist on the Mapping team, you will collaborate with our world class team of engineers, product managers, and designers to grow and improve the quality of recommended routes and accuracy of our travel time estimations. We're looking for a passionate, driven Data Scientist who is excited to dive into our spatial data and build a best-in-class mapping product that provides safe, efficient, and seamless navigation for our rideshare drivers.
Data Science is at the heart of Lyft’s products and decision-making. You will leverage data and rigorous, analytical thinking to shape our mapping products and make business decisions that put our customers first. This will involve identifying and scoping opportunities, shaping priorities, recommending technical solutions, designing experiments, and measuring the impact of new features. You will help us solve some of the most impactful problems in mapping, including:
Lyft is committed to creating an inclusive workforce that fosters belonging. Lyft believes that every person has a right to equal employment opportunities without discrimination because of race, ancestry, place of origin, colour, ethnic origin, citizenship, creed, sex, sexual orientation, gender identity, gender expression, age, marital status, family status, disability, pardoned record of offences, or any other basis protected by applicable law or by Company policy. Lyft also strives for a healthy and safe workplace and strictly prohibits harassment of any kind. Accommodation for persons with disabilities will be provided upon request in accordance with applicable law during the application and hiring process. Please contact your recruiter if you wish to make such a request.
Lyft highly values having employees working in-office to foster a collaborative work environment and company culture. This role will be in-office on a hybrid schedule — Team Members will be expected to work in the office at least 3 days per week, including on Mondays, Wednesdays, and Thursdays. Lyft considers working in the office at least 3 days per week to be an essential function of this hybrid role. Your recruiter can share more information about the various in-office perks Lyft offers. Additionally, hybrid roles have the flexibility to work from anywhere for up to 4 weeks per year. #Hybrid
The expected base pay range for this position in the Toronto area is $108,000 - $135,000, not inclusive of potential equity offering, bonus or benefits. Salary ranges are dependent on a variety of factors, including qualifications, experience and geographic location. Your recruiter can share more information about the salary range specific to your working location and other factors during the hiring process.
Lyft may use artificial intelligence to screen applicants, however, Lyft employees make the ultimate selection and hiring decisions.
This job fills an existing vacancy.
Ready to apply?
Apply to Lyft
Share this job
Upwork Inc.’s (Nasdaq: UPWK) family of companies connects businesses with global, AI-enabled talent across every contingent work type including freelance, fractional, and payrolled. This portfolio includes the Upwork Marketplace, which connects businesses with on-demand access to highly skilled talent across the globe, and Lifted, which provides a purpose-built solution for enterprise organizations to source, contract, manage, and pay talent across the full spectrum of contingent work. From Fortune 100 enterprises to entrepreneurs, businesses rely on Upwork Inc. to find and hire expert talent, leverage AI-powered work solutions, and drive business transformation. With access to professionals spanning more than 10,000 skills across AI & machine learning, software development, sales & marketing, customer support, finance & accounting, and more, the Upwork family of companies enables businesses of all sizes to scale, innovate, and transform their workforces for the age of AI and beyond.
Since its founding, Upwork Inc. has facilitated more than $30 billion in total transactions and services as it fulfills its purpose to create opportunity in every era of work. Learn more about the Upwork Marketplace at Upwork.com and follow us on LinkedIn, Facebook, Instagram, TikTok, and X; and learn more about Lifted at Go-Lifted and follow on LinkedIn.
The AI Foundations team leads core research and development across the training, evaluation, and deployment of AI systems that power Uma, Upwork’s flagship AI model, and other customer-facing generative AI capabilities. As a Sr. Lead AI Research Scientist focused on AI Evaluation and Reliability, you will drive high-impact research initiatives that improve the trustworthiness, robustness, and real-world performance of AI systems operating at marketplace scale.
At the Sr. Lead level, this role combines deep technical expertise with cross-functional leadership. You will identify and lead research efforts that address systemic reliability challenges, partner closely with engineering and product teams to translate research into production outcomes, and help shape how Upwork evaluates AI performance in real work scenarios. Your work will support AI systems embedded in retrieval-based workflows, agentic architectures, and human plus AI collaboration patterns, while contributing to Upwork’s broader AI research strategy and external presence.
Lead applied research initiatives focused on AI evaluation, reliability, and robustness, defining success metrics tied to customer impact and production readiness.
Design and validate methods to measure and mitigate AI reliability risks, including uncertainty estimation, hallucination detection, and identification of model failure modes.
Partner cross-functionally with engineering, data science, and product teams to integrate research outcomes into customer-facing AI systems and workflows.
Own research projects end to end, from problem framing and hypothesis development through experimentation, prototyping, and synthesis of results.
Influence technical direction across teams by surfacing insights, proposing scalable solutions, and aligning stakeholders on priorities and tradeoffs.
Mentor researchers and engineers through technical guidance, feedback, and collaborative leadership on shared initiatives.
Contribute to Upwork’s external research footprint through publications, presentations, and engagement with the broader AI research community.
Proven experience leading applied AI research that balances scientific rigor with real-world deployment constraints and business impact.
A strong record of research contribution through publications, internal innovation, or demonstrable influence on production AI systems.
Deep proficiency with Python and modern deep learning frameworks such as PyTorch, with hands-on experience evaluating and improving large-scale models.
An adaptive approach to integrating AI tools into research and development workflows to accelerate experimentation, improve evaluation quality, and share best practices with others.
Come change how the world works.
Upwork is establishing an operational hub in Toronto, Canada. The new office is expected to be fully operational by Q4 2026. This role will require 3 days in office once we have an office open.
This position will initially be employed through a partner to ensure a seamless hiring process while we establish the hub. Once the hub is established, there may be opportunities to transition to employment with Upwork, depending on business needs and other requirements. While employed by the partner, you’ll work as part of Upwork’s team, with access to our resources, culture, and growth opportunities.
Our partner will offer competitive benefits. When Upwork’s hub is established, we will be excited to offer employment and benefits directly as business needs require.
Upwork is committed to building a diverse, inclusive, and equitable workforce. Employment decisions are made without regard to race, color, religion, gender, sexual orientation, gender identity, national origin, disability, or any other status protected by applicable law.
We use BrightHire, an AI-enabled tool, to record interviews and summarize interview transcripts. The tool allows the interviewer to focus on the discussion and does not score or evaluate candidates or make recommendations. The interview transcripts are reviewed, and decisions are only made by humans. Candidates who prefer not to have their interview recorded through BrightHire can opt out when the interview is scheduled.
To learn more about how Upwork processes and protects your personal information as part of the application process, please review our Global Job Applicant Privacy Notice and the Applicant Privacy Addendum (Canada).
To learn more about how Upwork processes and protects your personal information as part of the application process, please review our Global Job Applicant Privacy Notice
Ready to apply?
Apply to Upwork
Share this job
We are a Canadian leader in digital automotive solutions. Our flagship brands — AutoTrader.ca, AutoSync, Dealertrack Canada and CMS — help Canadians buy, sell, and finance vehicles with confidence.
AutoTrader.ca is Canada’s largest automotive marketplace, with over 25 million monthly visits.
As part of AutoScout24 group, Europe’s largest online car marketplace, we’re shaping the future of automotive retail in Canada and beyond.
The base salary range for this position is CAD 180K – CAD 220K.
This range reflects the expected compensation at the time of posting. The final offer may vary and can be higher based on relevant skills, experience, location, and market conditions. Based on the role the total rewards package may also include benefits, bonus, and other employee offerings.
What's in it for you:
We understand that there is life at work and life outside of work. Here are a few benefits we all benefit from that support us to be our creative best.
For a career where you can drive our business and shape your future, apply now.
Use of Artificial Intelligence in Hiring: We use artificial intelligence (“AI”) in our hiring process, including to screen, assess, or select applicants for this position.
Vacancy Status: This job posting is for an existing vacancy.
Ready to apply?
Apply to AutoTrader.ca
Share this job
About the Role:
The Machine Learning team at Tubi drives the innovation behind personalized user experiences. With the largest inventory in the industry and hundreds of millions of viewers, we tackle problems in the space of recommendations, search, content understanding and ads optimization that shape the future of streaming.
We are seeking a highly skilled Machine Learning Engineer to contribute to transformative projects in video personalization. In this role, you will design and implement advanced algorithms and systems to improve our personalization strategy. As a senior technical expert, you will tackle complex problems in machine learning at scale, collaborating closely with cross-functional teams to develop and optimize machine learning-driven solutions.
What You'll Do:
Your Background:
#LI-Hybrid #LI-SC1
Pursuant to state and local pay disclosure requirements, the pay range for this role, with final offer amount dependent on education, skills, experience, and location is is listed annually below. This role is also eligible for an annual discretionary bonus, long-term incentive plan, and various benefits including medical/dental/vision, insurance, a 401(k) plan, paid time off and other benefits in accordance with applicable plan documents.
High cost labor markets such as but not limited to Los Angeles, New York City, and San Francisco
Tubi is a division of Fox Corporation, and the FOX Employee Benefits summarized here, covers the majority of all US employee benefits. The following distinctions below outline the differences between the Tubi and FOX benefits:
Boldly built for every fandom, Tubi is a free streaming service that entertains over 100 million monthly active users. Tubi offers the world's largest collection of Hollywood movies and TV shows, thousands of creator-led stories and hundreds of Tubi Originals made for the most passionate fans. Headquartered in San Francisco and founded in 2014, Tubi is part of Tubi Media Group, a division of Fox Corporation.
We are an equal opportunity employer and all qualified applicants will receive consideration for employment without regard to race, color, religion, sex, national origin, gender identity, disability, protected veteran status, or any other characteristic protected by law. We will consider for employment qualified applicants with criminal histories consistent with applicable law.
Ready to apply?
Apply to Tubi
Share this job
The Machine Learning team at Tubi drives the innovation behind personalized user experiences. With the largest inventory in the industry and hundreds of millions of viewers, we tackle problems in the space of recommendations, search, content understanding, and ads optimization that shape the future of streaming.
We are seeking a Director of Machine Learning Engineering and Infrastructure to lead a hybrid team bridging advanced ML engineering with world-class infrastructure design. In this role, you will own the strategic direction and execution for scaling our machine learning capabilities while ensuring our distributed systems and infrastructure can support innovation at massive scale. You will combine technical depth with leadership excellence to guide teams that deliver both foundational ML systems and high-performance distributed services.
What You'll Do:
Your Background:
Pursuant to state and local pay disclosure requirements, the pay range for this role, with final offer amount dependent on education, skills, experience, and location is listed annually below. This role is also eligible for an annual discretionary bonus, long-term incentive plan, and various benefits including medical/dental/vision, insurance, a 401(k) plan, paid time off and other benefits in accordance with applicable plan documents.
Tubi is a division of Fox Corporation, and the FOX Employee Benefits summarized here, covers the majority of all US employee benefits. The following distinctions below outline the differences between the Tubi and FOX benefits:
Boldly built for every fandom, Tubi is a free streaming service that entertains over 100 million monthly active users. Tubi offers the world's largest collection of Hollywood movies and TV shows, thousands of creator-led stories and hundreds of Tubi Originals made for the most passionate fans. Headquartered in San Francisco and founded in 2014, Tubi is part of Tubi Media Group, a division of Fox Corporation.
We are an equal opportunity employer and all qualified applicants will receive consideration for employment without regard to race, color, religion, sex, national origin, gender identity, disability, protected veteran status, or any other characteristic protected by law. We will consider for employment qualified applicants with criminal histories consistent with applicable law.
Ready to apply?
Apply to Tubi
Share this job
MaintainX is the world’s leading mobile-first Asset and Work Intelligence platform for industrial and frontline environments. We’re a modern, IoT-enabled, cloud-based solution that powers maintenance, safety, and operations on physical equipment and facilities.
We help 12,000+ organizations—including Duracell, Univar Solutions, Titan America, McDonald’s, Brenntag, Cintas, Xylem, and Shell—achieve operational excellence and reliability at scale.
Following our $150 million Series D led by Bain Capital Ventures, Bessemer Ventures, August Capital, Amity Ventures, and Ridge Ventures, MaintainX has raised a total of $254 million, valuing the company at $2.5 billion.
As we enter our next phase of growth, we’re investing deeply in AI/ML, LLMs, and Industrial IoT to transform how frontline teams operate—predicting failures before they happen, automating workflows, and embedding intelligence into every asset and procedure.
What you’ll do:
About you:
Bonus skills:
What’s in it for you:
About us:
We exist to make the lives of frontline and maintenance teams easier by building software that meets their real-world needs. Our product transforms how 80% of the global workforce—those who don’t sit behind a desk—manage their operations, assets, and teams.
MaintainX is committed to creating a diverse environment. All qualified applicants will receive consideration for employment without regard to race, color, religion, gender, gender identity or expression, sexual orientation, national origin, genetics, disability, age, or veteran status.
Ready to apply?
Apply to MaintainX
Share this job
MaintainX is the world’s leading mobile-first Asset and Work Intelligence platform for industrial and frontline environments. We’re a modern, IoT-enabled, cloud-based solution that powers maintenance, safety, and operations on physical equipment and facilities.
We help 12,000+ organizations—including Duracell, Univar Solutions, Titan America, McDonald’s, Brenntag, Cintas, Xylem, and Shell—achieve operational excellence and reliability at scale.
Following our $150 million Series D led by Bain Capital Ventures, Bessemer Ventures, August Capital, Amity Ventures, and Ridge Ventures, MaintainX has raised a total of $254 million, valuing the company at $2.5 billion.
As we enter our next phase of growth, we’re investing deeply in AI/ML, LLMs, and Industrial IoT to transform how frontline teams operate—predicting failures before they happen, automating workflows, and embedding intelligence into every asset and procedure.
We are seeking a highly skilled and motivated Senior Applied Machine Learning Developer to guide the technical direction and architecture of our Predictive Maintenance and Asset Intelligence initiatives.
You’ll combine deep ML expertise with strong software development and leadership skills—mentoring developers, scaling systems, and driving the roadmap for AI-enabled maintenance intelligence across thousands of industrial sites.
This role sits at the intersection of ML architecture, IoT data systems, and product impact, shaping the foundation for MaintainX’s predictive and generative AI strategy.
What you’ll do:
About you:
Bonus skills:
What’s in it for you:
About us:
We exist to make the lives of frontline and maintenance teams easier by building software that meets their real-world needs. Our product transforms how 80% of the global workforce—those who don’t sit behind a desk—manage their operations, assets, and teams.
MaintainX is committed to creating a diverse environment. All qualified applicants will receive consideration for employment without regard to race, color, religion, gender, gender identity or expression, sexual orientation, national origin, genetics, disability, age, or veteran status.
Ready to apply?
Apply to MaintainX
Share this job
MaintainX est la première plateforme mobile au monde dédiée à la gestion des actifs et des tâches dans les environnements industriels et de première ligne. Nous proposons une solution moderne, basée sur l'IoT et le cloud, qui facilite la maintenance, la sécurité et l'exploitation des équipements et installations physiques.
Nous aidons plus de 12 000 organisations, dont Duracell, Univar Solutions, Titan America, McDonald's, Brenntag, Cintas, Xylem et Shell, à atteindre l'excellence opérationnelle et la fiabilité à grande échelle.
À la suite de notre série D de 150 millions de dollars menée par Bain Capital Ventures, Bessemer Ventures, August Capital, Amity Ventures et Ridge Ventures, MaintainX a levé un total de 254 millions de dollars, valorisant la société à 2,5 milliards de dollars.
Alors que nous entrons dans notre prochaine phase de croissance, nous investissons massivement dans l'IA/ML, les LLM et l'IoT industriel afin de transformer le mode de fonctionnement des équipes de première ligne, en prédisant les pannes avant qu'elles ne se produisent, en automatisant les flux de travail et en intégrant l'intelligence dans chaque actif et chaque procédure.
Nous recherchons un développeur(se) senior en apprentissage automatique appliqué hautement qualifié et motivé pour orienter la direction technique et l'architecture de nos initiatives en matière de maintenance prédictive et d'intelligence des actifs.
Vous combinerez une expertise approfondie en apprentissage automatique avec de solides compétences en génie logiciel et en sens du leadership. Vous encadrerez des ingénieurs, développerez des systèmes et piloterez la feuille de route pour l'intelligence de maintenance basée sur l'IA sur des milliers de sites industriels.
Ce poste se situe à la croisée de l'architecture d'apprentissage automatique, des systèmes de données IoT et de l'impact des produits, et constitue le fondement de la stratégie d'IA prédictive et générative de MaintainX.
Ce que vous ferez:
À propos de vous:
Une attention particulière est accordée aux candidats présentant les caractéristiques suivantes:
Quels sont les avantages pour vous?:
Qui sommes-nous:
MaintainX s'engage à créer un environnement diversifié. Tous les candidats qualifiés seront pris en considération pour un emploi sans considération de race, de couleur, de religion, de sexe, d'identité ou d'expression de genre, d'orientation sexuelle, d'origine nationale, de génétique, d'invalidité, d'âge ou de statut d'ancien combattant.
Ready to apply?
Apply to MaintainX
Share this job
MaintainX est la première plateforme mobile au monde dédiée à la gestion des actifs et des tâches dans les environnements industriels et de première ligne. Nous proposons une solution moderne, basée sur l'IoT et le cloud, qui facilite la maintenance, la sécurité et l'exploitation des équipements et installations physiques. Nous aidons plus de 12 000 organisations, dont Duracell, Univar Solutions, Titan America, McDonald's, Brenntag, Cintas, Xylem et Shell, à atteindre l'excellence opérationnelle et la fiabilité à grande échelle. À la suite de notre série D de 150 millions de dollars menée par Bain Capital Ventures, Bessemer Ventures, August Capital, Amity Ventures et Ridge Ventures, MaintainX a levé un total de 254 millions de dollars, valorisant la société à 2,5 milliards de dollars. Alors que nous entrons dans notre prochaine phase de croissance, nous investissons massivement dans l'IA/ML, les LLM et l'IoT industriel afin de transformer le mode de fonctionnement des équipes de première ligne, en prédisant les pannes avant qu'elles ne se produisent, en automatisant les flux de travail et en intégrant l'intelligence dans chaque actif et chaque procédure.
Ce que vous ferez:
À propos de vous:
Atouts:
Quels sont les avantages pour vous?:
Qui sommes-nous:
Ready to apply?
Apply to MaintainX
Cresta is on a mission to turn every customer conversation into a competitive advantage by unlocking the true potential of the contact center. Our platform combines the best of AI and human intelligence to help contact centers discover customer insights and behavioral best practices, automate conversations and inefficient processes, and empower every team member to work smarter and faster. Born from the prestigious Stanford AI lab, Cresta's co-founder and chairman is Sebastian Thrun, the genius behind Google X, Waymo, Udacity, and more. Our leadership also includes CEO, Ping Wu, the co-founder of Google Contact Center AI and Vertex AI platform, and co-founder, Tim Shi, an early member of Open AI.
Join us on this thrilling journey to revolutionize the workforce with AI. The future of work is here, and it's at Cresta.
At Cresta, the Knowledge Assist (KA) team develops AI solutions for the contact center industry, focusing on improving agent productivity by providing access to the right knowledge at the right time.
Our current projects:
Our internships offer a dynamic, fast-paced environment where you’ll collaborate with top researchers and engineers in the field. We provide opportunities for interns to make significant contributions to AI research and apply novel techniques at scale.
This is a unique opportunity to shape the future of AI at Cresta by solving complex problems and bringing breakthrough AI advancements into production environments.
Responsibilities:
Perks & Benefits:
Compensation for this position includes a base salary, equity, and a variety of benefits. Actual base salaries will be based on candidate-specific factors, including experience, skillset, and location, and local minimum pay requirements as applicable. We are actively hiring for this role in the US and Canada. Your recruiter can provide further details.
This posting will be used to fill a newly-created role.
We have noticed a rise in recruiting impersonations across the industry, where scammers attempt to access candidates' personal and financial information through fake interviews and offers. All Cresta recruiting email communications will always come from the @cresta.ai domain. Any outreach claiming to be from Cresta via other sources should be ignored. If you are uncertain whether you have been contacted by an official Cresta employee, reach out to recruiting@cresta.ai
Ready to apply?
Apply to CrestaShare this job
Cerebras Systems builds the world's largest AI chip, 56 times larger than GPUs. Our novel wafer-scale architecture provides the AI compute power of dozens of GPUs on a single chip, with the programming simplicity of a single device. This approach allows Cerebras to deliver industry-leading training and inference speeds and empowers machine learning users to effortlessly run large-scale ML applications, without the hassle of managing hundreds of GPUs or TPUs.
Cerebras' current customers include top model labs, global enterprises, and cutting-edge AI-native startups. OpenAI recently announced a multi-year partnership with Cerebras, to deploy 750 megawatts of scale, transforming key workloads with ultra high-speed inference.
Thanks to the groundbreaking wafer-scale architecture, Cerebras Inference offers the fastest Generative AI inference solution in the world, over 10 times faster than GPU-based hyperscale cloud inference services. This order of magnitude increase in speed is transforming the user experience of AI applications, unlocking real-time iteration and increasing intelligence via additional agentic computation.
About the Role
We are seeking a versatile and experienced engineer to join our SOTA Training Platform team. This team is responsible to rapidly bring up state-of-the-art open-source models (like LLaMA, Qwen, etc) or customer-provided proprietary models on our Cerebras CSX systems. Success in this role requires a system-minded generalist who thrives in fast-paced bringup environments and is comfortable working across the entire Cerebras software stack.
Your work will play a critical role in achieving unprecedented levels of performance, efficiency, and scalability for AI applications.
People who are serious about software make their own hardware. At Cerebras we have built a breakthrough architecture that is unlocking new opportunities for the AI industry. With dozens of model releases and rapid growth, we’ve reached an inflection point in our business. Members of our team tell us there are five main reasons they joined Cerebras:
Read our blog: Five Reasons to Join Cerebras in 2026.
Cerebras Systems is committed to creating an equal and diverse environment and is proud to be an equal opportunity employer. We celebrate different backgrounds, perspectives, and skills. We believe inclusive teams build better products and companies. We try every day to build a work environment that empowers people to do their best work through continuous learning, growth and support of those around them.
This website or its third-party tools process personal data. For more details, click here to review our CCPA disclosure notice.
Ready to apply?
Apply to Cerebras Systems
Share this job
Cerebras Systems builds the world's largest AI chip, 56 times larger than GPUs. Our novel wafer-scale architecture provides the AI compute power of dozens of GPUs on a single chip, with the programming simplicity of a single device. This approach allows Cerebras to deliver industry-leading training and inference speeds and empowers machine learning users to effortlessly run large-scale ML applications, without the hassle of managing hundreds of GPUs or TPUs.
Cerebras' current customers include top model labs, global enterprises, and cutting-edge AI-native startups. OpenAI recently announced a multi-year partnership with Cerebras, to deploy 750 megawatts of scale, transforming key workloads with ultra high-speed inference.
Thanks to the groundbreaking wafer-scale architecture, Cerebras Inference offers the fastest Generative AI inference solution in the world, over 10 times faster than GPU-based hyperscale cloud inference services. This order of magnitude increase in speed is transforming the user experience of AI applications, unlocking real-time iteration and increasing intelligence via additional agentic computation.
About The Role
The Inference ML Engineering team at Cerebras Systems is dedicated to enabling our fast generative inference solution through simple APIs powered by a distributed runtime that runs on large clusters of our own hardware. Our mission is to empower enterprises, developers, and researchers to unlock the full potential of our platform, leveraging its performance, scalability, and flexibility. The team works closely with cross-functional groups, including compiler developers, cluster orchestrators, ML scientists, cloud architects, and product teams, to deliver high-impact solutions that redefine the boundaries of ML performance and usability.
As a Senior Software Engineer on the Inference ML Engineering team, you will play a key role in designing and implementing APIs, ML features, and tools that enable running state-of-the-art generative AI models on our custom hardware. You will architect solutions that enable seamless model translation and execution, ensuring high throughput and low latency, while maintaining ease of use. Your responsibilities will include leading technical initiatives, collaborating with other engineering teams to enhance the developer experience, enabling key ML features at scale, maintaining our speed advantage, achieving high throughput, and supporting a wide range of ML workloads. This role offers an opportunity to shape the evolution of our ML ecosystem while tackling complex technical challenges at the intersection of machine learning, software, and hardware.
Responsibilities
Skills and Qualifications
People who are serious about software make their own hardware. At Cerebras we have built a breakthrough architecture that is unlocking new opportunities for the AI industry. With dozens of model releases and rapid growth, we’ve reached an inflection point in our business. Members of our team tell us there are five main reasons they joined Cerebras:
Read our blog: Five Reasons to Join Cerebras in 2026.
Cerebras Systems is committed to creating an equal and diverse environment and is proud to be an equal opportunity employer. We celebrate different backgrounds, perspectives, and skills. We believe inclusive teams build better products and companies. We try every day to build a work environment that empowers people to do their best work through continuous learning, growth and support of those around them.
This website or its third-party tools process personal data. For more details, click here to review our CCPA disclosure notice.
Ready to apply?
Apply to Cerebras Systems
Share this job
Cerebras Systems builds the world's largest AI chip, 56 times larger than GPUs. Our novel wafer-scale architecture provides the AI compute power of dozens of GPUs on a single chip, with the programming simplicity of a single device. This approach allows Cerebras to deliver industry-leading training and inference speeds and empowers machine learning users to effortlessly run large-scale ML applications, without the hassle of managing hundreds of GPUs or TPUs.
Cerebras' current customers include top model labs, global enterprises, and cutting-edge AI-native startups. OpenAI recently announced a multi-year partnership with Cerebras, to deploy 750 megawatts of scale, transforming key workloads with ultra high-speed inference.
Thanks to the groundbreaking wafer-scale architecture, Cerebras Inference offers the fastest Generative AI inference solution in the world, over 10 times faster than GPU-based hyperscale cloud inference services. This order of magnitude increase in speed is transforming the user experience of AI applications, unlocking real-time iteration and increasing intelligence via additional agentic computation.
People who are serious about software make their own hardware. At Cerebras we have built a breakthrough architecture that is unlocking new opportunities for the AI industry. With dozens of model releases and rapid growth, we’ve reached an inflection point in our business. Members of our team tell us there are five main reasons they joined Cerebras:
Read our blog: Five Reasons to Join Cerebras in 2026.
Cerebras Systems is committed to creating an equal and diverse environment and is proud to be an equal opportunity employer. We celebrate different backgrounds, perspectives, and skills. We believe inclusive teams build better products and companies. We try every day to build a work environment that empowers people to do their best work through continuous learning, growth and support of those around them.
This website or its third-party tools process personal data. For more details, click here to review our CCPA disclosure notice.
Ready to apply?
Apply to Cerebras Systems
Share this job
Cerebras Systems builds the world's largest AI chip, 56 times larger than GPUs. Our novel wafer-scale architecture provides the AI compute power of dozens of GPUs on a single chip, with the programming simplicity of a single device. This approach allows Cerebras to deliver industry-leading training and inference speeds and empowers machine learning users to effortlessly run large-scale ML applications, without the hassle of managing hundreds of GPUs or TPUs.
Cerebras' current customers include top model labs, global enterprises, and cutting-edge AI-native startups. OpenAI recently announced a multi-year partnership with Cerebras, to deploy 750 megawatts of scale, transforming key workloads with ultra high-speed inference.
Thanks to the groundbreaking wafer-scale architecture, Cerebras Inference offers the fastest Generative AI inference solution in the world, over 10 times faster than GPU-based hyperscale cloud inference services. This order of magnitude increase in speed is transforming the user experience of AI applications, unlocking real-time iteration and increasing intelligence via additional agentic computation.
We are building the next generation of large-scale AI systems that power training and inference workloads at unprecedented scale and efficiency.
You will design and develop high-performance distributed software that orchestrates massive compute and data pipelines across heterogeneous clusters. Your work will push the limits of concurrency, throughput, and scalability—enabling efficient execution of models at massive scale. This role sits at the intersection of systems engineering and machine learning performance, demanding both architectural depth and low-level implementation skills. You will help shape how models are executed and optimized end-to-end, from data ingestion to distributed execution, across cutting-edge hardware platforms.
We’re hiring for runtime roles across both Training and Inference.
People who are serious about software make their own hardware. At Cerebras we have built a breakthrough architecture that is unlocking new opportunities for the AI industry. With dozens of model releases and rapid growth, we’ve reached an inflection point in our business. Members of our team tell us there are five main reasons they joined Cerebras:
Read our blog: Five Reasons to Join Cerebras in 2026.
Cerebras Systems is committed to creating an equal and diverse environment and is proud to be an equal opportunity employer. We celebrate different backgrounds, perspectives, and skills. We believe inclusive teams build better products and companies. We try every day to build a work environment that empowers people to do their best work through continuous learning, growth and support of those around them.
This website or its third-party tools process personal data. For more details, click here to review our CCPA disclosure notice.
Ready to apply?
Apply to Cerebras Systems
Share this job
Cerebras Systems builds the world's largest AI chip, 56 times larger than GPUs. Our novel wafer-scale architecture provides the AI compute power of dozens of GPUs on a single chip, with the programming simplicity of a single device. This approach allows Cerebras to deliver industry-leading training and inference speeds and empowers machine learning users to effortlessly run large-scale ML applications, without the hassle of managing hundreds of GPUs or TPUs.
Cerebras' current customers include top model labs, global enterprises, and cutting-edge AI-native startups. OpenAI recently announced a multi-year partnership with Cerebras, to deploy 750 megawatts of scale, transforming key workloads with ultra high-speed inference.
Thanks to the groundbreaking wafer-scale architecture, Cerebras Inference offers the fastest Generative AI inference solution in the world, over 10 times faster than GPU-based hyperscale cloud inference services. This order of magnitude increase in speed is transforming the user experience of AI applications, unlocking real-time iteration and increasing intelligence via additional agentic computation.
Cerebras builds wafer-scale AI processors—single chips delivering tens of PB/s of memory bandwidth and a dataflow architecture that accelerates at a granularity no multi-device system can match. The Advanced Technology Group (ATG) is Cerebras’ pathfinding organization. We work ahead of product to explore new architectures, demonstrate breakthrough performance on scientific and AI workloads, and shape the technical roadmap for future Cerebras hardware and software. Our work regularly appears at top-tier venues (Supercomputing, SIAM, IEEE, and NeurIPS) and directly influences the design of next-generation wafer-scale systems.
Most AI research today is shaped by the constraints of existing hardware. This role starts from the other direction: what would you build if the architecture let you rethink the fundamentals? You will design and develop AI models and training methodologies on wafer-scale hardware, working at the level of optimization theory, model architecture, and statistical foundations rather than assembling existing components.
The ATG sits at the intersection of AI, computational science, and computer architecture, and your work will draw on all three. You will collaborate closely with Cerebras’ ASIC, compiler, kernel, and AI teams as well as external partners at universities and national laboratories.
We are hiring for multiple positions across experience levels. If this work resonates, we encourage you to apply.
People who are serious about software make their own hardware. At Cerebras we have built a breakthrough architecture that is unlocking new opportunities for the AI industry. With dozens of model releases and rapid growth, we’ve reached an inflection point in our business. Members of our team tell us there are five main reasons they joined Cerebras:
Read our blog: Five Reasons to Join Cerebras in 2026.
Cerebras Systems is committed to creating an equal and diverse environment and is proud to be an equal opportunity employer. We celebrate different backgrounds, perspectives, and skills. We believe inclusive teams build better products and companies. We try every day to build a work environment that empowers people to do their best work through continuous learning, growth and support of those around them.
This website or its third-party tools process personal data. For more details, click here to review our CCPA disclosure notice.
Ready to apply?
Apply to Cerebras Systems
Share this job
Cerebras Systems builds the world's largest AI chip, 56 times larger than GPUs. Our novel wafer-scale architecture provides the AI compute power of dozens of GPUs on a single chip, with the programming simplicity of a single device. This approach allows Cerebras to deliver industry-leading training and inference speeds and empowers machine learning users to effortlessly run large-scale ML applications, without the hassle of managing hundreds of GPUs or TPUs.
Cerebras' current customers include top model labs, global enterprises, and cutting-edge AI-native startups. OpenAI recently announced a multi-year partnership with Cerebras, to deploy 750 megawatts of scale, transforming key workloads with ultra high-speed inference.
Thanks to the groundbreaking wafer-scale architecture, Cerebras Inference offers the fastest Generative AI inference solution in the world, over 10 times faster than GPU-based hyperscale cloud inference services. This order of magnitude increase in speed is transforming the user experience of AI applications, unlocking real-time iteration and increasing intelligence via additional agentic computation.
As an Applied Machine Learning Research Scientist at Cerebras, you will play a key role in turning modern machine learning techniques into scalable, high-performance systems. This role sits at the intersection of modeling and systems focused not on publishing new algorithms, but on understanding how they work and making them run effectively at scale. Your work will directly impact how large language models (LLMs) are trained, optimized, and deployed on one of the most advanced AI platforms in the world.
You will work closely with researchers and senior engineers to implement and improve workflows for LLM pretraining, fine-tuning, and reinforcement learning-based post-training. This includes building training pipelines, debugging complex system behaviors, improving model quality, and iterating on data and evaluation strategies. Your contributions will help translate cutting-edge ML ideas into reliable, production-ready systems that solve real-world problems.
This role is ideal for candidates who enjoy hands-on engineering, want to build deep intuition for ML systems, and are excited about working on LLMs and reinforcement learning in practice, not just in theory.
Responsibilities
Skills & Qualifications
Preferred Skills & Qualifications
People who are serious about software make their own hardware. At Cerebras we have built a breakthrough architecture that is unlocking new opportunities for the AI industry. With dozens of model releases and rapid growth, we’ve reached an inflection point in our business. Members of our team tell us there are five main reasons they joined Cerebras:
Read our blog: Five Reasons to Join Cerebras in 2026.
Cerebras Systems is committed to creating an equal and diverse environment and is proud to be an equal opportunity employer. We celebrate different backgrounds, perspectives, and skills. We believe inclusive teams build better products and companies. We try every day to build a work environment that empowers people to do their best work through continuous learning, growth and support of those around them.
This website or its third-party tools process personal data. For more details, click here to review our CCPA disclosure notice.
Ready to apply?
Apply to Cerebras Systems
Share this job
Cerebras Systems builds the world's largest AI chip, 56 times larger than GPUs. Our novel wafer-scale architecture provides the AI compute power of dozens of GPUs on a single chip, with the programming simplicity of a single device. This approach allows Cerebras to deliver industry-leading training and inference speeds and empowers machine learning users to effortlessly run large-scale ML applications, without the hassle of managing hundreds of GPUs or TPUs.
Cerebras' current customers include top model labs, global enterprises, and cutting-edge AI-native startups. OpenAI recently announced a multi-year partnership with Cerebras, to deploy 750 megawatts of scale, transforming key workloads with ultra high-speed inference.
Thanks to the groundbreaking wafer-scale architecture, Cerebras Inference offers the fastest Generative AI inference solution in the world, over 10 times faster than GPU-based hyperscale cloud inference services. This order of magnitude increase in speed is transforming the user experience of AI applications, unlocking real-time iteration and increasing intelligence via additional agentic computation.
The Inference ML Engineering team at Cerebras builds the runtime, APIs, and systems that power the fastest generative AI inference platform in the world.
As an Engineering Manager, Inference ML Runtime, you will lead a team responsible for designing and scaling the systems that enable seamless execution of state-of-the-art AI models on Cerebras hardware. You will operate at the intersection of machine learning, distributed systems, and high-performance runtime engineering, translating cutting-edge research into production-ready infrastructure to serve a variety of text-only and multimodal models.
This role combines technical leadership, people management, and execution ownership, with direct impact on Cerebras’ core inference platform.
Technical Leadership
Team Leadership
Execution & Delivery
Platform & Performance Ownership
Cross-Functional Collaboration
Required
Preferred
Why This Role Matters
This team is central to Cerebras’ mission of delivering the fastest AI inference in the world. Your work will directly enable real-time AI applications and unlock new capabilities across enterprise and frontier AI use cases.
People who are serious about software make their own hardware. At Cerebras we have built a breakthrough architecture that is unlocking new opportunities for the AI industry. With dozens of model releases and rapid growth, we’ve reached an inflection point in our business. Members of our team tell us there are five main reasons they joined Cerebras:
Read our blog: Five Reasons to Join Cerebras in 2026.
Cerebras Systems is committed to creating an equal and diverse environment and is proud to be an equal opportunity employer. We celebrate different backgrounds, perspectives, and skills. We believe inclusive teams build better products and companies. We try every day to build a work environment that empowers people to do their best work through continuous learning, growth and support of those around them.
This website or its third-party tools process personal data. For more details, click here to review our CCPA disclosure notice.
Ready to apply?
Apply to Cerebras Systems
Share this job
Cerebras Systems builds the world's largest AI chip, 56 times larger than GPUs. Our novel wafer-scale architecture provides the AI compute power of dozens of GPUs on a single chip, with the programming simplicity of a single device. This approach allows Cerebras to deliver industry-leading training and inference speeds and empowers machine learning users to effortlessly run large-scale ML applications, without the hassle of managing hundreds of GPUs or TPUs.
Cerebras' current customers include top model labs, global enterprises, and cutting-edge AI-native startups. OpenAI recently announced a multi-year partnership with Cerebras, to deploy 750 megawatts of scale, transforming key workloads with ultra high-speed inference.
Thanks to the groundbreaking wafer-scale architecture, Cerebras Inference offers the fastest Generative AI inference solution in the world, over 10 times faster than GPU-based hyperscale cloud inference services. This order of magnitude increase in speed is transforming the user experience of AI applications, unlocking real-time iteration and increasing intelligence via additional agentic computation.
About the Role
We are seeking a versatile and experienced engineer to join our Inference Core Model Bringup team. This team is responsible to rapidly bring up state-of-the-art open-source models (like LLaMA, Qwen, etc) or customer-provided proprietary models on our Cerebras CSX systems. Success in this role requires a system-minded generalist who thrives in fast-paced bringup environments and is comfortable working across the entire Cerebras software stack.
Your work will play a critical role in achieving unprecedented levels of performance, efficiency, and scalability for AI applications.
People who are serious about software make their own hardware. At Cerebras we have built a breakthrough architecture that is unlocking new opportunities for the AI industry. With dozens of model releases and rapid growth, we’ve reached an inflection point in our business. Members of our team tell us there are five main reasons they joined Cerebras:
Read our blog: Five Reasons to Join Cerebras in 2026.
Cerebras Systems is committed to creating an equal and diverse environment and is proud to be an equal opportunity employer. We celebrate different backgrounds, perspectives, and skills. We believe inclusive teams build better products and companies. We try every day to build a work environment that empowers people to do their best work through continuous learning, growth and support of those around them.
This website or its third-party tools process personal data. For more details, click here to review our CCPA disclosure notice.
Ready to apply?
Apply to Cerebras Systems
Share this job
Cerebras Systems builds the world's largest AI chip, 56 times larger than GPUs. Our novel wafer-scale architecture provides the AI compute power of dozens of GPUs on a single chip, with the programming simplicity of a single device. This approach allows Cerebras to deliver industry-leading training and inference speeds and empowers machine learning users to effortlessly run large-scale ML applications, without the hassle of managing hundreds of GPUs or TPUs.
Cerebras' current customers include top model labs, global enterprises, and cutting-edge AI-native startups. OpenAI recently announced a multi-year partnership with Cerebras, to deploy 750 megawatts of scale, transforming key workloads with ultra high-speed inference.
Thanks to the groundbreaking wafer-scale architecture, Cerebras Inference offers the fastest Generative AI inference solution in the world, over 10 times faster than GPU-based hyperscale cloud inference services. This order of magnitude increase in speed is transforming the user experience of AI applications, unlocking real-time iteration and increasing intelligence via additional agentic computation.
As a Kernel Engineer on our team, you will develop high-performance software solutions at the intersection of hardware and software, developing high-performance software for cutting-edge AI and HPC workloads. Your focus will be on implementing, optimizing, and scaling deep learning operations to fully leverage our custom, massively parallel processor architecture.
You will be part of a world-class team responsible for the design, performance tuning, and validation of foundational ML and HPC kernels. This includes building a library of parallel and distributed algorithms that maximize compute utilization and push the boundaries of training efficiency for state-of-the-art AI models. Your work will be critical to unlocking the full potential of our hardware and accelerating the pace of AI innovation.
Responsibilities
Skills And Qualifications
Preferred Skills And Qualifications
People who are serious about software make their own hardware. At Cerebras we have built a breakthrough architecture that is unlocking new opportunities for the AI industry. With dozens of model releases and rapid growth, we’ve reached an inflection point in our business. Members of our team tell us there are five main reasons they joined Cerebras:
Read our blog: Five Reasons to Join Cerebras in 2026.
Cerebras Systems is committed to creating an equal and diverse environment and is proud to be an equal opportunity employer. We celebrate different backgrounds, perspectives, and skills. We believe inclusive teams build better products and companies. We try every day to build a work environment that empowers people to do their best work through continuous learning, growth and support of those around them.
This website or its third-party tools process personal data. For more details, click here to review our CCPA disclosure notice.
Ready to apply?
Apply to Cerebras Systems
About Faire
Faire is a technology wholesale platform built on the belief that the future is local. Independent retailers around the globe collectively represent a multi-hundred-billion-dollar wholesale market that has historically been fragmented and offline. At Faire, we're using the power of tech, data, and machine learning to connect this thriving community of entrepreneurs across the globe. Picture your favorite boutique in town — we help them discover the best products from around the world to sell in their stores. With the right tools and insights, we believe that we can level the playing field so businesses can grow and local communities can thrive.
We’re looking for smart, resourceful and passionate people to join us as we power the shop local movement. If you believe in community, come join ours.
About this Role
The Ads Data team is building the next generation of advertising products for the wholesale industry. As a key member of this team, you’ll help drive the ML algorithm strategy and system design behind one of the most critical levers for customer value and company growth—Search Ads. You’ll lead the advancement of real-time systems that decide which ads to show for a query, where to place them, and how to optimize for relevance, marketplace health, and advertiser outcomes. This role mirrors many of the technical expectations of Faire’s organic Search roles (modern NLP/LLMs, query understanding, real-time ranking), while operating in an ads environment with auctions, budgets, and pacing constraints.
You’ll operate at the forefront of algorithms—combining large language models, natural language processing, query understanding, deep learning, and structured behavioral data to deliver highly relevant sponsored results for any given query.
What You'll Do
Qualifications
Great to Haves
Salary Range
San Francisco: the pay range for this role is $196,000 to $269,500 per year.
This role will also be eligible for equity and benefits. Actual base pay will be determined based on permissible factors such as transferable skills, work experience, market demands, and primary work location. The base pay range provided is subject to change and may be modified in the future.
Hybrid Faire employees currently go into the office 3 days per week on Tuesdays, Thursdays, and a third flex day of their choosing (Monday, Wednesday, or Friday). Additionally, hybrid in-office roles will have the flexibility to work remotely up to 4 weeks per year. Specific Workplace and Information Technology positions may require onsite attendance 5 days per week as will be indicated in the job posting.
Why you’ll love working at Faire
Faire was founded in 2017 by a team of early product and engineering leads from Square. We’re backed by some of the top investors in retail and tech including: Y Combinator, Lightspeed Venture Partners, Forerunner Ventures, Khosla Ventures, Sequoia Capital, Founders Fund, and DST Global. We have headquarters in San Francisco and Kitchener-Waterloo, and a global employee presence across offices in Toronto, London, and New York. To learn more about Faire and our customers, you can read more on our blog.
Faire provides equal employment opportunities (EEO) to all employees and applicants for employment without regard to race, color, religion, sex, national origin, age, disability, genetics, sexual orientation, gender identity or gender expression.
Faire is committed to providing access, equal opportunity and reasonable accommodation for individuals with disabilities in employment, its services, programs, and activities. Accommodations are available throughout the recruitment process and applicants with a disability may request to be accommodated throughout the recruitment process. We will work with all applicants to accommodate their individual accessibility needs. To request reasonable accommodation, please fill out our Accommodation Request Form (https://bit.ly/faire-form)
Privacy
For information about the type of personal data Faire collects from applicants, as well as your choices regarding the data collected about you, please visit Faire’s Privacy Notice (https://www.faire.com/privacy)
Ready to apply?
Apply to Faire
About Faire
Faire is a technology wholesale platform built on the belief that the future is local. Independent retailers around the globe collectively represent a multi-hundred-billion-dollar wholesale market that has historically been fragmented and offline. At Faire, we're using the power of tech, data, and machine learning to connect this thriving community of entrepreneurs across the globe. Picture your favorite boutique in town — we help them discover the best products from around the world to sell in their stores. With the right tools and insights, we believe that we can level the playing field so businesses can grow and local communities can thrive.
We’re looking for smart, resourceful and passionate people to join us as we power the shop local movement. If you believe in community, come join ours.
About this Role
The Ads Data team is building the next generation of advertising products for the wholesale industry. As a key member of this team, you’ll help drive the ML algorithm strategy and system design behind one of the most critical levers for customer value and company growth—Search Ads. You’ll lead the advancement of real-time systems that decide which ads to show for a query, where to place them, and how to optimize for relevance, marketplace health, and advertiser outcomes. This role mirrors many of the technical expectations of Faire’s organic Search roles (modern NLP/LLMs, query understanding, real-time ranking), while operating in an ads environment with auctions, budgets, and pacing constraints.
You’ll operate at the forefront of algorithms—combining large language models, natural language processing, query understanding, deep learning, and structured behavioral data to deliver highly relevant sponsored results for any given query.
What You'll Do
Qualifications
Great to Haves
Salary Range
San Francisco: the pay range for this role is $165,500 to $227,500 per year.
This role will also be eligible for equity and benefits. Actual base pay will be determined based on permissible factors such as transferable skills, work experience, market demands, and primary work location. The base pay range provided is subject to change and may be modified in the future.
Hybrid Faire employees currently go into the office 3 days per week on Tuesdays, Thursdays, and a third flex day of their choosing (Monday, Wednesday, or Friday). Additionally, hybrid in-office roles will have the flexibility to work remotely up to 4 weeks per year. Specific Workplace and Information Technology positions may require onsite attendance 5 days per week as will be indicated in the job posting.
Why you’ll love working at Faire
Faire was founded in 2017 by a team of early product and engineering leads from Square. We’re backed by some of the top investors in retail and tech including: Y Combinator, Lightspeed Venture Partners, Forerunner Ventures, Khosla Ventures, Sequoia Capital, Founders Fund, and DST Global. We have headquarters in San Francisco and Kitchener-Waterloo, and a global employee presence across offices in Toronto, London, and New York. To learn more about Faire and our customers, you can read more on our blog.
Faire provides equal employment opportunities (EEO) to all employees and applicants for employment without regard to race, color, religion, sex, national origin, age, disability, genetics, sexual orientation, gender identity or gender expression.
Faire is committed to providing access, equal opportunity and reasonable accommodation for individuals with disabilities in employment, its services, programs, and activities. Accommodations are available throughout the recruitment process and applicants with a disability may request to be accommodated throughout the recruitment process. We will work with all applicants to accommodate their individual accessibility needs. To request reasonable accommodation, please fill out our Accommodation Request Form (https://bit.ly/faire-form)
Privacy
For information about the type of personal data Faire collects from applicants, as well as your choices regarding the data collected about you, please visit Faire’s Privacy Notice (https://www.faire.com/privacy)
Ready to apply?
Apply to Faire
Share this job
Affinity stitches together billions of data points from massive datasets to create a powerful, accurate representation of the world's professional relationship graph. Based on this data, we offer our users the insights and visibility they need to nurture and tap into the opportunities in their team's network.
This role is part of the AI Platform team, which owns the AI services that power Affinity's industry-leading relationship intelligence platform. We extract and retrieve information from billions of structured and unstructured data points to deliver actionable insights to customers.
As a Senior Machine Learning Engineer, you will collaborate with data engineers, software engineers, and product managers to shape the future of private capital's leading CRM platform. You will design and build AI systems that efficiently uncover insights from compelling business interaction data – an exciting and unique opportunity within the industry.
This is an applied machine learning position with a strong emphasis on engineering, rather than research. You will play a key role in advancing our ML Engineering capabilities, particularly in information retrieval and eventually recommendation systems.
What you’ll be doing:
Qualifications
Don’t meet every single requirement? Studies have shown that women and people of color are less likely to apply to jobs unless they meet every qualification. At Affinity, we are dedicated to building a diverse, inclusive, and authentic workplace, so if you’re excited about this role, but your past experience doesn’t perfectly align with the qualifications above, we encourage you to apply anyways. You may be just the right candidate for this or other roles.
Required:
Nice to Have:
Tech stack: Our ML pipeline manages multiple Python services that support various AI features, including utilizing OCR to extract information from unstructured data, serving embedding models to vectorize chunks, and ranking a list of recommendations based on relevance and user preference.
How we work:
Our culture is a key part of how we operate, as well as our hiring process:
If you’d want to learn more about our values click here.
What you'll enjoy at Affinity:
Please note that the role compensation details below reflect the base salary only and do not include any equity or benefits. This represents the salary range that Affinity believes, in good faith, at the time of this posting, that it will pay for the posted job.
A reasonable estimate of the current range is $160,000 to $220,000 CAD. Within the range, individual pay depends on various factors including geographical location and review of experience, knowledge, skills, abilities of the applicant.
At Affinity, we may use artificial intelligence (AI) tools as part of our recruitment process to help screen and evaluate candidate applications. While AI assists our hiring team in managing applications efficiently, it never replaces decisions made by real people. We are committed to fair and transparent hiring practices, and our AI tools are regularly monitored to ensure they support—not replace—human judgment.
Additional Information:
About Affinity
With more than 3,000 customers worldwide and backed by some of Silicon Valley's best firms, Affinity has raised $120M to empower dealmakers to find, manage, and close more deals. How? Our Relationship Intelligence platform uses the wealth of data exhaust from trillions of interactions between Investment Bankers, Venture Capitalists, Consultants, and other strategic dealmakers to deliver automated relationship insights that drive over 450,000 deals every month. We are are proud to have received Inc. and Fortune Best Workplaces awards as well as to be Great Places to Work certified for the last 5 years running. Join us on our mission to make it possible for anyone to cultivate and fully harness their network to succeed.
We use E-Verify
Our company uses E-Verify to confirm the employment eligibility of all newly hired employees. To learn more about E-Verify, including your rights and responsibilities, please visit www.dhs.gov/E-Verify.
Ready to apply?
Apply to Affinity.co
About Faire
Faire is a technology wholesale platform built on the belief that the future is local. Independent retailers around the globe collectively represent a multi-hundred-billion-dollar wholesale market that has historically been fragmented and offline. At Faire, we're using the power of tech, data, and machine learning to connect this thriving community of entrepreneurs across the globe. Picture your favorite boutique in town — we help them discover the best products from around the world to sell in their stores. With the right tools and insights, we believe that we can level the playing field so businesses can grow and local communities can thrive.
We’re looking for smart, resourceful and passionate people to join us as we power the shop local movement. If you believe in community, come join ours.
About this role
As a Senior Applied AI/ML Scientist on the Search ranking team, you will help shape the technical vision, machine-learning algorithm strategy, and system design behind one of our most important growth levers: Search (the primary tool used by customers on any e-commerce site). You will advance real-time search and recommendation systems that power next-generation shopping experiences.
You’ll work at the frontier of algorithms, combining query understanding, deep learning, transformer-based sequential modeling, graph neural networks, and structured behavioral data to return hyper-relevant, personalized products and brands for every user query.
This is a rare chance to influence the end-to-end personalized discovery experience at Faire within a high-scale, deeply multi-modal environment, while collaborating closely with a talented team of scientists and engineers.
What you'll do
You're a great fit if you have...
Bonus points for...
Salary Range
California: the pay range for this role is $192,000 to $264,000 per year.
This role will also be eligible for equity and benefits. Actual base pay will be determined based on permissible factors such as transferable skills, work experience, market demands, and primary work location. The base pay range provided is subject to change and may be modified in the future.
Hybrid Faire employees currently go into the office 3 days per week on Tuesdays, Thursdays, and a third flex day of their choosing (Monday, Wednesday, or Friday). Additionally, hybrid in-office roles will have the flexibility to work remotely up to 4 weeks per year. Specific Workplace and Information Technology positions may require onsite attendance 5 days per week as will be indicated in the job posting.
Why you’ll love working at Faire
Faire was founded in 2017 by a team of early product and engineering leads from Square. We’re backed by some of the top investors in retail and tech including: Y Combinator, Lightspeed Venture Partners, Forerunner Ventures, Khosla Ventures, Sequoia Capital, Founders Fund, and DST Global. We have headquarters in San Francisco and Kitchener-Waterloo, and a global employee presence across offices in Toronto, London, and New York. To learn more about Faire and our customers, you can read more on our blog.
Faire provides equal employment opportunities (EEO) to all employees and applicants for employment without regard to race, color, religion, sex, national origin, age, disability, genetics, sexual orientation, gender identity or gender expression.
Faire is committed to providing access, equal opportunity and reasonable accommodation for individuals with disabilities in employment, its services, programs, and activities. Accommodations are available throughout the recruitment process and applicants with a disability may request to be accommodated throughout the recruitment process. We will work with all applicants to accommodate their individual accessibility needs. To request reasonable accommodation, please fill out our Accommodation Request Form (https://bit.ly/faire-form)
Privacy
For information about the type of personal data Faire collects from applicants, as well as your choices regarding the data collected about you, please visit Faire’s Privacy Notice (https://www.faire.com/privacy)
Ready to apply?
Apply to Faire
Share this job
We are an innovative AI startup focused on transforming professional services through cutting-edge Generative AI and deep domain expertise. Our agent-driven solutions automate complex workflows, engaging humans only when needed to maximize efficiency and accuracy. Join us at the forefront of AI innovation, where your expertise will directly shape the future of professional services.
Ready to apply?
Apply to Evolver
Share this job
At HeyGen, our mission is to make visual storytelling accessible to all. Over the last decade, visual content has become the preferred method of information creation, consumption, and retention. But the ability to create such content, in particular videos, continues to be costly and challenging to scale. Our ambition is to build technology that equips more people with the power to reach, captivate, and inspire audiences.
Learn more at www.heygen.com. Visit our Mission and Culture doc here.
We are seeking a seasoned Technical Leader to build and scale the foundational compute infrastructure that powers our state-of-the-art AI models—from multimodal training data pipelines to high-throughput, low-latency video generation.
You will be the core engineer responsible for building the robust, efficient, and scalable platform that enables our research and production teams to rapidly iterate on HeyGen's generative video models. Your contributions will directly impact model performance, developer productivity, and the final quality of every AI-generated video.
Optimize GPU Utilization: Design and implement mechanisms to aggressively optimize GPU and cluster utilization across thousands of devices for inference, training, data processing and large-scale deployment of our state-of-art video generation models.
Develop Large-Scale AI Job Framework: Build highly scalable, reliable frameworks for launching and managing massive, heterogeneous compute jobs, including multi-modal high-volume data ingestion/processing, distributed model training, and continuous evaluation/benchmarking.
Enhance Observability: Develop world-class observability, tracing, and visualization tools for our compute cluster to ensure reliability, diagnose performance bottlenecks (e.g., memory, bandwidth, communication).
Accelerate Pipelines: Collaborate closely with AI researchers and AI engineers to integrate innovative acceleration techniques (e.g., custom CUDA kernels, distributed training libraries) into production-ready, scalable training and inference pipelines.
Infrastructure Management: Champion the adoption and optimization of modern cloud and container technologies (Kubernetes, Ray) for elastic, cost-efficient scaling of our distributed systems.
We are looking for a highly motivated engineer with deep experience operating and optimizing AI infrastructure at scale.
Bachelor's degree in Computer Science, Engineering, or a related field, or equivalent practical experience.
5+ years of full-time industry experience in large-scale MLOps, AI infrastructure, or HPC systems.
Experience with data frameworks and standards like Ray, Apache Spark, LanceDB
Strong proficiency in Python and a high-performance language such as C++ for developing core infrastructure components.
Deep understanding and hands-on experience with modern orchestration and distributed computing frameworks such as Kubernetes and Ray.
Experience with core ML frameworks such as PyTorch, TensorFlow, or JAX.
Master's or PhD in Computer Science or a related technical field.
Demonstrated Tech Lead experience, driving projects from conceptual design through to production deployment across cross-functional teams.
Prior experience building infrastructure specifically for Generative AI models (e.g., diffusion models, GANs, or large language models) where cost and latency are critical.
Proven background in building and operating large-scale data infrastructure (e.g., Ray, Apache Spark) to manage petabytes of multi-modal data (video, audio, text).
HeyGen is an Equal Opportunity Employer. We celebrate diversity and are committed to creating an inclusive environment for all employees.
Ready to apply?
Apply to HeyGen
Share this job
At HeyGen, our mission is to make visual storytelling accessible to all. Over the last decade, visual content has become the preferred method of information creation, consumption, and retention. But the ability to create such content, in particular videos, continues to be costly and challenging to scale. Our ambition is to build technology that equips more people with the power to reach, captivate, and inspire audiences.
Learn more at www.heygen.com. Visit our Mission and Culture doc here.
We are seeking a seasoned Software Engineer to build and scale the foundational compute infrastructure that powers our state-of-the-art AI models—from multimodal training data pipelines to high-throughput, low-latency video generation.
You will be the core engineer responsible for building the robust, efficient, and scalable platform that enables our research and production teams to rapidly iterate on HeyGen's generative video models. Your contributions will directly impact model performance, developer productivity, and the final quality of every AI-generated video.
Optimize GPU Utilization: Design and implement mechanisms to aggressively optimize GPU and cluster utilization across thousands of devices for inference, training, data processing and large-scale deployment of our state-of-art video generation models.
Develop Large-Scale AI Job Framework: Build highly scalable, reliable frameworks for launching and managing massive, heterogeneous compute jobs, including multi-modal high-volume data ingestion/processing, distributed model training, and continuous evaluation/benchmarking.
Enhance Observability: Develop world-class observability, tracing, and visualization tools for our compute cluster to ensure reliability, diagnose performance bottlenecks (e.g., memory, bandwidth, communication).
Accelerate Pipelines: Collaborate closely with AI researchers and AI engineers to integrate innovative acceleration techniques (e.g., custom CUDA kernels, distributed training libraries) into production-ready, scalable training and inference pipelines.
Infrastructure Management: Champion the adoption and optimization of modern cloud and container technologies (Kubernetes, Ray) for elastic, cost-efficient scaling of our distributed systems.
We are looking for a highly motivated engineer with deep experience operating and optimizing AI infrastructure at scale.
Bachelor's degree in Computer Science, Engineering, or a related field, or equivalent practical experience.
5+ years of full-time industry experience in large-scale MLOps, AI infrastructure, or HPC systems.
Experience with data frameworks and standards like Ray, Apache Spark, LanceDB
Strong proficiency in Python and a high-performance language such as C++ for developing core infrastructure components.
Deep understanding and hands-on experience with modern orchestration and distributed computing frameworks such as Kubernetes and Ray.
Experience with core ML frameworks such as PyTorch, TensorFlow, or JAX.
Master's or PhD in Computer Science or a related technical field.
Demonstrated Tech Lead experience, driving projects from conceptual design through to production deployment across cross-functional teams.
Prior experience building infrastructure specifically for Generative AI models (e.g., diffusion models, GANs, or large language models) where cost and latency are critical.
Proven background in building and operating large-scale data infrastructure (e.g., Ray, Apache Spark) to manage petabytes of multi-modal data (video, audio, text).
HeyGen is an Equal Opportunity Employer. We celebrate diversity and are committed to creating an inclusive environment for all employees.
Ready to apply?
Apply to HeyGen
Share this job
At HeyGen, our mission is to make visual storytelling accessible to all. Over the last decade, visual content has become the preferred method of information creation, consumption, and retention. But the ability to create such content, in particular videos, continues to be costly and challenging to scale. Our ambition is to build technology that equips more people with the power to reach, captivate, and inspire audiences.
Learn more at www.heygen.com. Visit our Mission and Culture doc here.
Position Summary:
At HeyGen, we are at the forefront of developing applications powered by our cutting-edge AI research. As a Data Infrastructure Engineer, you will lead the development of fundamental data systems and infrastructure. These systems are essential for powering our innovative applications, including Avatar IV, Photo Avatar, Instant Avatar, Interactive Avatar, and Video Translation. Your role will be crucial in enhancing the efficiency and scalability of these systems, which are vital to HeyGen's success.
Key Responsibilities:
Qualifications:
Preferred Qualifications:
What HeyGen Offers
HeyGen is an Equal Opportunity Employer. We celebrate diversity and are committed to creating an inclusive environment for all employees.
Ready to apply?
Apply to HeyGen
At HeyGen, our mission is to make visual storytelling accessible to all. Over the last decade, visual content has become the preferred method of information creation, consumption, and retention. But the ability to create such content, in particular videos, continues to be costly and challenging to scale. Our ambition is to build technology that equips more people with the power to reach, captivate, and inspire audiences.
Learn more at www.heygen.com. Visit our Mission and Culture doc here.
Driving Product Innovation with Advanced Computer Vision Technology
HeyGen is a dynamic startup at the forefront of revolutionizing video content with state-of-the-art artificial intelligence technology. With a robust track record of growth and a proven business model, we are rapidly carving out a leadership position in the AI-powered video creation industry. We are in search of passionate Research Scientists with a specialization in Computer Vision to propel our mission forward. If you are eager to deploy your advanced skills to disrupt and redefine the video creation and interaction experience, your next big opportunity awaits with us.
Key Responsibilities:
Applied Innovation and Product Development: Harness your knowledge of generative models, including GANs, Diffusion models, and other multi-modal models, to propel our video creation tools forward. Your role involves translating state-of-the-art research into innovative features that directly enhance our user experience and set our products apart.
Real-World Solutions: Channel your research capabilities to develop scalable solutions that have a significant impact on our platform, making the video creation process more intuitive, powerful, and accessible to our users.
Collaborative Engineering and Growth: Engage with a diverse team where collaborative efforts lead to extraordinary outcomes. Embrace an environment that promotes learning, mentorship, and teamwork, furthering both personal development and the advancement of our technology.
Qualifications:
What HeyGen Offers:
Ready to apply?
Apply to HeyGen
Cookies & analytics
This site uses cookies from third-party services to deliver its features and to analyze traffic.