Two markets, two realities
If you're measuring AI hiring health by what's happening at Anthropic, OpenAI, Google DeepMind, and Meta, you see a market that has cooled significantly. Hiring is selective, timelines are long, and the bar for ML research roles has never been higher.
If you're measuring by what's happening at enterprise technology companies, financial services firms, healthcare systems, and manufacturing companies, you see a different story entirely: frantic demand for people who can deploy AI reliably in production.
These are not the same market.
What enterprises are actually hiring for
The most in-demand roles in enterprise AI aren't ML researchers. They're:
- AI implementation engineers: People who can take foundation models and integrate them into existing enterprise systems, handle the data pipelines, and manage the operational complexity of deployed AI.
- AI product managers: People who can translate business requirements into AI system specifications and manage the gap between what models can do and what the business needs.
- Applied ML engineers: The hybrid role — enough ML depth to debug model failures, enough software engineering to build reliable systems.
Research roles are concentrated almost entirely in frontier labs and a handful of well-funded startups.
The compensation gap
Research scientists at top labs command compensation packages that have become disconnected from the broader market. $500K total compensation is table stakes for senior research roles. This is not replicating in enterprise.
Enterprise AI roles pay well — often $200-350K for senior positions — but organizations expecting research talent at implementation salaries are consistently disappointed.
What this means for candidates
The path optimization differs completely depending on which market you're targeting.
For frontier lab research: publications, top-tier institution credentials, demonstrated novel contributions. The bar is genuinely high and increasingly competitive.
For enterprise deployment: practical experience shipping AI systems, familiarity with MLOps tooling, ability to explain model behavior to non-technical stakeholders. Publications are largely irrelevant.
Most candidates we spoke to are hedging — trying to build a profile that works for both markets and ending up optimized for neither.