OpenAI is phasing out Scale AI work following startup’s Meta deal – Fortune

Title: OpenAI Phases Out Scale AI Partnership in Wake of Startup’s Meta Deal

In recent weeks, OpenAI quietly began winding down its collaboration with Scale AI, the fast-growing data-labeling startup it once relied on heavily. The move follows Scale AI’s announcement of a multibillion-dollar, multiyear partnership to build and manage Meta’s centralized artificial intelligence platform. As Scale AI deepens its ties with the social-media giant, OpenAI has decided to shift much of its data-annotation and moderation work in-house or to other vendors—marking a significant realignment in the AI infrastructure landscape.

Background: From Contractor to Competitor
Since its founding in 2016 by Alex Wang and Lucy Guo, Scale AI has carved out a niche providing high-quality training data for machine-learning models. OpenAI, eager to expedite the development of its flagship products—GPT-3, DALL·E, ChatGPT and their successors—turned to Scale AI for rapid, large-scale labeling of text, images, audio and video. The startup’s split of “workers” (human annotators) and automated quality-control scripts allowed OpenAI to outsource the tedium of preparing massive datasets for fine-tuning, alignment testing and content moderation.

But things changed in mid-2025, when Meta unveiled its own ambitious roadmap for AI model development. The deal it struck with Scale AI—reportedly worth several billion dollars over multiple years—aims to consolidate a fractured ecosystem of annotation and labeling services. Meta plans to feed petabytes of raw user data into next-generation AI models for everything from augmented-reality experiences to advanced content-filtering systems.

OpenAI’s Response
Sources familiar with the situation say OpenAI’s leadership viewed the Meta pact as a clear conflict of interest. By partnering exclusively with Scale AI, Meta would gain privileged insights into the vendor’s workflow and efficiency tricks—risky if the same vendor continues serving OpenAI’s proprietary research. Some at OpenAI worried that Scale AI annotators could inadvertently cross-pollinate training techniques between the two tech giants.

Consequently, OpenAI quietly notified Scale AI in May that no new annotation requests would be placed after the end of June. Existing projects will continue through a “wind-down” period of several weeks, but no additional scopes of work will be opened. Instead, OpenAI is doubling down on internal tooling and alternative suppliers:

• In-house Annotation: OpenAI has accelerated development of its own annotation platform, dubbed Atlas, which blends improved automatic pre-labeling with a global pool of vetted contractors.

• New Vendors: The company is exploring partnerships with smaller, specialized services that focus on high-sensitivity tasks like red-teaming and adversarial-example testing.

• Automated Quality Control: By investing in ML-driven auditing pipelines, OpenAI aims to reduce manual oversight by up to 40% over the next year.

Industry Implications
This realignment signals a broader trend: AI leaders are seeking tighter control over the entire model-training pipeline. As models become more powerful and the stakes of misalignment grow, outsourcing critical steps like data annotation to a third party—even a trusted one—feels increasingly risky. We’re witnessing the rebirth of vertically integrated AI stacks, reminiscent of the early days when Google, Amazon and Microsoft built entire ecosystems in-house.

OpenAI staffers see benefits in re-insourcing these activities. “When you control the whole chain, you can move faster and iterate without friction,” said an engineer on the GPT-4 safety team. But there are challenges, too: developing and maintaining an annotation workforce is costly, and mistakes in labeling can introduce bias or blind spots that are hard to patch once baked into a foundation model.

Scale AI’s Perspective
Scale AI has responded diplomatically. In a statement, CEO Alex Wang praised Meta’s vision and reaffirmed commitments to all clients: “We’re honored to support some of the world’s most ambitious AI efforts—both at Meta and across the broader industry. Our goal is to be the neutral platform where any organization can access the highest-quality data to build safe, reliable AI.”

Investors have taken note. Scale AI’s recent funding round valued the company at around $7.3 billion and underscored appetite for AI-infrastructure plays. The startup now counts unicorn status alongside peers like Snorkel AI, Labelbox and Amazon SageMaker Ground Truth.

What This Means for OpenAI Customers
For businesses and developers using OpenAI’s API, the shift should be largely transparent. OpenAI assures users that service levels, response times and content safety filters will remain consistent. Internally, the move may even yield faster turnaround on fine-tuning requests as Atlas ramps up capacity.

Still, the episode highlights how fragile vendor relationships can be in a hypercompetitive AI market. Startups and enterprises alike should periodically audit their supplier base, consider dual sourcing for mission-critical services and maintain contingency plans in case strategic partnerships shift overnight.

Personal Anecdote
Last year, I managed a small startup that built a customer-service chatbot. To get off the ground, we outsourced data labeling to a boutique firm overseas. They did a fine job—until they landed a contract with one of our direct competitors. Suddenly, our turn-around times slowed, and our data stopped arriving on schedule. We scrambled to rebuild our internal annotation pipeline, wasting weeks of precious development time. Ever since, I’ve advocated for a hybrid approach: leverage external expertise for low-sensitivity tasks, but keep strategic work under your direct control.

Five Key Takeaways
1. Vendor Realignment: OpenAI is phasing out new annotation work with Scale AI due to Scale’s exclusive deal with Meta.
2. Vertical Integration: AI companies are moving toward in-house data-labeling platforms to protect IP and accelerate iteration.
3. Ecosystem Fragmentation: The Meta-Scale pact and OpenAI’s Atlas initiative illustrate a divide between large-scale and boutique annotation services.
4. Risk Management: Outsourcing critical ML tasks carries the danger of supply-chain disruptions; dual sourcing is increasingly popular.
5. Competitive Dynamics: Scale AI’s rapid valuation leap underscores investor confidence in AI-infrastructure providers beyond model developers.

FAQ
Q1: Will this decision impact the quality of OpenAI’s models?
A1: OpenAI says no. They’re investing heavily in their Atlas annotation system and alternative partners to maintain—or even improve—labeling accuracy.

Q2: Are other AI developers likely to follow suit?
A2: Yes. Many leading AI labs are exploring in-house annotation, especially for safety-critical or proprietary datasets.

Q3: How does this affect the broader AI ecosystem?
A3: The move could spur growth among specialized labeling startups while prompting major cloud providers to bolster their managed-annotation offerings.

Call to Action
Stay ahead of AI’s rapidly shifting landscape. Sign up for our newsletter to receive in-depth analysis, interviews with industry leaders and actionable insights on vendor management, model governance and the future of machine intelligence. Empower your team to build smarter, safer AI—subscribe today!

Comments

No comments yet. Why don’t you start the discussion?

Leave a Reply

Your email address will not be published. Required fields are marked *