Vivold Consulting

Why Cohere’s ex-AI research lead is betting against the scaling race

Key Insights

Former Cohere AI research head Sara Hooker argues that the industry's obsession with ever-larger language models has hit diminishing returns. Her new venture focuses on efficient architectures and data quality, not sheer scale.

Stay Updated

Get the latest insights delivered to your inbox

Betting against the AI scaling race

Sara Hooker, once leading research at Cohere, has become one of the most vocal critics of the industry's 'bigger is always better' mentality. In her recent talk and interview with TechCrunch, she argues that model scaling has plateaued — and that breakthroughs will come from smarter training data, specialized models, and adaptive reasoning systems.

What Hooker is saying


- Massive models like GPT-4, Claude 3, and Gemini Ultra show marginal accuracy gains but at exponential cost in compute and energy.
- The new research focus is data curation — improving signal-to-noise ratios and dynamic sampling instead of blindly increasing corpus size.
- Hooker calls for a future of modular AI — smaller, task-specific systems that communicate rather than one monolithic model.

The technical case


- Studies from Hooker’s lab show that training efficiency doubles when data is pre-filtered for reasoning diversity rather than scale.
- Mixture-of-experts architectures are regaining attention, offering large-model performance at a fraction of cost.
- She advocates for open, interpretable benchmarks beyond MMLU and HELM to measure real-world reliability.

Why this matters for the AI ecosystem


- The “scaling ceiling” could reshape the AI race. Firms like Anthropic and OpenAI may pivot toward data efficiency and reasoning-enhanced training.
- VC interest is shifting from GPU-driven startups to optimization-focused ventures building inference-efficient stacks.
- For enterprise buyers, the new competitive edge becomes: Can you get 90% of GPT-4 performance for 10% of the cost?

The human and policy angle


- Hooker also highlights the carbon footprint of scaling and urges policymakers to include compute transparency standards.
- “We’ve mistaken size for intelligence,” she says. “Now we have to make AI more human by making it smaller.”

Related Articles

Salesforce Unveils AI-Powered Slack Makeover with 30 New Features

Salesforce has announced a major update to Slack, introducing over 30 new AI-driven features aimed at enhancing workplace productivity and collaboration. Key enhancements include: - Advanced Slackbot capabilities for drafting content, summarizing conversations, and answering queries. - Integration with Salesforce CRM and third-party apps to provide context-aware assistance. - Proactive recommendations during video calls, such as surfacing relevant Salesforce records when key names are mentioned.

Salesforce Ramps Up Agentic AI Research with New Foundry Project

Salesforce has launched the AI Foundry, a new initiative aimed at accelerating agentic AI research and development. The project focuses on: - Bridging foundational research and product innovation through collaboration with strategic customers and academic partners. - Developing AI tools for high-impact enterprise areas, including simulated environments for testing AI agents and enhancing solutions like Agentforce Voice. - Exploring ambient intelligence to provide proactive, context-aware assistance without constant user input.

VHA Deploys Salesforce-Powered Agentic Operating System, Saving Thousands of Staff Hours for Front-Line Veteran Care

The Veterans Health Administration (VHA) has implemented a Salesforce-powered agentic operating system, resulting in significant operational efficiencies. Key outcomes include: - Transitioning from static reporting to automated problem-solving, eliminating administrative silos. - Freeing thousands of staff hours, allowing more focus on direct Veteran support. - Creating a connected performance management layer, enhancing care delivery across facilities.