Researchers from Stanford University and Google DeepMind have unveiled Step-Wise Reinforcement Learning (SWiRL), a technique designed to enhance the ability of large language models (LLMs) to tackle complex tasks requiring multi-step reasoning and tool use.
As the interest in AI agents and LLM tool use continues to increase, this technique could offer substantial benefits for enterprises looking…
When AI reasoning goes wrong: Microsoft Research shows more tokens can mean more problems
April 16, 2025
In a Nutshell
Microsoft Research finds that inference-time scaling methods for large language models don’t universally improve performance. Varying benefits, token inefficiency, and cost unpredictability challenge assumptions. Verification mechanisms enhance model…
New open source AI company Deep Cogito releases first models and they’re already topping the charts
April 9, 2025
Deep Cogito, a new AI research startup based in San Francisco, officially emerged from stealth today with Cogito v1, a new line of open source large language models (LLMs) fine-tuned from Meta’s Llama 3.2 and equipped with hybrid reasoning capabilities — the ability to…
AI lie detector: How HallOumi’s open-source approach to hallucination could unlock enterprise AI adoption
April 4, 2025
In the race to deploy enterprise AI, one obstacle consistently blocks the path: hallucinations. These fabricated responses from AI systems have caused everything from legal sanctions for attorneys to companies being forced to honor fictitious policies.
Organizations have tried different approaches to solving the hallucination challenge, including fine-tuning with better data, retrieval augmented…
At the DataGrail Summit 2024 this week, industry leaders delivered a stark warning about the rapidly advancing risks associated with AI.
Dave Zhou, CISO of Instacart, and Jason Clinton, CISO of Anthropic, highlighted the urgent need for robust security measures to keep pace…