Moreover, they show a counter-intuitive scaling limit: their reasoning hard work boosts with problem complexity up to a degree, then declines In spite of possessing an suitable token funds. By evaluating LRMs with their regular LLM counterparts underneath equal inference compute, we recognize a few overall performance regimes: (1) https://www.youtube.com/watch?v=snr3is5MTiU