News
Given enough time to "think," small language models can beat LLMs at math and coding tasks by generating and verifying multiple answers.
Different ways to use inference-time compute The researchers explored two main strategies for using inference-time compute to improve LLM performance.
Test-time scaling means OpenAI is using more compute during ChatGPT’s inference phase, the period of time after you press enter on a prompt.
Results that may be inaccessible to you are currently showing.
Hide inaccessible results