Muhammad Umair Mohsin’s Post

View profile for Muhammad Umair Mohsin, graphic

Gen AI | Data Analytics | Python | LLMs | Business Analyst | Consultant | SAP

🍓 𝐓𝐡𝐞 𝐦𝐮𝐜𝐡 𝐚𝐰𝐚𝐢𝐭𝐞𝐝 𝐧𝐞𝐰 𝐎𝐩𝐞𝐧𝐀𝐈 𝐨1 𝐢𝐬 𝐡𝐞𝐫𝐞  ⭐ The model uses a "hidden" 𝐜𝐡𝐚𝐢𝐧-𝐨𝐟-𝐭𝐡𝐨𝐮𝐠𝐡𝐭 𝐩𝐫𝐨𝐜𝐞𝐬𝐬 (COT), enabling it to think through problems in a more human style. 1️⃣ It turns out that this more in-depth thinking improves test-time performance considerably, enabling better, more accurate findings with extended processing (10–20 seconds). 2️⃣ The longer the model takes to evaluate a task, the more accurate and robust its results are usually. 📈 Performance and Benchmarks: ➡ The model outperforms human PhD-level accuracy on a benchmark of problems in chemistry, biology, and physics (GPQA) ➡ Programming: Ranked in the 89th percentile in Codeforces, showcasing advanced problem-solving and coding abilities. Well the downside is that I believe the cost of using this will be higher than GPT-4 and other models but didn't find the exact breakdown on the website yet. Lastly for the approach this model uses can be read in the paper below: https://lnkd.in/eHTJvdxA https://lnkd.in/eFyt3ggF #OpenAI #OpenAIo1 #LLMs #Reasoning #ChainofThought

View profile for Jim Fan, graphic
Jim Fan Jim Fan is an Influencer

NVIDIA Senior Research Manager & Lead of Embodied AI (GEAR Lab). Stanford Ph.D. Building Humanoid Robots and Physical AI. OpenAI's first intern. Sharing insights on the bleeding edge of AI.

OpenAI Strawberry (o1) is out! We are finally seeing the paradigm of inference-time scaling popularized and deployed in production. As Sutton said in the Bitter Lesson, there're only 2 techniques that scale indefinitely with compute: learning & search. It's time to shift focus to the latter. 1. You don't need a huge model to perform reasoning. Lots of parameters are dedicated to memorizing facts, in order to perform well in benchmarks like trivia QA. It is possible to factor out reasoning from knowledge, i.e. a small "reasoning core" that knows how to call tools like browser and code verifier. Pre-training compute may be decreased. 2. A huge amount of compute is shifted to serving inference instead of pre/post-training. LLMs are text-based simulators. By rolling out many possible strategies and scenarios in the simulator, the model will eventually converge to good solutions. The process is a well-studied problem like AlphaGo's monte carlo tree search (MCTS). 3. OpenAI must have figured out the inference scaling law a long time ago, which academia is just recently discovering. Two papers came out on Arxiv a week apart last month: - Large Language Monkeys: Scaling Inference Compute with Repeated Sampling. Brown et al. finds that DeepSeek-Coder increases from 15.9% with one sample to 56% with 250 samples on SWE-Bench, beating Sonnet-3.5. - Scaling LLM Test-Time Compute Optimally can be More Effective than Scaling Model Parameters. Snell et al. finds that PaLM 2-S beats a 14x larger model on MATH with test-time search. 4. Productionizing o1 is much harder than nailing the academic benchmarks. For reasoning problems in the wild, how to decide when to stop searching? What's the reward function? Success criterion? When to call tools like code interpreter in the loop? How to factor in the compute cost of those CPU processes? Their research post didn't share much. 5. Strawberry easily becomes a data flywheel. If the answer is correct, the entire search trace becomes a mini dataset of training examples, which contain both positive and negative rewards. This in turn improves the reasoning core for future versions of GPT, similar to how AlphaGo’s value network — used to evaluate quality of each board position — improves as MCTS generates more and more refined training data.

  • No alternative text description for this image

To view or add a comment, sign in

Explore topics