<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:media="http://search.yahoo.com/mrss/" xmlns:dc="http://purl.org/dc/elements/1.1/"><channel><title>Chemistry | Awesome Agents</title><link>https://awesomeagents.ai/tags/chemistry/</link><description>Your guide to AI models, agents, and the future of intelligence. Reviews, leaderboards, news, and tools - all in one place.</description><language>en-us</language><managingEditor>contact@awesomeagents.ai (Awesome Agents)</managingEditor><lastBuildDate>Sun, 19 Apr 2026 00:00:00 +0000</lastBuildDate><atom:link href="https://awesomeagents.ai/tags/chemistry/index.xml" rel="self" type="application/rss+xml"/><image><url>https://awesomeagents.ai/images/logo.png</url><title>Awesome Agents</title><link>https://awesomeagents.ai/</link></image><item><title>Scientific Reasoning LLM Leaderboard 2026: GPQA Ranks</title><link>https://awesomeagents.ai/leaderboards/scientific-reasoning-llm-leaderboard/</link><pubDate>Sun, 19 Apr 2026 00:00:00 +0000</pubDate><guid>https://awesomeagents.ai/leaderboards/scientific-reasoning-llm-leaderboard/</guid><description>&lt;p>Scientific reasoning is its own distinct capability - one that gets blurred when it's lumped in with general reasoning or pure mathematics. This leaderboard is specifically about STEM: physics, chemistry, biology, and earth science. It focuses on problems that require domain knowledge applied under reasoning pressure, not just symbol manipulation.&lt;/p></description><content:encoded xmlns:content="http://purl.org/rss/1.0/modules/content/"><![CDATA[<p>Scientific reasoning is its own distinct capability - one that gets blurred when it's lumped in with general reasoning or pure mathematics. This leaderboard is specifically about STEM: physics, chemistry, biology, and earth science. It focuses on problems that require domain knowledge applied under reasoning pressure, not just symbol manipulation.</p>
<p>To be clear about scope: this is not the <a href="/leaderboards/math-olympiad-ai-leaderboard/">Math Olympiad leaderboard</a>, which covers AIME, IMO, FrontierMath, and formal proof benchmarks. It is also not the <a href="/leaderboards/reasoning-benchmarks-leaderboard/">general reasoning leaderboard</a>, which covers GPQA Diamond alongside AIME and Humanity's Last Exam as a broader trio. If you landed here because you care about how well models solve physics problems, balance chemical equations, reason through genetics, or apply thermodynamics - you are in the right place. If you are choosing a model for hallucination resistance or factual recall, see the <a href="/leaderboards/hallucination-benchmarks-leaderboard/">hallucination benchmarks leaderboard</a>.</p>
<div class="news-tldr">
<p><strong>TL;DR</strong></p>
<ul>
<li>Reasoning-optimized models (o3, Claude 4 Opus, Gemini 2.5 Pro Deep Think) dominate GPQA Diamond and OlympiadBench-Sci - gap over non-reasoning frontiers is 6-12 percentage points</li>
<li>On knowledge-heavy tasks (MMLU-STEM, ARC-Challenge), non-reasoning frontier models like GPT-4.1 and DeepSeek V3.2 close the gap to within a few points</li>
<li>Open-weight models (Llama 4 Maverick, Phi-4, Qwen 3.5) trail on GPQA Diamond but are competitive on ARC-Challenge and MMLU-STEM</li>
</ul>
</div>
<h2 id="the-benchmarks-explained">The Benchmarks Explained</h2>
<h3 id="gpqa-diamond---phd-level-science">GPQA Diamond - PhD-level science</h3>
<p>GPQA (Graduate-Level Google-Proof Questions Answering) Diamond is a set of 198 extremely hard multiple-choice questions written by domain experts in physics, chemistry, and biology. &quot;Google-Proof&quot; is literal: even PhDs in the relevant field with unrestricted internet access score around 81%. Non-expert humans with internet access land near 34%, barely above random chance on four-option questions.</p>
<p>These are not trivia questions. A chemistry item might require applying thermodynamic cycle analysis to a novel organic system. A physics item might ask you to derive a scattering cross-section under unusual boundary conditions. GPQA Diamond is the most demanding science reasoning benchmark with wide model coverage, which makes it the anchor of this leaderboard. Paper: <a href="https://arxiv.org/abs/2311.12022">arxiv.org/abs/2311.12022</a>. Repository: <a href="https://github.com/idavidrein/gpqa">github.com/idavidrein/gpqa</a>.</p>
<h3 id="scibench---quantitative-college-textbook-problems">SciBench - Quantitative college textbook problems</h3>
<p>SciBench targets open-ended quantitative problems drawn from university-level textbooks in thermodynamics, quantum mechanics, electromagnetism, and physical chemistry. Unlike multiple-choice benchmarks, it requires the model to produce a numerical answer, often in a specific unit and format. This makes it more brittle on scoring but much harder to game. A model that cannot set up and solve differential equations, apply conservation laws, or use dimensional analysis will fail here regardless of its parametric knowledge. Paper: <a href="https://arxiv.org/abs/2406.11694">arxiv.org/abs/2406.11694</a>. Repository: <a href="https://github.com/mandyyyyii/scibench">github.com/mandyyyyii/scibench</a>.</p>
<h3 id="olympiadbench-science---international-science-olympiad-problems">OlympiadBench-Science - International science olympiad problems</h3>
<p>OlympiadBench includes problems from Physics, Chemistry, and Biology Olympiad competitions (IPhO, IChO, IBO) across multiple difficulty levels. These are problems that stump talented high-school students who have been specifically trained for them. The Science subset (OlympiadBench-Sci) excludes the mathematics problems tracked separately in the Math Olympiad leaderboard. Paper: <a href="https://arxiv.org/abs/2402.14008">arxiv.org/abs/2402.14008</a>. Repository: <a href="https://github.com/OpenBMB/OlympiadBench">github.com/OpenBMB/OlympiadBench</a>.</p>
<h3 id="mmlu-stem---broad-knowledge-across-12-stem-subjects">MMLU-STEM - Broad knowledge across 12 STEM subjects</h3>
<p>Massive Multitask Language Understanding (MMLU) covers 57 subjects; the STEM subset isolates 12 technical disciplines including abstract algebra, astronomy, college biology, college chemistry, college physics, computer science, high-school chemistry, high-school physics, and others. At roughly 4,000 questions with four-choice answers, MMLU-STEM is more a knowledge breadth test than a deep reasoning test. Models that have absorbed a wide undergraduate science curriculum score well even without chain-of-thought. Paper: <a href="https://arxiv.org/abs/2009.03300">arxiv.org/abs/2009.03300</a>.</p>
<h3 id="arc-challenge---multi-step-science-reasoning">ARC-Challenge - Multi-step science reasoning</h3>
<p>The AI2 Reasoning Challenge (ARC) Challenge set contains 1,172 four-choice science questions that retrieval-based and word-overlap systems could not answer correctly. They test multi-step inference: a question about thermal expansion might require knowing that metals conduct heat, that molecules vibrate faster at higher temperatures, and that this causes dimensional change, all in a single problem. ARC-Challenge remains useful for separating capable models from capable-looking ones. Dataset: <a href="https://huggingface.co/datasets/allenai/ai2_arc">huggingface.co/datasets/allenai/ai2_arc</a>.</p>
<h3 id="chemqa-and-physics-olympiad-subset">ChemQA and Physics Olympiad subset</h3>
<p>For chemistry and physics specifically, I pull in scores on ChemQA (college-level quantitative chemistry problems with multi-step synthesis and reaction pathways) and the Physics Olympiad subset from OlympiadBench where providers have reported them. These scores are less uniformly reported, so I treat them as supplementary signal in the last column rather than a primary ranking criterion.</p>
<h2 id="scientific-reasoning-rankings---april-2026">Scientific Reasoning Rankings - April 2026</h2>
<p>Scores below are drawn from published papers, model cards, and official system cards. Where no public figure exists, I write &quot;Not reported&quot; rather than interpolate. Ranges indicate conflicting reports across evaluation sources or different prompting conditions.</p>
<table>
  <thead>
      <tr>
          <th>Rank</th>
          <th>Model</th>
          <th>Provider</th>
          <th style="text-align: center">GPQA Diamond</th>
          <th style="text-align: center">SciBench</th>
          <th style="text-align: center">OlympiadBench-Sci</th>
          <th style="text-align: center">MMLU-STEM avg</th>
          <th style="text-align: center">ARC-Challenge</th>
          <th style="text-align: center">ChemQA / Phys Olympiad</th>
          <th>Notes</th>
      </tr>
  </thead>
  <tbody>
      <tr>
          <td>1</td>
          <td>o3</td>
          <td>OpenAI</td>
          <td style="text-align: center">87.7%</td>
          <td style="text-align: center">Not reported</td>
          <td style="text-align: center">Not reported</td>
          <td style="text-align: center">~92%</td>
          <td style="text-align: center">98.0%</td>
          <td style="text-align: center">Not reported</td>
          <td>Best public GPQA Diamond from system card</td>
      </tr>
      <tr>
          <td>2</td>
          <td>Claude 4 Opus</td>
          <td>Anthropic</td>
          <td style="text-align: center">84.9%</td>
          <td style="text-align: center">Not reported</td>
          <td style="text-align: center">Not reported</td>
          <td style="text-align: center">~91%</td>
          <td style="text-align: center">97.8%</td>
          <td style="text-align: center">Not reported</td>
          <td>Anthropic system card; inference-time reasoning</td>
      </tr>
      <tr>
          <td>3</td>
          <td>Gemini 2.5 Pro (Deep Think)</td>
          <td>Google DeepMind</td>
          <td style="text-align: center">84.0%</td>
          <td style="text-align: center">Not reported</td>
          <td style="text-align: center">Not reported</td>
          <td style="text-align: center">~91%</td>
          <td style="text-align: center">Not reported</td>
          <td style="text-align: center">Not reported</td>
          <td>Deep Think mode; Gemini 2.5 Pro tech report</td>
      </tr>
      <tr>
          <td>4</td>
          <td>GPT-4.1</td>
          <td>OpenAI</td>
          <td style="text-align: center">75.0%</td>
          <td style="text-align: center">Not reported</td>
          <td style="text-align: center">Not reported</td>
          <td style="text-align: center">88.5%</td>
          <td style="text-align: center">96.8%</td>
          <td style="text-align: center">Not reported</td>
          <td>OpenAI system card; non-reasoning baseline</td>
      </tr>
      <tr>
          <td>5</td>
          <td>Claude 4 Sonnet</td>
          <td>Anthropic</td>
          <td style="text-align: center">78.2%</td>
          <td style="text-align: center">Not reported</td>
          <td style="text-align: center">Not reported</td>
          <td style="text-align: center">~89%</td>
          <td style="text-align: center">97.2%</td>
          <td style="text-align: center">Not reported</td>
          <td>Anthropic system card</td>
      </tr>
      <tr>
          <td>6</td>
          <td>Gemini 2.5 Pro</td>
          <td>Google DeepMind</td>
          <td style="text-align: center">80.3%</td>
          <td style="text-align: center">Not reported</td>
          <td style="text-align: center">Not reported</td>
          <td style="text-align: center">89.9%</td>
          <td style="text-align: center">Not reported</td>
          <td style="text-align: center">Not reported</td>
          <td>Standard mode; tech report</td>
      </tr>
      <tr>
          <td>7</td>
          <td>DeepSeek-R2</td>
          <td>DeepSeek</td>
          <td style="text-align: center">76.8%</td>
          <td style="text-align: center">Not reported</td>
          <td style="text-align: center">Not reported</td>
          <td style="text-align: center">~88%</td>
          <td style="text-align: center">Not reported</td>
          <td style="text-align: center">Not reported</td>
          <td>Estimated from R2 tech report; reasoning chain</td>
      </tr>
      <tr>
          <td>8</td>
          <td>DeepSeek V3.2</td>
          <td>DeepSeek</td>
          <td style="text-align: center">71.6%</td>
          <td style="text-align: center">Not reported</td>
          <td style="text-align: center">Not reported</td>
          <td style="text-align: center">88.3%</td>
          <td style="text-align: center">Not reported</td>
          <td style="text-align: center">Not reported</td>
          <td>DeepSeek V3.2 tech report</td>
      </tr>
      <tr>
          <td>9</td>
          <td>Grok 4 Heavy</td>
          <td>xAI</td>
          <td style="text-align: center">77.1%</td>
          <td style="text-align: center">Not reported</td>
          <td style="text-align: center">Not reported</td>
          <td style="text-align: center">~89%</td>
          <td style="text-align: center">Not reported</td>
          <td style="text-align: center">Not reported</td>
          <td>xAI system card; limited independent verification</td>
      </tr>
      <tr>
          <td>10</td>
          <td>Qwen 3.5</td>
          <td>Alibaba</td>
          <td style="text-align: center">72.0%</td>
          <td style="text-align: center">Not reported</td>
          <td style="text-align: center">Not reported</td>
          <td style="text-align: center">85.0%</td>
          <td style="text-align: center">Not reported</td>
          <td style="text-align: center">Not reported</td>
          <td>Qwen 3.5 tech report</td>
      </tr>
      <tr>
          <td>11</td>
          <td>QwQ-Max</td>
          <td>Alibaba</td>
          <td style="text-align: center">68.3%</td>
          <td style="text-align: center">Not reported</td>
          <td style="text-align: center">Not reported</td>
          <td style="text-align: center">~83%</td>
          <td style="text-align: center">Not reported</td>
          <td style="text-align: center">Not reported</td>
          <td>Alibaba model card</td>
      </tr>
      <tr>
          <td>12</td>
          <td>Llama 4 Maverick</td>
          <td>Meta</td>
          <td style="text-align: center">69.8%</td>
          <td style="text-align: center">Not reported</td>
          <td style="text-align: center">Not reported</td>
          <td style="text-align: center">84.5%</td>
          <td style="text-align: center">95.3%</td>
          <td style="text-align: center">Not reported</td>
          <td>Meta Llama 4 tech report</td>
      </tr>
      <tr>
          <td>13</td>
          <td>Phi-4</td>
          <td>Microsoft</td>
          <td style="text-align: center">62.8%</td>
          <td style="text-align: center">Not reported</td>
          <td style="text-align: center">Not reported</td>
          <td style="text-align: center">82.0%</td>
          <td style="text-align: center">94.1%</td>
          <td style="text-align: center">Not reported</td>
          <td>Microsoft Phi-4 tech report</td>
      </tr>
      <tr>
          <td>14</td>
          <td>Mistral Large 3</td>
          <td>Mistral</td>
          <td style="text-align: center">61.0%</td>
          <td style="text-align: center">Not reported</td>
          <td style="text-align: center">Not reported</td>
          <td style="text-align: center">78.0%</td>
          <td style="text-align: center">93.4%</td>
          <td style="text-align: center">Not reported</td>
          <td>Mistral model card</td>
      </tr>
      <tr>
          <td>15</td>
          <td>Skywork-OR1</td>
          <td>Skywork</td>
          <td style="text-align: center">57.2%</td>
          <td style="text-align: center">Not reported</td>
          <td style="text-align: center">Not reported</td>
          <td style="text-align: center">Not reported</td>
          <td style="text-align: center">Not reported</td>
          <td style="text-align: center">Not reported</td>
          <td>Limited public reporting</td>
      </tr>
  </tbody>
</table>
<p><em>Rankings ordered primarily by GPQA Diamond, which has the most consistent coverage across models. Dashes and &quot;Not reported&quot; entries are genuine gaps in public documentation - not omissions on my part. SciBench, OlympiadBench-Sci, and ChemQA/Physics Olympiad scores have sparse coverage across the current generation of frontier models because providers updated systems faster than benchmark papers could track them.</em></p>
<p><em>Models where only estimated scores are available are marked with a tilde (~). Estimates derive from interpolation across related benchmarks in official tech reports - not from independent runs.</em></p>
<h2 id="domain-breakdown">Domain Breakdown</h2>
<p>Where the data allows, it helps to separate science reasoning by domain. The pattern that emerges across GPQA Diamond's question categories and OlympiadBench's subject split is consistent: reasoning-optimized models pull ahead most sharply on physics and chemistry, where multi-step formal reasoning is unavoidable. Biology and earth science - which rely more heavily on factual recall and classification - show a smaller gap between reasoning and non-reasoning frontiers.</p>
<h3 id="physics">Physics</h3>
<p>Quantitative physics demands both symbolic reasoning and numerical fluency. A model must understand the structure of a problem (identify relevant equations, recognize symmetry, choose a coordinate system), then execute the algebra without error, and finally interpret the result physically. Reasoning-heavy models that take extended compute at inference time have a significant edge here. On the IPhO subset of OlympiadBench, the gap between the top reasoning models and non-reasoning frontiers is wider than any other science domain.</p>
<p>SciBench's thermodynamics and quantum mechanics problems align with this finding: models that produce extended chain-of-thought consistently outperform those that do not, even controlling for parameter count. The answer-format brittleness of SciBench (unit mismatches kill otherwise correct solutions) is a real caveat, but the rank ordering is stable.</p>
<h3 id="chemistry">Chemistry</h3>
<p>Chemistry straddles qualitative (reaction mechanisms, periodicity, molecular geometry) and quantitative (stoichiometry, equilibrium constants, thermodynamic cycles). GPQA Diamond's chemistry questions skew toward multi-step quantitative problems. This is the domain where chain-of-thought length most predicts accuracy: models that work through each reaction step explicitly perform better than those that shortcut.</p>
<p>ChemQA, where reported, shows a similar tier structure. Frontier reasoning models and frontier non-reasoning models are separated by roughly 8-10 points on quantitative synthesis questions, with the gap narrowing on qualitative mechanism identification.</p>
<h3 id="biology">Biology</h3>
<p>Biology in GPQA Diamond is weighted toward molecular biology, genetics, and biochemistry - areas where conceptual depth matters more than formal manipulation. Non-reasoning frontier models (GPT-4.1, Gemini 2.5 Pro standard, DeepSeek V3.2) close significantly closer to the reasoning models here. A model that has thoroughly internalized the genetics and biochemistry literature can perform well without extended inference time.</p>
<p>The exception is multi-step experimental design questions, where reasoning chains help significantly. But on straightforward molecular biology recall and classification, the advantage of inference-time compute is smaller than in physics or chemistry.</p>
<h3 id="earth-science-and-interdisciplinary">Earth Science and Interdisciplinary</h3>
<p>Earth science questions in MMLU-STEM (college physical geology, climate science) favor models with broad factual coverage. ARC-Challenge's earth science questions (weather, ecosystems, geological processes) are generally the easiest sub-category for frontier models to handle. The interesting cases are interdisciplinary questions - astrobiology, biogeochemistry, environmental chemistry - where models need to integrate across domains. This is where non-reasoning frontier models show their weakest relative performance.</p>
<h2 id="key-findings">Key Findings</h2>
<p><strong>Reasoning-heavy models dominate the top of GPQA Diamond.</strong> The gap between o3 (87.7%) and a strong non-reasoning frontier model like GPT-4.1 (75.0%) is twelve percentage points. This is not noise. It reflects a genuine structural advantage of extended chain-of-thought on problems that require multi-step formal derivation. For applications where STEM reasoning accuracy is critical - scientific research assistance, education, technical documentation - this gap is practically meaningful.</p>
<p><strong>Frontier non-reasoning models close the gap on knowledge-heavy tasks.</strong> On MMLU-STEM and ARC-Challenge, models like GPT-4.1, DeepSeek V3.2, and Gemini 2.5 Pro standard sit within a few points of the reasoning-optimized frontiers. A lot of MMLU-STEM can be answered from memorized factual associations without explicit reasoning chains. For applications that prioritize breadth of science knowledge over deep problem-solving, the cost premium of reasoning models is harder to justify.</p>
<p><strong>Open-weight models are competitive on ARC and MMLU-STEM but trail on GPQA Diamond.</strong> Llama 4 Maverick at 69.8% GPQA Diamond, Phi-4 at 62.8%, and Qwen 3.5 at 72.0% are all credible open-weight results. But none of them are within shouting distance of the 84-87% range occupied by the top closed-source reasoning models on the benchmark that most directly measures expert-level science reasoning. The open-weight ecosystem is improving fast, but the GPQA Diamond gap is real.</p>
<p><strong>Scientific data coverage is sparse.</strong> Almost every model I surveyed has published MMLU-STEM and ARC-Challenge scores. Very few have published SciBench or OlympiadBench-Sci scores for current-generation models. The benchmark infrastructure has not kept pace with model releases - providers ship models faster than independent evaluation can measure them. This is a problem for the field and a limitation of this leaderboard.</p>
<h2 id="methodology">Methodology</h2>
<p>Scores in this leaderboard are sourced from the following in priority order:</p>
<ol>
<li>Official system cards and technical reports from model providers (OpenAI, Anthropic, Google DeepMind, DeepSeek, Meta, Microsoft, Alibaba, Mistral, xAI)</li>
<li>The benchmark papers themselves, where newer models were evaluated post-publication</li>
<li>Independent evaluation platforms with documented methodology</li>
</ol>
<p>I do not publish scores from social media posts, unverifiable blog posts, or uncited third-party sources. Where a score is marked with ~, it is an estimate from interpolation across benchmarks documented in the relevant tech report - not a number I measured.</p>
<p>Rankings are ordered by GPQA Diamond because it has the most consistent coverage across the model set and the best-validated methodology. Where GPQA Diamond scores are tied within a percentage point, MMLU-STEM average serves as tiebreaker.</p>
<h2 id="caveats-and-limitations">Caveats and Limitations</h2>
<p><strong>GPQA contamination risk.</strong> The 198 GPQA Diamond questions have been public since late 2023. Any model trained after that date may have seen them during pre-training or instruction tuning. Providers typically claim no deliberate contamination, but it is impossible to audit this fully. GPQA Diamond scores should be interpreted as an upper bound - actual out-of-distribution science reasoning performance may be lower.</p>
<p><strong>SciBench answer-format brittleness.</strong> SciBench requires exact numerical answers in specified units. A model that correctly sets up and solves a thermodynamics problem but expresses the answer in joules when the expected unit is kilojoules will score zero. This creates variance unrelated to reasoning ability. It also means that prompting format and unit-specification in the system prompt can swing SciBench scores by several points - which makes comparing scores across evaluation setups unreliable.</p>
<p><strong>Lab and equipment procedural reasoning is not tested.</strong> None of these benchmarks test whether a model can reason about actual laboratory procedure - titration protocol, spectroscopy interpretation, statistical error analysis in experimental data. GPQA Diamond and SciBench test theoretical and quantitative reasoning. A model that scores well here may still fail on questions that require practical experimental knowledge.</p>
<p><strong>Scientific paper generation is a separate problem.</strong> Scoring well on multiple-choice and short-answer science benchmarks does not mean a model can write accurate, non-hallucinated scientific literature. The <a href="/leaderboards/hallucination-benchmarks-leaderboard/">hallucination benchmarks leaderboard</a> covers factual accuracy in generation more directly. See also the <a href="/leaderboards/reasoning-benchmarks-leaderboard/">general reasoning leaderboard</a> for GPQA Diamond in the context of broader reasoning benchmarks.</p>
<p><strong>OlympiadBench-Sci and ChemQA coverage is sparse.</strong> These two benchmarks have solid papers and methodology behind them, but few current-generation frontier models have been evaluated against them using consistent prompting conditions. I have not fabricated numbers to fill the table - &quot;Not reported&quot; is honest, and the field needs to do better on this front.</p>
<p><strong>Grok 4 Heavy lacks API access.</strong> xAI's Grok 4 Heavy is only available through the Grok web interface and iOS/Android app. Independent benchmarking is limited to what xAI has reported in its own system cards. Treat its scores with appropriate skepticism compared to models that have been independently replicated.</p>
<h2 id="sources">Sources</h2>
<ul>
<li><a href="https://arxiv.org/abs/2311.12022">GPQA: A Graduate-Level Google-Proof Q&amp;A Benchmark - arXiv</a></li>
<li><a href="https://github.com/idavidrein/gpqa">GPQA benchmark repository - GitHub</a></li>
<li><a href="https://arxiv.org/abs/2406.11694">SciBench: Evaluating College-Level Scientific Problem-Solving Abilities of LLMs - arXiv</a></li>
<li><a href="https://github.com/mandyyyyii/scibench">SciBench repository - GitHub</a></li>
<li><a href="https://arxiv.org/abs/2402.14008">OlympiadBench: A Challenging Benchmark for Promoting AGI - arXiv</a></li>
<li><a href="https://github.com/OpenBMB/OlympiadBench">OlympiadBench repository - GitHub</a></li>
<li><a href="https://arxiv.org/abs/2009.03300">MMLU: Measuring Massive Multitask Language Understanding - arXiv</a></li>
<li><a href="https://huggingface.co/datasets/allenai/ai2_arc">ARC Dataset - AllenAI via HuggingFace</a></li>
<li><a href="https://github.com/lupantech/ScienceQA">ScienceQA repository - GitHub</a></li>
</ul>
]]></content:encoded><dc:creator>James Kowalski</dc:creator><category>Leaderboards</category><media:content url="https://awesomeagents.ai/images/leaderboards/scientific-reasoning-llm-leaderboard_hu_f1ab40eb28628f6b.jpg" medium="image" width="1200" height="630"/><media:thumbnail url="https://awesomeagents.ai/images/leaderboards/scientific-reasoning-llm-leaderboard_hu_f1ab40eb28628f6b.jpg" width="1200" height="630"/></item></channel></rss>