Datasets:

Modalities:
Image
Text
Formats:
parquet
ArXiv:
Libraries:
Datasets
pandas
License:
AI4Industry commited on
Commit
04f485b
·
verified ·
1 Parent(s): bcaadc5

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -2
README.md CHANGED
@@ -53,13 +53,13 @@ tags:
53
  RxnBench is a visual question answering (VQA) benchmark comprising 1,525 multiple-choice questions (MCQs) at the PhD-level of organic chemistry reaction understanding.
54
 
55
  The benchmark is built from 305 scientific figures drawn from high-impact OpenAssess journals.
56
- For each figure, domain experts carefully designed five multiple-choice VQA questions targeting the interpretation of organic reaction diagrams.
57
  These questions were further refined through multiple rounds of rigorous review and revision to ensure both clarity and scientific accuracy.
58
  The questions cover a variety of types, including the description of chemical reaction images, extraction of reaction content, recognition of molecules or Markush structures, and determination of mechanisms.
59
  This benchmark challenges visual-language models on their foundational knowledge of organic chemistry, multimodal contextual reasoning, and chemical reasoning skills.
60
 
61
 
62
- The benchmark is released in both English and Chinese versions. All data are annotated by experts in the organic chemistry domain and subsequently validated through two rounds of expert review.
63
 
64
 
65
  ## 🎯 Benchmark Evaluation
 
53
  RxnBench is a visual question answering (VQA) benchmark comprising 1,525 multiple-choice questions (MCQs) at the PhD-level of organic chemistry reaction understanding.
54
 
55
  The benchmark is built from 305 scientific figures drawn from high-impact OpenAssess journals.
56
+ For each figure, five multiple-choice VQA questions were constructed by domain experts to assess the interpretation of organic reaction diagrams, with all annotations subsequently validated through two independent rounds of expert review.
57
  These questions were further refined through multiple rounds of rigorous review and revision to ensure both clarity and scientific accuracy.
58
  The questions cover a variety of types, including the description of chemical reaction images, extraction of reaction content, recognition of molecules or Markush structures, and determination of mechanisms.
59
  This benchmark challenges visual-language models on their foundational knowledge of organic chemistry, multimodal contextual reasoning, and chemical reasoning skills.
60
 
61
 
62
+ The benchmark is released in both English and Chinese versions.
63
 
64
 
65
  ## 🎯 Benchmark Evaluation