DATASET
Open Source Community
meta-math/GSM8K_zh
GSM8K_zh is a Chinese dataset tailored for mathematical reasoning, consisting of problems and answers translated from the English GSM8K dataset. It includes 7,473 training samples and 1,319 test samples. Training samples contain full questions and answers; test samples provide only the translated questions. The dataset is suitable for Chinese–English question‑answering tasks, especially for mathematical problem solving.
Updated 12/4/2023
hugging_face
Description
Dataset Overview
Basic Information
- License: MIT
- Task Category: Question Answering
- Languages: English, Chinese
- Tags: Mathematics, Math QA, Chinese Math QA
- Scale: n<1K
Dataset Description
- Name: GSM8K_zh
- Description: GSM8K_zh is a dataset for Chinese mathematical reasoning. Question‑answer pairs are translated from the GSM8K dataset (https://github.com/openai/grade-school-math/tree/master) using few‑shot prompting with GPT‑3.5‑Turbo.
- Sample Count: 7,473 training samples and 1,319 test samples. Training data are used for supervised fine‑tuning, while test data are for evaluation.
- Sample Structure:
- Training: includes
question_zh(question) andanswer_zh(answer) keys. - Test: provides only the translated question (
question_zh).
- Training: includes
Citation
- Reference:
@article{yu2023metamath, title={MetaMath: Bootstrap Your Own Mathematical Questions for Large Language Models}, author={Yu, Longhui and Jiang, Weisen and Shi, Han and Yu, Jincheng and Liu, Zhengying and Zhang, Yu and Kwok, James T and Li, Zhenguo and Weller, Adrian and Liu, Weiyang}, journal={arXiv preprint arXiv:2309.12284}, year={2023} }
AI studio
Generate PPTs instantly with Nano Banana Pro.
Generate PPT NowAccess Dataset
Login to Access
Please login to view download links and access full dataset details.
Topics
Mathematical Reasoning
Chinese Math QA
Source
Organization: hugging_face
Created: Unknown
Power Your Data Analysis with Premium AI Models
Supporting GPT-5, Claude-4, DeepSeek v3, Gemini and more.
Enjoy a free trial and save 20%+ compared to official pricing.