JUHE API Marketplace
DATASET
Open Source Community

stanfordnlp/SHP

The Stanford Human Preferences Dataset (SHP) comprises 385 K Reddit user preference records across 18 distinct topical domains, intended for training RLHF reward models and NLG evaluation models. Each example consists of a Reddit post containing a question or instruction and a pair of top‑level comments, where one comment is preferred by the Reddit community. Preferences are inferred from timestamp information to ensure they reflect comment helpfulness rather than harmfulness. The dataset includes training, validation, and test splits for 18 sub‑forums, with each sub‑forum stored as JSONL files.

Updated 10/10/2023
hugging_face

Description

Dataset Overview

Dataset Name

Stanford Human Preferences Dataset (SHP)

Dataset Size

385K entries

Dataset Task Categories

  • Text Generation
  • Question Answering

Dataset Tags

  • Human Feedback
  • RLHF
  • Preference
  • Reddit
  • Preference Model
  • RL
  • NLG
  • Evaluation

Dataset Language

English

Dataset Content

SHP contains 385K human preference records for question/instruction answering, covering 18 domains such as cooking, legal advice, etc. Each example includes a Reddit post, a question/instruction, and two top comments, with one comment preferentially chosen by the Reddit community.

Dataset Structure

The dataset is divided into 18 directories, each representing a sub‑forum, containing JSONL files for training, validation, and testing.

Dataset Uses

For training RLHF reward models and NLG evaluation models.

Dataset Differences from Other Datasets

  • Compared with the Anthropics HH‑RLHF dataset, SHP’s data are naturally occurring and human‑written, whereas HH‑RLHF’s responses are machine‑generated.
  • Compared with the ELI5 dataset, SHP infers preferences using timestamps, while ELI5 only provides comments and scores.

Dataset Preprocessing

Preprocessing is minimal, limited to expanding sub‑forum‑specific abbreviations and removing URLs from hyperlinks.

Building Preference Models

It is recommended to fine‑tune large models, such as FLAN‑T5‑xl, to predict human preferences, and to report performance curves based on score_ratio.

Dataset Limitations

  • SHP is not intended for harm reduction and does not contain toxic content needed for toxicity detection training.
  • More preferred responses are not necessarily more factually accurate.

Dataset License

Data were scraped in accordance with Reddit API terms of use; user‑generated content remains owned by the users, and Reddit grants a non‑exclusive, non‑transferable, non‑sublicensable, revocable license.

Contact Information

Email: kawin@stanford.edu

Dataset Creators

Kawin Ethayarajh, Heidi (Chenyu) Zhang, Yizhong Wang, Dan Jurafsky

AI studio

Generate PPTs instantly with Nano Banana Pro.

Generate PPT Now

Access Dataset

Login to Access

Please login to view download links and access full dataset details.

Topics

Machine Learning
Dataset Difficulty Evaluation

Source

Organization: hugging_face

Created: Unknown

Power Your Data Analysis with Premium AI Models

Supporting GPT-5, Claude-4, DeepSeek v3, Gemini and more.

Enjoy a free trial and save 20%+ compared to official pricing.