+
Skip to content

Metadata correction for 2024.genbench-1.1 #6187

@kddubey

Description

@kddubey

JSON data block

{
  "anthology_id": "2024.genbench-1.1",
  "abstract": "Few-shot learning benchmarks are critical for evaluating modern NLP techniques. It is possible, however, that benchmarks favor methods which easily make use of unlabeled text, because researchers can use unlabeled text from the test set to pretrain their models. Given the dearth of research on this potential problem, we run experiments to quantify the bias caused by pretraining on unlabeled test set text instead of on unlabeled, independently drawn text. Controlled few-shot and zero-shot experiments on 25 classification tasks and 3 language models—BERT, GPT-2, and Mistral 7B—do not find evidence of overoptimism. Furthermore, we demonstrate the importance of repeated subsampling when studying few-shot text classification, and recommend that few-shot learning benchmarks include multiple training folds. Code and data are available here: <url>https://github.com/kddubey/pretrain-on-test</url>."
}

Metadata

Metadata

Labels

approvedUsed to note team approval of metadata requestscorrectionfor corrections submitted to the anthologymetadataCorrection to metadata

Type

No type

Projects

No projects

Milestone

No milestone

Relationships

None yet

Development

No branches or pull requests

Issue actions

    点击 这是indexloc提供的php浏览器服务,不要输入任何密码和下载