Datasets:

Modalities:
Text
Formats:
parquet
ArXiv:
Libraries:
Datasets
pandas
matthieumeeus97 commited on
Commit
5a41585
1 Parent(s): 10048f4

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +5 -0
README.md CHANGED
@@ -21,6 +21,11 @@ This dataset contains **sequences from** ArXiv papers randomly sampled from the
21
  We randomly sample 1,000 documents members and 1,000 non-members, ensuring that the selected documents have at least 5,000 words (any sequences of characters seperated by a white space).
22
  This dataset contains the first 25 sequences of 200 words from all the documents made available in full [here](https://huggingface.co/datasets/imperial-cpg/pile_arxiv_doc_mia).
23
 
 
 
 
 
 
24
  The dataset can be used to develop and evaluate document-level MIAs against LLMs trained on The Pile.
25
  Target models include the suite of Pythia and GPTNeo models, to be found [here](https://huggingface.co/EleutherAI). Our understanding is that the deduplication executed on the Pile to create the "Pythia-dedup" models has been only done on the training dataset, suggesting this dataset of members/non-members also to be valid for these models.
26
 
 
21
  We randomly sample 1,000 documents members and 1,000 non-members, ensuring that the selected documents have at least 5,000 words (any sequences of characters seperated by a white space).
22
  This dataset contains the first 25 sequences of 200 words from all the documents made available in full [here](https://huggingface.co/datasets/imperial-cpg/pile_arxiv_doc_mia).
23
 
24
+ The dataset contains as columns:
25
+ - text: the raw text of the sequence
26
+ - label: binary label for membership (1=member)
27
+ - doc_idx: index allowing to group sequences to the same, original document
28
+
29
  The dataset can be used to develop and evaluate document-level MIAs against LLMs trained on The Pile.
30
  Target models include the suite of Pythia and GPTNeo models, to be found [here](https://huggingface.co/EleutherAI). Our understanding is that the deduplication executed on the Pile to create the "Pythia-dedup" models has been only done on the training dataset, suggesting this dataset of members/non-members also to be valid for these models.
31