honicky's picture
Add train/val/test splits for HDFS logs
977c62c verified
---
language: en
tags:
- log-analysis
- hdfs
- anomaly-detection
license: mit
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
- split: validation
path: data/validation-*
- split: test
path: data/test-*
dataset_info:
features:
- name: event_encoded
dtype: string
- name: tokenized_block
sequence: int64
- name: block_id
dtype: string
- name: label
dtype: string
- name: __index_level_0__
dtype: int64
splits:
- name: train
num_bytes: 1159074302
num_examples: 460048
- name: validation
num_bytes: 145089712
num_examples: 57506
- name: test
num_bytes: 144844752
num_examples: 57507
download_size: 173888975
dataset_size: 1449008766
---
# HDFS Logs Train/Val/Test Splits
This dataset contains preprocessed HDFS log sequences split into train, validation, and test sets for anomaly detection tasks.
## Dataset Description
The dataset is derived from the HDFS log dataset, which contains system logs from a Hadoop Distributed File System (HDFS).
Each sequence represents a block of log messages, labeled as either normal or anomalous. The dataset has been preprocessed
using the Drain algorithm to extract structured fields and identify event types.
### Data Fields
- `block_id`: Unique identifier for each HDFS block, used to group log messages into blocks
- `event_encoded`: The preprocessed log sequence with event IDs and parameters
- `tokenized_block`: The tokenized log sequence, used for training
- `label`: Classification label ('Normal' or 'Anomaly')
### Data Splits
- Training set: 460,049 sequences (80%)
- Validation set: 57,506 sequences (10%)
- Test set: 57,506 sequences (10%)
The splits are stratified by the Label field to maintain class distribution across splits.
## Source Data
Original data source: https://zenodo.org/records/8196385/files/HDFS_v1.zip?download=1
## Preprocessing
We preprocess the logs using the Drain algorithm to extract structured fields and identify event types.
We then encode the logs using a pretrained tokenizer and add special tokens to separate event types. This
dataset should be immediately usable for training and testing models for log-based anomaly detection.
## Intended Uses
This dataset is designed for:
- Training log anomaly detection models
- Evaluating log sequence prediction models
- Benchmarking different approaches to log-based anomaly detection
see [honicky/pythia-14m-hdfs-logs](https://huggingface.co/honicky/pythia-14m-hdfs-logs) for an example model.
## Citation
If you use this dataset, please cite the original HDFS paper:
```bibtex
@inproceedings{xu2009detecting,
title={Detecting large-scale system problems by mining console logs},
author={Xu, Wei and Huang, Ling and Fox, Armando and Patterson, David and Jordan, Michael I},
booktitle={Proceedings of the ACM SIGOPS 22nd symposium on Operating systems principles},
pages={117--132},
year={2009}
}
```