DSBench / data_modeling /data /task /playground-series-s3e17.txt
liqiang888's picture
Upload 738 files
fe8d248 verified
raw
history blame
2.53 kB
Description
Welcome to the 2023 edition of Kaggle's Playground Series! Thank you to everyone who participated in and contributed to Season 3 Playground Series so far! With the same goal to give the Kaggle community a variety of fairly light-weight challenges that can be used to learn and sharpen skills in different aspects of machine learning and data science, we will continue launching the Tabular Tuesday in June every Tuesday 00:00 UTC, with each competition running for 2 weeks. Again, these will be fairly light-weight datasets that are synthetically generated from real-world data and will provide an opportunity to quickly iterate through various model and feature engineering ideas, create visualizations, etc.
Synthetically-Generated Datasets
Using synthetic data for Playground competitions allows us to strike a balance between having real-world data (with named features) and ensuring test labels are not publicly available. This allows us to host competitions with more interesting datasets than in the past. While there are still challenges with synthetic data generation, the state-of-the-art is much better now than when we started the Tabular Playground Series two years ago, and that goal is to produce datasets that have far fewer artifacts. Please feel free to give us feedback on the datasets for the different competitions so that we can continue to improve!
Evaluation
Submissions are evaluated on area under the ROC curve between the predicted probability and the observed target.
Submission File
For each id in the test set, you must predict the probability of a Machine failure. The file should contain a header and have the following format:
id, Machine failure
136429, 0.5
136430, 0.1
136431, 0.9
etc.
Dataset Description
The dataset for this competition (both train and test) was generated from a deep learning model trained on the Machine Failure Predictions. Feature distributions are close to, but not exactly the same, as the original. Feel free to use the original dataset as part of this competition, both to explore differences as well as to see whether incorporating the original in training improves model performance.
Files
- train.csv - The training dataset; Machine failure is the (binary) target (which, in order to align with the ordering of the original dataset, is not in the last column position)
- test.csv - The test dataset; your objective is to predict the probability of Machine failure
- sample_submission.csv - A sample submission file in the correct format