Datasets:
Dataset Card for Patch the Planet
Dataset Description
This data was produced by ThinkOnward for the Patch the Planet Challenge, using a synthetic seismic dataset generator called Synthoseis. This dataset consists of 500 training volumes and 15 test volumes. You will also be provided with a training data generation code in the starter notebook to build the training data. This code allows experimentation with different-sized missing data volumes in the seismic data. The challenger can increase the percentage of the missing section in each seismic volume to increase the difficulty. The default missing section will be set to 25%.
- Created by: Mike McIntire at ThinkOnward
- License: CC 4.0
Uses
How to generate a dataset
This dataset is provided as whole seismic volumes. It is the users responsibility to generate the missing sections of the seismic volumes. Please follow the steps below to generate the missing sections of the seismic volumes.
Step 1: Load the seismic volume and convert from parquet to numpy array
import pandas as pd
import numpy as np
def parquet2array(parquet_file, original_shape=(300,300,1259)):
df = pd.read_parquet(parquet_file)
data_only = df.drop(columns=['Row', 'Col'])
# Convert the DataFrame back to a 2D numpy array
reshaped_array = data_only.values
# Reshape the 2D array back into a 3D array
array = reshaped_array.reshape(original_shape)
return array
Step 2: Generate the missing sections of the seismic volume. This code will delete a random section of the seismic volume and return the target region and the mask of the target region.
def training_data_generator(seismic: np.ndarray, axis: Literal['i_line', 'x_line', None]=None, percentile: int=25):
"""Function to delete part of original seismic volume and extract target region
Parameters:
seismic: np.ndarray 3D matrix with original survey
axis: one of 'i_line','x_line' or None. Axis along which part of survey will be deleted.
If None (default), random will be chosen
percentile: int, size of deleted part relative to axis. Any integer between 1 and 99 (default 20)
Returns:
seismic: np.ndarray, original survey 3D matrix with deleted region
target: np.ndarray, 3D deleted region
target_mask: np.ndarray, position of target 3D matrix in seismic 3D matrix.
This mask is used to reconstruct original survey -> seismic[target_mask]=target.reshape(-1)
"""
# check parameters
assert isinstance(seismic, np.ndarray) and len(seismic.shape)==3, 'seismic must be 3D numpy.ndarray'
assert axis in ['i_line', 'x_line', None], 'axis must be one of: i_line, x_line or None'
assert type(percentile) is int and 0<percentile<100, 'percentile must be an integer between 0 and 100'
# rescale volume
minval = np.percentile(seismic, 2)
maxval = np.percentile(seismic, 98)
seismic = np.clip(seismic, minval, maxval)
seismic = ((seismic - minval) / (maxval - minval)) * 255
# if axis is None get random choice
if axis is None:
axis = np.random.choice(['i_line', 'x_line'], 1)[0]
# crop subset
if axis == 'i_line':
sample_size = np.round(seismic.shape[0]*(percentile/100)).astype('int')
sample_start = np.random.choice(range(seismic.shape[0]-sample_size), 1)[0]
sample_end = sample_start+sample_size
target_mask = np.zeros(seismic.shape).astype('bool')
target_mask[sample_start:sample_end, :, :] = True
target = seismic[sample_start:sample_end, :, :].copy()
seismic[target_mask] = np.nan
else:
sample_size = np.round(seismic.shape[1]*(percentile/100)).astype('int')
sample_start = np.random.choice(range(seismic.shape[1]-sample_size), 1)[0]
sample_end = sample_start+sample_size
target_mask = np.zeros(seismic.shape).astype('bool')
target_mask[:, sample_start:sample_end, :] = True
target = seismic[:, sample_start:sample_end, :].copy()
seismic[target_mask] = np.nan
return seismic, target, target_mask
Dataset Structure
train (500 volumes)
- seismicCubes_RFC_fullstack_2023_1234567.parquet
- seismicCubes_RFC_fullstack_2023_1234568.parquet
- ...
- seismicCubes_RFC_fullstack_2023_1234568.parquet
test (15 volumes, 25% missing, target region provided)
- seismicCubes_RFC_fullstack_2023_1234567.parquet
- seismicCubes_RFC_fullstack_2023_1234568.parquet
- ...
- seismicCubes_RFC_fullstack_2023_1234568.parquet
Dataset Creation
Source Data
This data was produced by ThinkOnward for the Patch the Planet Challenge, using a synthetic seismic dataset generator called Synthoseis.
Who are the source data producers?
This data was produced by ThinkOnward for the Patch the Planet Challenge, using a synthetic seismic dataset generator called Synthoseis. The data is provided as whole seismic volumes. It is the users responsibility to generate the missing sections of the seismic volumes. using the provided code.
Recommendations
This is a synthetically generated dataset, and differs from real-world seismic data. It is recommended that this dataset be used for research purposes only.
Citation
This dataset was released in conjunction with the presentation of a poster at the 2024 IMAGE Conference in Houston, Texas (August 26-29th, 2024)
BibTeX:
@misc {thinkonward_2024, author = { {ThinkOnward} }, title = { patch-the-planet (Revision 5e94745) }, year = 2024, url = { https://huggingface.co/datasets/thinkonward/patch-the-planet }, doi = { 10.57967/hf/2909 }, publisher = { Hugging Face } }
APA:
McIntire, M., Tanovic, O., Mazura, J., Suurmeyer, N., & Pisel, J. (n.d.). Geophysical Foundation Model: Improving results with trace masking. In https://imageevent.aapg.org/portals/26/abstracts/2024/4092088.pdf. 2024 IMAGE Conference, Houston, United States of America.
Dataset Card Contact
Please contact [email protected]
for questions, comments, or concerns about this model.
- Downloads last month
- 116