bytetrack / tools /mix_data_test_mot17.py
AK391
all files
7734d5b
raw
history blame
3.92 kB
import json
import os
"""
cd datasets
mkdir -p mix_det/annotations
cp mot/annotations/val_half.json mix_det/annotations/val_half.json
cp mot/annotations/test.json mix_det/annotations/test.json
cd mix_det
ln -s ../mot/train mot_train
ln -s ../crowdhuman/CrowdHuman_train crowdhuman_train
ln -s ../crowdhuman/CrowdHuman_val crowdhuman_val
ln -s ../Cityscapes cp_train
ln -s ../ETHZ ethz_train
cd ..
"""
mot_json = json.load(open('datasets/mot/annotations/train_half.json','r'))
img_list = list()
for img in mot_json['images']:
img['file_name'] = 'mot_train/' + img['file_name']
img_list.append(img)
ann_list = list()
for ann in mot_json['annotations']:
ann_list.append(ann)
video_list = mot_json['videos']
category_list = mot_json['categories']
print('mot17')
max_img = 10000
max_ann = 2000000
max_video = 10
crowdhuman_json = json.load(open('datasets/crowdhuman/annotations/train.json','r'))
img_id_count = 0
for img in crowdhuman_json['images']:
img_id_count += 1
img['file_name'] = 'crowdhuman_train/' + img['file_name']
img['frame_id'] = img_id_count
img['prev_image_id'] = img['id'] + max_img
img['next_image_id'] = img['id'] + max_img
img['id'] = img['id'] + max_img
img['video_id'] = max_video
img_list.append(img)
for ann in crowdhuman_json['annotations']:
ann['id'] = ann['id'] + max_ann
ann['image_id'] = ann['image_id'] + max_img
ann_list.append(ann)
print('crowdhuman_train')
video_list.append({
'id': max_video,
'file_name': 'crowdhuman_train'
})
max_img = 30000
max_ann = 10000000
crowdhuman_val_json = json.load(open('datasets/crowdhuman/annotations/val.json','r'))
img_id_count = 0
for img in crowdhuman_val_json['images']:
img_id_count += 1
img['file_name'] = 'crowdhuman_val/' + img['file_name']
img['frame_id'] = img_id_count
img['prev_image_id'] = img['id'] + max_img
img['next_image_id'] = img['id'] + max_img
img['id'] = img['id'] + max_img
img['video_id'] = max_video
img_list.append(img)
for ann in crowdhuman_val_json['annotations']:
ann['id'] = ann['id'] + max_ann
ann['image_id'] = ann['image_id'] + max_img
ann_list.append(ann)
print('crowdhuman_val')
video_list.append({
'id': max_video,
'file_name': 'crowdhuman_val'
})
max_img = 40000
max_ann = 20000000
ethz_json = json.load(open('datasets/ETHZ/annotations/train.json','r'))
img_id_count = 0
for img in ethz_json['images']:
img_id_count += 1
img['file_name'] = 'ethz_train/' + img['file_name'][5:]
img['frame_id'] = img_id_count
img['prev_image_id'] = img['id'] + max_img
img['next_image_id'] = img['id'] + max_img
img['id'] = img['id'] + max_img
img['video_id'] = max_video
img_list.append(img)
for ann in ethz_json['annotations']:
ann['id'] = ann['id'] + max_ann
ann['image_id'] = ann['image_id'] + max_img
ann_list.append(ann)
print('ETHZ')
video_list.append({
'id': max_video,
'file_name': 'ethz'
})
max_img = 50000
max_ann = 25000000
cp_json = json.load(open('datasets/Cityscapes/annotations/train.json','r'))
img_id_count = 0
for img in cp_json['images']:
img_id_count += 1
img['file_name'] = 'cp_train/' + img['file_name'][11:]
img['frame_id'] = img_id_count
img['prev_image_id'] = img['id'] + max_img
img['next_image_id'] = img['id'] + max_img
img['id'] = img['id'] + max_img
img['video_id'] = max_video
img_list.append(img)
for ann in cp_json['annotations']:
ann['id'] = ann['id'] + max_ann
ann['image_id'] = ann['image_id'] + max_img
ann_list.append(ann)
print('Cityscapes')
video_list.append({
'id': max_video,
'file_name': 'cityperson'
})
mix_json = dict()
mix_json['images'] = img_list
mix_json['annotations'] = ann_list
mix_json['videos'] = video_list
mix_json['categories'] = category_list
json.dump(mix_json, open('datasets/mix_det/annotations/train.json','w'))