katielink commited on
Commit
268d2cb
·
1 Parent(s): 23f4506

Update README Formatting

Browse files
Files changed (3) hide show
  1. README.md +2 -5
  2. configs/metadata.json +2 -1
  3. docs/README.md +2 -5
README.md CHANGED
@@ -8,10 +8,8 @@ license: apache-2.0
8
  # Model Overview
9
  A pre-trained model for simultaneous segmentation and classification of nuclei within multi-tissue histology images based on CoNSeP data. The details of the model can be found in [1].
10
 
11
- ## Workflow
12
  The model is trained to simultaneously segment and classify nuclei. Training is done via a two-stage approach. First initialized the model with pre-trained weights on the [ImageNet dataset](https://ieeexplore.ieee.org/document/5206848), trained only the decoders for the first 50 epochs, and then fine-tuned all layers for another 50 epochs. There are two training modes in total. If "original" mode is specified, [270, 270] and [80, 80] are used for `patch_size` and `out_size` respectively. If "fast" mode is specified, [256, 256] and [164, 164] are used for `patch_size` and `out_size` respectively. The results shown below are based on the "fast" mode.
13
 
14
- ### Pre-trained weights
15
  The first stage is trained with pre-trained weights from some internal data.The [original author's repo](https://github.com/vqdang/hover_net#data-format) also provides pre-trained weights but for non-commercial use.
16
  Each user is responsible for checking the content of models/datasets and the applicable licenses and determining if suitable for the intended use.
17
 
@@ -29,6 +27,8 @@ The training data is from <https://warwick.ac.uk/fac/cross_fac/tia/data/hovernet
29
 
30
  The provided labelled data was partitioned, based on the original split, into training (27 tiles) and testing (14 tiles) datasets.
31
 
 
 
32
  After download the datasets, please run `scripts/prepare_patches.py` to prepare patches from tiles. Prepared patches are saved in `your-concep-dataset-path`/Prepared. The implementation is referring to <https://github.com/vqdang/hover_net/blob/master/extract_patches.py>. The command is like:
33
 
34
  ```
@@ -125,9 +125,6 @@ python -m monai.bundle run --config_file "['configs/train.json','configs/evaluat
125
  python -m monai.bundle run --config_file configs/inference.json
126
  ```
127
 
128
- # Disclaimer
129
- This is an example, not to be used for diagnostic purposes.
130
-
131
  # References
132
  [1] Simon Graham, Quoc Dang Vu, Shan E Ahmed Raza, Ayesha Azam, Yee Wah Tsang, Jin Tae Kwak, Nasir Rajpoot, Hover-Net: Simultaneous segmentation and classification of nuclei in multi-tissue histology images, Medical Image Analysis, 2019 https://doi.org/10.1016/j.media.2019.101563
133
 
 
8
  # Model Overview
9
  A pre-trained model for simultaneous segmentation and classification of nuclei within multi-tissue histology images based on CoNSeP data. The details of the model can be found in [1].
10
 
 
11
  The model is trained to simultaneously segment and classify nuclei. Training is done via a two-stage approach. First initialized the model with pre-trained weights on the [ImageNet dataset](https://ieeexplore.ieee.org/document/5206848), trained only the decoders for the first 50 epochs, and then fine-tuned all layers for another 50 epochs. There are two training modes in total. If "original" mode is specified, [270, 270] and [80, 80] are used for `patch_size` and `out_size` respectively. If "fast" mode is specified, [256, 256] and [164, 164] are used for `patch_size` and `out_size` respectively. The results shown below are based on the "fast" mode.
12
 
 
13
  The first stage is trained with pre-trained weights from some internal data.The [original author's repo](https://github.com/vqdang/hover_net#data-format) also provides pre-trained weights but for non-commercial use.
14
  Each user is responsible for checking the content of models/datasets and the applicable licenses and determining if suitable for the intended use.
15
 
 
27
 
28
  The provided labelled data was partitioned, based on the original split, into training (27 tiles) and testing (14 tiles) datasets.
29
 
30
+ ### Preprocessing
31
+
32
  After download the datasets, please run `scripts/prepare_patches.py` to prepare patches from tiles. Prepared patches are saved in `your-concep-dataset-path`/Prepared. The implementation is referring to <https://github.com/vqdang/hover_net/blob/master/extract_patches.py>. The command is like:
33
 
34
  ```
 
125
  python -m monai.bundle run --config_file configs/inference.json
126
  ```
127
 
 
 
 
128
  # References
129
  [1] Simon Graham, Quoc Dang Vu, Shan E Ahmed Raza, Ayesha Azam, Yee Wah Tsang, Jin Tae Kwak, Nasir Rajpoot, Hover-Net: Simultaneous segmentation and classification of nuclei in multi-tissue histology images, Medical Image Analysis, 2019 https://doi.org/10.1016/j.media.2019.101563
130
 
configs/metadata.json CHANGED
@@ -1,7 +1,8 @@
1
  {
2
  "schema": "https://github.com/Project-MONAI/MONAI-extra-test-data/releases/download/0.8.1/meta_schema_hovernet_20221124.json",
3
- "version": "0.1.6",
4
  "changelog": {
 
5
  "0.1.6": "add non-deterministic note",
6
  "0.1.5": "update benchmark on A100",
7
  "0.1.4": "adapt to BundleWorkflow interface",
 
1
  {
2
  "schema": "https://github.com/Project-MONAI/MONAI-extra-test-data/releases/download/0.8.1/meta_schema_hovernet_20221124.json",
3
+ "version": "0.1.7",
4
  "changelog": {
5
+ "0.1.7": "Update README Formatting",
6
  "0.1.6": "add non-deterministic note",
7
  "0.1.5": "update benchmark on A100",
8
  "0.1.4": "adapt to BundleWorkflow interface",
docs/README.md CHANGED
@@ -1,10 +1,8 @@
1
  # Model Overview
2
  A pre-trained model for simultaneous segmentation and classification of nuclei within multi-tissue histology images based on CoNSeP data. The details of the model can be found in [1].
3
 
4
- ## Workflow
5
  The model is trained to simultaneously segment and classify nuclei. Training is done via a two-stage approach. First initialized the model with pre-trained weights on the [ImageNet dataset](https://ieeexplore.ieee.org/document/5206848), trained only the decoders for the first 50 epochs, and then fine-tuned all layers for another 50 epochs. There are two training modes in total. If "original" mode is specified, [270, 270] and [80, 80] are used for `patch_size` and `out_size` respectively. If "fast" mode is specified, [256, 256] and [164, 164] are used for `patch_size` and `out_size` respectively. The results shown below are based on the "fast" mode.
6
 
7
- ### Pre-trained weights
8
  The first stage is trained with pre-trained weights from some internal data.The [original author's repo](https://github.com/vqdang/hover_net#data-format) also provides pre-trained weights but for non-commercial use.
9
  Each user is responsible for checking the content of models/datasets and the applicable licenses and determining if suitable for the intended use.
10
 
@@ -22,6 +20,8 @@ The training data is from <https://warwick.ac.uk/fac/cross_fac/tia/data/hovernet
22
 
23
  The provided labelled data was partitioned, based on the original split, into training (27 tiles) and testing (14 tiles) datasets.
24
 
 
 
25
  After download the datasets, please run `scripts/prepare_patches.py` to prepare patches from tiles. Prepared patches are saved in `your-concep-dataset-path`/Prepared. The implementation is referring to <https://github.com/vqdang/hover_net/blob/master/extract_patches.py>. The command is like:
26
 
27
  ```
@@ -118,9 +118,6 @@ python -m monai.bundle run --config_file "['configs/train.json','configs/evaluat
118
  python -m monai.bundle run --config_file configs/inference.json
119
  ```
120
 
121
- # Disclaimer
122
- This is an example, not to be used for diagnostic purposes.
123
-
124
  # References
125
  [1] Simon Graham, Quoc Dang Vu, Shan E Ahmed Raza, Ayesha Azam, Yee Wah Tsang, Jin Tae Kwak, Nasir Rajpoot, Hover-Net: Simultaneous segmentation and classification of nuclei in multi-tissue histology images, Medical Image Analysis, 2019 https://doi.org/10.1016/j.media.2019.101563
126
 
 
1
  # Model Overview
2
  A pre-trained model for simultaneous segmentation and classification of nuclei within multi-tissue histology images based on CoNSeP data. The details of the model can be found in [1].
3
 
 
4
  The model is trained to simultaneously segment and classify nuclei. Training is done via a two-stage approach. First initialized the model with pre-trained weights on the [ImageNet dataset](https://ieeexplore.ieee.org/document/5206848), trained only the decoders for the first 50 epochs, and then fine-tuned all layers for another 50 epochs. There are two training modes in total. If "original" mode is specified, [270, 270] and [80, 80] are used for `patch_size` and `out_size` respectively. If "fast" mode is specified, [256, 256] and [164, 164] are used for `patch_size` and `out_size` respectively. The results shown below are based on the "fast" mode.
5
 
 
6
  The first stage is trained with pre-trained weights from some internal data.The [original author's repo](https://github.com/vqdang/hover_net#data-format) also provides pre-trained weights but for non-commercial use.
7
  Each user is responsible for checking the content of models/datasets and the applicable licenses and determining if suitable for the intended use.
8
 
 
20
 
21
  The provided labelled data was partitioned, based on the original split, into training (27 tiles) and testing (14 tiles) datasets.
22
 
23
+ ### Preprocessing
24
+
25
  After download the datasets, please run `scripts/prepare_patches.py` to prepare patches from tiles. Prepared patches are saved in `your-concep-dataset-path`/Prepared. The implementation is referring to <https://github.com/vqdang/hover_net/blob/master/extract_patches.py>. The command is like:
26
 
27
  ```
 
118
  python -m monai.bundle run --config_file configs/inference.json
119
  ```
120
 
 
 
 
121
  # References
122
  [1] Simon Graham, Quoc Dang Vu, Shan E Ahmed Raza, Ayesha Azam, Yee Wah Tsang, Jin Tae Kwak, Nasir Rajpoot, Hover-Net: Simultaneous segmentation and classification of nuclei in multi-tissue histology images, Medical Image Analysis, 2019 https://doi.org/10.1016/j.media.2019.101563
123