English
phytoClassUCSC / README.md
Patrick Daniel
updated readme
d6d27e6
metadata
license: cc-by-sa-4.0
language:
  - en

phytoClassUCSC - A phytoplankton classifier for IFCB data

Note: Sections and prompts from the model cards paper, v2.

Jump to section:

Model details

  • Developed by the Kudela Lab from the Ocean Sciences Department at University of California, Santa Cruz.
  • Current version trained in February, 2023.
  • Version 1.0
  • phytoClassUCSC is a depthwise- CNN based on the Xception architecture Chollet, F., 2017 with 134 layers using weights pretrained on ImageNet.
  • An average pooling layer is used.
  • Licensed under CC-BY-SA-4.0
  • For Questions email Patrick Daniel ([email protected])

Intended use

This model was designed and trained to work with IFCB data generated in Monterey Bay. While that does not mean it may not perform well in other locations, the distribution of training images reflects common phytoplankton observed at the Santa Cruz Wharf and Power Buoy locations.

Independent model validation should be used when applying the model to other sites.

Review section 4.2 of the model cards paper.

Primary intended uses

Generalized phytoplankton classifier for common taxa found in the Monterey Bay. This

Primary intended users

Researchers intersted in a general.

Out-of-scope use cases

Observing and identifying rare or non-endemic taxa.

Factors

Model classes were chosen based on common and resolvable phytoplankton taxa. Taxonomic groupings were chosen based on what researchers in the lab felt groups that could be confidently identified, given the expertise and research intersts of the lab.

Instrument

Model was trained on images from Imaging FlowCytobot (IFCB) instruments primary deployed at the Santa Cruz Wharf and the Monterey Bay Aquarium Research Institute (MBARI) Power Buoy. The Santa Cruz Wharf IFCB (#104) is an early generation

Review section 4.3 of the model cards paper.

Relevant factors

Evaluation factors

Metrics

The appropriate metrics to feature in a model card depend on the type of model that is being tested. For example, classification systems in which the primary output is a class label differ significantly from systems whose primary output is a score. In all cases, the reported metrics should be determined based on the model’s structure and intended use.

Review section 4.4 of the model cards paper.

Model performance measures

Decision thresholds

Approaches to uncertainty and variability

Evaluation data

All referenced datasets would ideally point to any set of documents that provide visibility into the source and composition of the dataset. Evaluation datasets should include datasets that are publicly available for third-party use. These could be existing datasets or new ones provided alongside the model card analyses to enable further benchmarking.

Review section 4.5 of the model cards paper.

Datasets

Motivation

Preprocessing

Training data

Review section 4.6 of the model cards paper.

Quantitative analyses

Quantitative analyses should be disaggregated, that is, broken down by the chosen factors. Quantitative analyses should provide the results of evaluating the model according to the chosen metrics, providing confidence interval values when possible.

Review section 4.7 of the model cards paper.

Unitary results

Intersectional result

Ethical considerations

None

Data

Use cases

Caveats and recommendations

This section should list additional concerns that were not covered in the previous sections.

Review section 4.9 of the model cards paper.