--- license: cc-by-sa-4.0 language: - en --- # phytoClassUCSC Sections and prompts from the [model cards paper](https://arxiv.org/abs/1810.03993), v2. Jump to section: - [Model details](#model-details) - [Intended use](#intended-use) - [Factors](#factors) - [Metrics](#metrics) - [Evaluation data](#evaluation-data) - [Training data](#training-data) - [Quantitative analyses](#quantitative-analyses) - [Ethical considerations](#ethical-considerations) - [Caveats and recommendations](#caveats-and-recommendations) ## Model details Review section 4.1 of the [model cards paper](https://arxiv.org/abs/1810.03993). - Developed by the Kudela Lab from the Ocean Sciences Department at University of California, Santa Cruz. - Current version trained in February, 2023. - phytoClassUCSC-SoftNone02162023 - phytoClassUCSC is a depthwise- CNN based on the Xception architecture [Chollet, F., 2017](https://arxiv.org/abs/1610.02357) with 134 layers using weights pretrained on ImageNet. - An average pooling layer is used. - Paper or other resource for more information - Citation details - License - Email Patrick Daniel ([pcdaniel@ucsc.edu](pcdaniel@ucsc.edu)) for questions ## Intended use _Use cases that were envisioned during development._ Review section 4.2 of the [model cards paper](https://arxiv.org/abs/1810.03993). ### Primary intended uses ### Primary intended users ### Out-of-scope use cases ## Factors _Factors could include demographic or phenotypic groups, environmental conditions, technical attributes, or others listed in Section 4.3._ Review section 4.3 of the [model cards paper](https://arxiv.org/abs/1810.03993). ### Relevant factors ### Evaluation factors ## Metrics _The appropriate metrics to feature in a model card depend on the type of model that is being tested. For example, classification systems in which the primary output is a class label differ significantly from systems whose primary output is a score. In all cases, the reported metrics should be determined based on the model’s structure and intended use._ Review section 4.4 of the [model cards paper](https://arxiv.org/abs/1810.03993). ### Model performance measures ### Decision thresholds ### Approaches to uncertainty and variability ## Evaluation data _All referenced datasets would ideally point to any set of documents that provide visibility into the source and composition of the dataset. Evaluation datasets should include datasets that are publicly available for third-party use. These could be existing datasets or new ones provided alongside the model card analyses to enable further benchmarking._ Review section 4.5 of the [model cards paper](https://arxiv.org/abs/1810.03993). ### Datasets ### Motivation ### Preprocessing ## Training data Review section 4.6 of the [model cards paper](https://arxiv.org/abs/1810.03993). ## Quantitative analyses _Quantitative analyses should be disaggregated, that is, broken down by the chosen factors. Quantitative analyses should provide the results of evaluating the model according to the chosen metrics, providing confidence interval values when possible._ Review section 4.7 of the [model cards paper](https://arxiv.org/abs/1810.03993). ### Unitary results ### Intersectional result ## Ethical considerations _This section is intended to demonstrate the ethical considerations that went into model development, surfacing ethical challenges and solutions to stakeholders. Ethical analysis does not always lead to precise solutions, but the process of ethical contemplation is worthwhile to inform on responsible practices and next steps in future work._ Review section 4.8 of the [model cards paper](https://arxiv.org/abs/1810.03993). ### Data ### Human life ### Mitigations ### Risks and harms ### Use cases ## Caveats and recommendations _This section should list additional concerns that were not covered in the previous sections._ Review section 4.9 of the [model cards paper](https://arxiv.org/abs/1810.03993).