Spaces:
Runtime error
Runtime error
metadata
title: T5S
emoji: π―
colorFrom: yellow
colorTo: red
sdk: streamlit
app_file: src/visualization/visualize.py
pinned: false
t5s
T5 Summarisation Using Pytorch Lightning
Usage
To use and run the DVC pipeline install the t5s
package
pip install t5s
Firstly we need to clone the repo containing the code so we can do that using:
t5s clone
We would then have to create the required directories to run the pipeline
t5s dirs
Then we need to pull the models from DVC
t5s pull
Now to run the training pipeline we can run:
t5s run
Finally to push the model to DVC
t5s push
To push this model to HuggingFace Hub for inference you can run:
t5s push_to_hf_hub
Next if we would like to test the model and visualise the results we can run:
t5s visualize
And this would create a streamlit app for testing
Project Organization
βββ LICENSE
βββ Makefile <- Makefile with commands like `make dirs` or `make clean`
βββ README.md <- The top-level README for developers using this project.
βββ data
β βββ processed <- The final, canonical data sets for modeling.
β βββ raw <- The original, immutable data dump.
β
βββ models <- Trained and serialized models, model predictions, or model summaries
β
βββ notebooks <- Jupyter notebooks. Naming convention is a number (for ordering),
β the creator's initials, and a short `-` delimited description, e.g.
β `1.0-jqp-initial-data-exploration`.
βββ references <- Data dictionaries, manuals, and all other explanatory materials.
β
βββ reports <- Generated analysis as HTML, PDF, LaTeX, etc.
β βββ metrics.txt <- Relevant metrics after evaluating the model.
β βββ training_metrics.txt <- Relevant metrics from training the model.
β
βββ requirements.txt <- The requirements file for reproducing the analysis environment
β
βββ setup.py <- makes project pip installable (pip install -e .) so src can be imported
βββ src <- Source code for use in this project.
β βββ __init__.py <- Makes src a Python module
β β
β βββ data <- Scripts to download or generate data
β β βββ make_dataset.py
β β βββ process_data.py
β β
β βββ models <- Scripts to train models
β β βββ predict_model.py
β β βββ train_model.py
β β βββ evaluate_model.py
β β βββ model.py
β β
β βββ visualization <- Scripts to create exploratory and results oriented visualizations
β βββ visualize.py
β
βββ tox.ini <- tox file with settings for running tox; see tox.testrun.org
βββ data.dvc <- Traing a model on the processed data.