Tijmen2 commited on
Commit
06c8985
1 Parent(s): 4e04e79

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +3 -2
README.md CHANGED
@@ -4,6 +4,7 @@ language:
4
  - en
5
  datasets:
6
  - Open-Orca/SlimOrca-Dedup
 
7
  ---
8
  # cosmosage
9
 
@@ -53,6 +54,6 @@ ASSISTANT:
53
 
54
  cosmosage_v0.2 performs much better than cosmosage_v0.1. While v0.1 did not seem to have picked up much knowledge from the ArXiV papers it was trained on, v0.2 can often give surprisingly good answers to highly technical questions about cosmology. It gives certain answers which it could not have known without having read these recent papers, leading me to conclude that it has learned knowledge from the ArXiV papers.
55
 
56
- I've also been impressed by cosmosage's knowledge about astronomy, as well as other branches of physics. However, in these areas it is less clear how much the performance is due to the pretraining of the Mistral model versus the fine-tuning I did. I included quite a bit of data from older textbooks, so it can make mistakes. For example, it sometimes thinks Pluto is a planet.
57
 
58
- Currently, the main challenge of cosmosage_v0.2 lies in its _reliability_. In many of its answers it confidently makes incorrect statements that appear to be completely made up.
 
4
  - en
5
  datasets:
6
  - Open-Orca/SlimOrca-Dedup
7
+ pipeline_tag: question-answering
8
  ---
9
  # cosmosage
10
 
 
54
 
55
  cosmosage_v0.2 performs much better than cosmosage_v0.1. While v0.1 did not seem to have picked up much knowledge from the ArXiV papers it was trained on, v0.2 can often give surprisingly good answers to highly technical questions about cosmology. It gives certain answers which it could not have known without having read these recent papers, leading me to conclude that it has learned knowledge from the ArXiV papers.
56
 
57
+ I've also been impressed by cosmosage's knowledge about astronomy, as well as other branches of physics. However, in these areas it is less clear how much the performance is due to the pretraining of the Mistral model versus the fine-tuning I did.
58
 
59
+ Currently, the main area where cosmosage_v0.2 needs to be improved is its _reliability_. In many of its answers it confidently makes incorrect statements that appear to be completely made up. This means that the outputs of cosmosage_v0.2 should not be trusted but rather thought of as potential explanations, food for thought, or inspiration.