farimafatahi commited on
Commit
2a17a1c
Β·
verified Β·
1 Parent(s): 8e86e3b

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +1 -1
app.py CHANGED
@@ -184,7 +184,7 @@ with tab1:
184
  <strong> 🎯 Factual Precision </strong> measures the ratio of supported units divided by all units averaged over model responses. <strong> πŸŒ€ Hallucination Score </strong> quantifies the incorrect or inconclusive contents within a model response, as described in the paper. We also provide statistics on the average length of the response in terms of the number of tokens, the average verifiable units existing in the model responses (<strong>Avg. # Units</strong>), the average number of units labelled as undecidable (<strong>Avg. # Undecidable</strong>), and the average number of units labelled as unsupported (<strong>Avg. # Unsupported</strong>).
185
  </p>
186
  <p>
187
- πŸ”’ for closed LLMs; πŸ”‘ for open-weights LLMs; 🚨 for newly added models"
188
  </p>
189
  </div>
190
  """,
 
184
  <strong> 🎯 Factual Precision </strong> measures the ratio of supported units divided by all units averaged over model responses. <strong> πŸŒ€ Hallucination Score </strong> quantifies the incorrect or inconclusive contents within a model response, as described in the paper. We also provide statistics on the average length of the response in terms of the number of tokens, the average verifiable units existing in the model responses (<strong>Avg. # Units</strong>), the average number of units labelled as undecidable (<strong>Avg. # Undecidable</strong>), and the average number of units labelled as unsupported (<strong>Avg. # Unsupported</strong>).
185
  </p>
186
  <p>
187
+ πŸ”’ for closed LLMs; πŸ”‘ for open-weights LLMs; 🚨 for newly added models
188
  </p>
189
  </div>
190
  """,