s-a-malik commited on
Commit
f3099db
·
1 Parent(s): 3dc5f5e

add paper links

Browse files
Files changed (1) hide show
  1. app.py +9 -1
app.py CHANGED
@@ -13,6 +13,7 @@ from transformers import AutoModelForCausalLM, AutoTokenizer
13
  # TODO Sentence level highlighting instead (prediction after every word is not what it was trained on). Also solves token-level highlighting issues.
14
  # TODO log prob output scaling highlighting instead?
15
  # TODO make it look nicer
 
16
  # TODO streaming output (need custom generation function because of probes)
17
  # TODO add options to switch between models, SLT/TBG, layers?
18
  # TODO full semantic entropy calculation
@@ -23,12 +24,18 @@ MAX_INPUT_TOKEN_LENGTH = int(os.getenv("MAX_INPUT_TOKEN_LENGTH", "4096"))
23
 
24
  DESCRIPTION = """
25
  <h1>Llama-2 7B Chat with Uncertainty Probes</h1>
26
- <p>This Space demonstrates the Llama-2-7b-chat model with a semantic uncertainty probe.</p>
27
  <p>The highlighted text shows the model's uncertainty in real-time:</p>
28
  <ul>
29
  <li><span style="background-color: #00FF00; color: black">Green</span> indicates more certain generations</li>
30
  <li><span style="background-color: #FF0000; color: black">Red</span> indicates more uncertain generations</li>
31
  </ul>
 
 
 
 
 
 
32
  """
33
 
34
  EXAMPLES = [
@@ -101,6 +108,7 @@ def generate(
101
  se_highlighted_text = ""
102
  acc_highlighted_text = ""
103
 
 
104
  for i in range(1, len(hidden)):
105
 
106
  # Semantic Uncertainty Probe
 
13
  # TODO Sentence level highlighting instead (prediction after every word is not what it was trained on). Also solves token-level highlighting issues.
14
  # TODO log prob output scaling highlighting instead?
15
  # TODO make it look nicer
16
+ # TODO better examples.
17
  # TODO streaming output (need custom generation function because of probes)
18
  # TODO add options to switch between models, SLT/TBG, layers?
19
  # TODO full semantic entropy calculation
 
24
 
25
  DESCRIPTION = """
26
  <h1>Llama-2 7B Chat with Uncertainty Probes</h1>
27
+ <p>This Space demonstrates the Llama-2-7b-chat model with a semantic uncertainty probe. This demo is based on our paper: <a href="https://arxiv.org/abs/2406.15927" target="_blank">"Semantic Uncertainty and Accuracy Probes for Language Models"</a> by Jannik Kossen*, Jiatong Han*, Muhammed Razzak*, Lisa Schut, Shreshth Malik and Yarin Gal.</p>
28
  <p>The highlighted text shows the model's uncertainty in real-time:</p>
29
  <ul>
30
  <li><span style="background-color: #00FF00; color: black">Green</span> indicates more certain generations</li>
31
  <li><span style="background-color: #FF0000; color: black">Red</span> indicates more uncertain generations</li>
32
  </ul>
33
+ <p>The demo compares the model's uncertainty with two different probes:</p>
34
+ <ul>
35
+ <li><b>Semantic Uncertainty Probe:</b> Predicts the semantic uncertainty of the model's generations.</li>
36
+ <li><b>Accuracy Probe:</b> Predicts the accuracy of the model's generations.</li>
37
+ </ul>
38
+ <p>Please see our paper for more details.</p>
39
  """
40
 
41
  EXAMPLES = [
 
108
  se_highlighted_text = ""
109
  acc_highlighted_text = ""
110
 
111
+ # skip the first hidden state as it is the prompt
112
  for i in range(1, len(hidden)):
113
 
114
  # Semantic Uncertainty Probe