seanpedrickcase commited on
Commit
f0668f6
Β·
1 Parent(s): b7f4700

Changed readme sdk back to gradio, updated intro text

Browse files
Files changed (2) hide show
  1. README.md +2 -2
  2. app.py +1 -1
README.md CHANGED
@@ -3,7 +3,7 @@ title: Large language model topic modelling
3
  emoji: πŸ“
4
  colorFrom: purple
5
  colorTo: yellow
6
- sdk: 5.8.0
7
  app_file: app.py
8
  pinned: true
9
  license: cc-by-nc-4.0
@@ -11,7 +11,7 @@ license: cc-by-nc-4.0
11
 
12
  # Large language model topic modelling
13
 
14
- Extract topics and summarise outputs using Large Language Models (LLMs, Gemini Flash/Pro, or Claude 3 through AWS Bedrock if running on AWS). The app will query the LLM with batches of responses to produce summary tables, which are then compared iteratively to output a table with the general topics, subtopics, topic sentiment, and relevant text rows related to them. The prompts are designed for topic modelling public consultations, but they can be adapted to different contexts (see the LLM settings tab to modify).
15
 
16
  You can use an AWS Bedrock model (Claude 3, paid), or Gemini (a free API, but with strict limits for the Pro model). Due to the strict API limits for the best model (Pro 1.5), the use of Gemini requires an API key. To set up your own Gemini API key, go here: https://aistudio.google.com/app/u/1/plan_information.
17
 
 
3
  emoji: πŸ“
4
  colorFrom: purple
5
  colorTo: yellow
6
+ sdk: gradio
7
  app_file: app.py
8
  pinned: true
9
  license: cc-by-nc-4.0
 
11
 
12
  # Large language model topic modelling
13
 
14
+ Extract topics and summarise outputs using Large Language Models (LLMs, Gemma 2b instruct if local, Gemini Flash/Pro, or Claude 3 through AWS Bedrock if running on AWS). The app will query the LLM with batches of responses to produce summary tables, which are then compared iteratively to output a table with the general topics, subtopics, topic sentiment, and relevant text rows related to them. The prompts are designed for topic modelling public consultations, but they can be adapted to different contexts (see the LLM settings tab to modify).
15
 
16
  You can use an AWS Bedrock model (Claude 3, paid), or Gemini (a free API, but with strict limits for the Pro model). Due to the strict API limits for the best model (Pro 1.5), the use of Gemini requires an API key. To set up your own Gemini API key, go here: https://aistudio.google.com/app/u/1/plan_information.
17
 
app.py CHANGED
@@ -154,7 +154,7 @@ with app:
154
  gr.Markdown(
155
  """# Large language model topic modelling
156
 
157
- Extract topics and summarise outputs using Large Language Models (LLMs, Gemini Flash/Pro, or Claude 3 through AWS Bedrock if running on AWS). The app will query the LLM with batches of responses to produce summary tables, which are then compared iteratively to output a table with the general topics, subtopics, topic sentiment, and relevant text rows related to them. The prompts are designed for topic modelling public consultations, but they can be adapted to different contexts (see the LLM settings tab to modify). Instructions on use can be found in the README.md file.
158
 
159
  You can use an AWS Bedrock model (Claude 3, paid), or Gemini (a free API, but with strict limits for the Pro model). Due to the strict API limits for the best model (Pro 1.5), the use of Gemini requires an API key. To set up your own Gemini API key, go here: https://aistudio.google.com/app/u/1/plan_information.
160
 
 
154
  gr.Markdown(
155
  """# Large language model topic modelling
156
 
157
+ Extract topics and summarise outputs using Large Language Models (LLMs, Gemma 2b instruct if local, Gemini Flash/Pro, or Claude 3 through AWS Bedrock if running on AWS). The app will query the LLM with batches of responses to produce summary tables, which are then compared iteratively to output a table with the general topics, subtopics, topic sentiment, and relevant text rows related to them. The prompts are designed for topic modelling public consultations, but they can be adapted to different contexts (see the LLM settings tab to modify). Instructions on use can be found in the README.md file.
158
 
159
  You can use an AWS Bedrock model (Claude 3, paid), or Gemini (a free API, but with strict limits for the Pro model). Due to the strict API limits for the best model (Pro 1.5), the use of Gemini requires an API key. To set up your own Gemini API key, go here: https://aistudio.google.com/app/u/1/plan_information.
160