Nevil9 commited on
Commit
f06dcc5
1 Parent(s): d7e678a

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +41 -5
README.md CHANGED
@@ -1,6 +1,42 @@
1
- PsyLlama: A Conversational AI for Mental Health Assessment
2
 
3
- Model Name: PsyLlama
4
- Model Architecture: LLaMA-based model (fine-tuned)
5
- Model Type: Instruct-tuned, conversational AI model
6
- Primary Use: Mental health assessment through psychometric analysis
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # PsyLlama: A Conversational AI for Mental Health Assessment
2
 
3
+ **Model Name**: `PsyLlama`
4
+ **Model Architecture**: LLaMA-based model (fine-tuned)
5
+ **Model Type**: Instruct-tuned, conversational AI model
6
+ **Primary Use**: Mental health assessment through psychometric analysis
7
+
8
+ ---
9
+
10
+ ### Model Description
11
+
12
+ **PsyLlama** is a conversational AI model based on LLaMA architecture, fine-tuned for mental health assessments. It is designed to assist healthcare professionals in conducting initial psychometric evaluations and mental health assessments by generating context-aware conversational responses. The model uses structured questions and answers to assess patients' mental states and supports clinical decision-making in telemedicine environments.
13
+
14
+ **Applications**:
15
+ - Psychometric evaluation
16
+ - Mental health chatbot
17
+ - Symptom analysis for mental health assessment
18
+
19
+ ---
20
+
21
+ ### Model Usage
22
+
23
+ To use **PsyLlama**, you can load it from Hugging Face using the `transformers` library. Below is a code snippet showing how to initialize and use the model:
24
+
25
+ ```python
26
+ from transformers import AutoModelForCausalLM, AutoTokenizer
27
+
28
+ # Load the model and tokenizer from Hugging Face
29
+ model_name = "Nevil9/PsyLlama"
30
+ tokenizer = AutoTokenizer.from_pretrained(model_name)
31
+ model = AutoModelForCausalLM.from_pretrained(model_name)
32
+
33
+ # Example input
34
+ input_text = "How are you feeling today? Have you been experiencing any anxiety or stress?"
35
+
36
+ # Tokenize input and generate response
37
+ inputs = tokenizer(input_text, return_tensors="pt")
38
+ output = model.generate(**inputs, max_length=100)
39
+
40
+ # Decode and print the response
41
+ response = tokenizer.decode(output[0], skip_special_tokens=True)
42
+ print(response)