Abhaykoul commited on
Commit
193dea9
·
verified ·
1 Parent(s): e027fc4

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -3
README.md CHANGED
@@ -46,7 +46,7 @@ HelpingAI-6B has achieved an impressive Emotional Quotient (EQ) of 91.57, making
46
  import torch
47
  from transformers import AutoModelForCausalLM, AutoTokenizer
48
 
49
- # Load the HelpingAI-3B-coder model
50
  model = AutoModelForCausalLM.from_pretrained("OEvortex/HelpingAI-6B", trust_remote_code=True)
51
  # Load the tokenizer
52
  tokenizer = AutoTokenizer.from_pretrained("OEvortex/HelpingAI-6B", trust_remote_code=True)
@@ -75,8 +75,6 @@ outputs = model.generate(
75
  eos_token_id=tokenizer.eos_token_id,
76
  )
77
 
78
- # Decode the generated text
79
- output = tokenizer.decode(generated_text[0], skip_special_tokens=True)
80
 
81
  response = outputs[0][inputs.shape[-1]:]
82
  print(tokenizer.decode(response, skip_special_tokens=True))
 
46
  import torch
47
  from transformers import AutoModelForCausalLM, AutoTokenizer
48
 
49
+ # Load the HelpingAI-6B model
50
  model = AutoModelForCausalLM.from_pretrained("OEvortex/HelpingAI-6B", trust_remote_code=True)
51
  # Load the tokenizer
52
  tokenizer = AutoTokenizer.from_pretrained("OEvortex/HelpingAI-6B", trust_remote_code=True)
 
75
  eos_token_id=tokenizer.eos_token_id,
76
  )
77
 
 
 
78
 
79
  response = outputs[0][inputs.shape[-1]:]
80
  print(tokenizer.decode(response, skip_special_tokens=True))