Quantization made by Richard Erkhov. [Github](https://github.com/RichardErkhov) [Discord](https://discord.gg/pvy7H8DZMG) [Request more models](https://github.com/RichardErkhov/quant_request) opt-350m-instruct - AWQ - Model creator: https://huggingface.co/nnpy/ - Original model: https://huggingface.co/nnpy/opt-350m-instruct/ Original model description: --- license: apache-2.0 datasets: - openchat/openchat_sharegpt4_dataset --- ## Usage ``` from transformers import AutoTokenizer, AutoModelForCausalLM tok = AutoTokenizer.from_pretrained('facebook/opt-350m') model = AutoModelForCausalLM.from_pretrained('prasanna2003/opt-350m-instruct') system_message = "You are AI language model helps the human." input_prompt = "Define data science." prompt = '' + system_message + '' + input_prompt + '' prompt = tokenizer(prompt, return_tensors='pt') out = model.generate(**prompt, max_length=120) print(tok.decode(out[0])) ```