library_name: transformers
base_model: meta-llama/Meta-Llama-3-8B
tags:
- biology
- medical
Model Details
Protein-Llama-3-8B is a specialized version of the Llama-3-8B language model, fine-tuned for the task of protein language modeling. This model has been continually pre-trained using LoRA technique on extensive datasets of protein sequences, enabling it to generate novel protein sequences based on natural language prompts. It supports both uncontrollable and controllable protein generation, allowing users to specify desired characteristics for the proteins. The model is designed to facilitate advancements in protein engineering, making it a valuable tool for drug development, chemical synthesis, and other biotechnological applications. For full details please read our paper.
Model Description
Generating novel protein sequences possessing desired properties, termed as protein engineering, is crucial for industries like drug development and chemical synthesis. Traditional protein engineering techniques often involve introducing random mutations into the gene encoding the protein of interest. This is followed by expression and screening to identify variants with improved or novel functions, which are then reproduced. While effective, these approaches are labor-intensive and time-consuming, as they rely on iterating over known protein sequences. This limits their ability to generate diverse protein sequences with entirely new capabilities, as they are constrained by existing protein templates. Moreover, the need to analyze numerous protein variants can waste valuable experimental resources. However, leveraging a Large Language Model (LLM) that has learned the "protein language" significantly accelerates this process. An LLM can generate and evaluate protein sequences in a matter of seconds. The inherent randomness of LLM-generated sequences enhances diversity, enabling the creation of completely novel proteins with potentially unprecedented functions. This not only streamlines the discovery and development process but also expands the scope of possibilities in protein engineering. This model is based on the Llama-3-8B architecture and is capable of generating proteins based on user defined characteristics.
Usage
To download and use the Protein-Llama-3 model for inference, follow these steps:
Installation
Ensure you have the transformers
library installed. You can install it using pip:
pip install transformers
Uncontrollable Generation
Uncontrollable generation can be handled via prompting the model with the phrase 'Seq=<'.
generator = pipeline('text-generation', model="Esperanto/Protein-Llama-3-8B")
sequences = generator("Seq=<",temperature=0.2,
top_k=40,
top_p=0.9,
do_sample=True,
repetition_penalty=1.2,
max_new_tokens=30,
num_return_sequences=500)
for sequence in sequences:
print(sequence['generated_text'])
Controllable Generation
Controllable generation can be done by prompting the model with '[Generate xxx protein] Seq=<'. Here, xxx can be any family from the 10 classes supported by this model.
generator = pipeline('text-generation', model="Esperanto/Protein-Llama-3-8B")
sequences = generator("[Generate Ligase enzyme protein] Seq=<",temperature=0.2,
top_k=40,
top_p=0.9,
do_sample=True,
repetition_penalty=1.2,
max_new_tokens=30,
num_return_sequences=500)
for sequence in sequences:
print(sequence['generated_text'])
Contributors
Aayush Shah, Shankar Jayaratnam