Edit model card

Model Details

Protein-Phi-3-mini is a specialized version of the Phi-3-mini-4k-instruct large language model, fine-tuned for the task of protein language modeling. This model has been continually pre-trained using LoRA technique on extensive datasets of protein sequences, enabling it to generate novel protein sequences based on natural language prompts. It supports both uncontrollable and controllable protein generation, allowing users to specify desired characteristics for the proteins. The model is designed to facilitate advancements in protein engineering, making it a valuable tool for drug development, chemical synthesis, and other biotechnological applications. For full details please read our paper.

Model Description

Generating novel protein sequences possessing desired properties, termed as protein engineering, is crucial for industries like drug development and chemical synthesis. Traditional protein engineering techniques often involve introducing random mutations into the gene encoding the protein of interest. This is followed by expression and screening to identify variants with improved or novel functions, which are then reproduced. While effective, these approaches are labor-intensive and time-consuming, as they rely on iterating over known protein sequences. This limits their ability to generate diverse protein sequences with entirely new capabilities, as they are constrained by existing protein templates. Moreover, the need to analyze numerous protein variants can waste valuable experimental resources.

However, leveraging a Large Language Model (LLM) that has learned the "protein language" significantly accelerates this process. An LLM can generate and evaluate protein sequences in a matter of seconds. The inherent randomness of LLM-generated sequences enhances diversity, enabling the creation of completely novel proteins with potentially unprecedented functions. This not only streamlines the discovery and development process but also expands the scope of possibilities in protein engineering.

This model is based on the Phi-3-mini-4k-instruct architecture and is capable of generating proteins based on user defined characteristics.

Usage

To download and use the Protein Phi 3 model for inference, follow these steps:

Installation

Ensure you have the transformers library installed. You can install it using pip:

pip install transformers

Uncontrollable Generation

Uncontrollable generation can be handled via prompting the model with the phrase 'Seq=<'.

generator = pipeline('text-generation', model="Esperanto/Protein-Phi-3-mini", trust_remote_code=True)

sequences = generator("Seq=<",temperature=0.2,
    top_k=40,
    top_p=0.9,
    do_sample=True,
    repetition_penalty=1.2,
    max_new_tokens=30,
    num_return_sequences=500)

for sequence in sequences:
    print(sequence['generated_text'])

Controllable Generation

Controllable generation can be done by prompting the model with '[Generate xxx protein] Seq=<'. Here, xxx can be any family from the 10 classes supported by this model.

generator = pipeline('text-generation', model="Esperanto/Protein-Phi-3-mini", trust_remote_code=True)

sequences = generator("[Generate Ligase enzyme protein] Seq=<",temperature=0.2,
    top_k=40,
    top_p=0.9,
    do_sample=True,
    repetition_penalty=1.2,
    max_new_tokens=30,
    num_return_sequences=500)

for sequence in sequences:
    print(sequence['generated_text'])

Contributors

Aayush Shah, Shankar Jayaratnam

Downloads last month
15
Safetensors
Model size
3.82B params
Tensor type
BF16
ยท
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Model tree for Esperanto/Protein-Phi-3-mini

Finetuned
(146)
this model

Space using Esperanto/Protein-Phi-3-mini 1

Collection including Esperanto/Protein-Phi-3-mini