| | --- |
| | library_name: transformers |
| | pipeline_tag: text-generation |
| | --- |
| | |
| | # ChatNT |
| |
|
| | [ChatNT](https://www.biorxiv.org/content/10.1101/2024.04.30.591835v1) is the first multimodal conversational agent designed with a deep understanding of biological sequences (DNA, RNA, proteins). |
| | It enables users — even those with no coding background — to interact with biological data through natural language and it generalizes across multiple biological tasks and modalities. |
| |
|
| | **Developed by:** [InstaDeep](https://huggingface.co/InstaDeepAI) |
| |
|
| | ### Model Sources |
| |
|
| | <!-- Provide the basic links for the model. --> |
| |
|
| | - **Repository:** [Nucleotide Transformer](https://github.com/instadeepai/nucleotide-transformer) |
| | - **Paper:** [ChatNT: A Multimodal Conversational Agent for DNA, RNA and Protein Tasks](https://www.biorxiv.org/content/10.1101/2024.04.30.591835v1.full.pdf) |
| |
|
| |
|
| | ### License Summary |
| | 1. The Licensed Models are **only** available under this License for Non-Commercial Purposes. |
| | 2. You are permitted to reproduce, publish, share and adapt the Output generated by the Licensed Model only for Non-Commercial Purposes and in accordance with this License. |
| | 3. You may **not** use the Licensed Models or any of its Outputs in connection with: |
| | 1. any Commercial Purposes, unless agreed by Us under a separate licence; |
| | 2. to train, improve or otherwise influence the functionality or performance of any other third-party derivative model that is commercial or intended for a Commercial Purpose and is similar to the Licensed Models; |
| | 3. to create models distilled or derived from the Outputs of the Licensed Models, unless such models are for Non-Commercial Purposes and open-sourced under the same license as the Licensed Models; or |
| | 4. in violation of any applicable laws and regulations. |
| |
|
| | ### Architecture and Parameters |
| | ChatNT is built on a three‑module design: a 500M‑parameter [Nucleotide Transformer v2](https://www.nature.com/articles/s41592-024-02523-z) DNA encoder pre‑trained on genomes from 850 species |
| | (handling up to 12 kb per sequence, Dalla‑Torre et al., 2024), an English‑aware Perceiver Resampler that linearly projects and gated‑attention compresses |
| | 2048 DNA‑token embeddings into 64 task‑conditioned vectors (REF), and a frozen 7B‑parameter [Vicuna‑7B](https://lmsys.org/blog/2023-03-30-vicuna/) decoder. |
| |
|
| | Users provide a natural‑language prompt containing one or more `<DNA>` placeholders and the corresponding DNA sequences (tokenized as 6‑mers). |
| | The projection layer inserts 64 resampled DNA embeddings at each placeholder, and the Vicuna decoder generates free‑form English responses in |
| | an autoregressive fashion, using low‑temperature sampling to produce classification labels, multi‑label statements, or numeric values. |
| |
|
| | ### Training Data |
| | ChatNT was instruction‑tuned on a unified corpus covering 27 diverse tasks from DNA, RNA and proteins, spanning multiple species, tissues and biological processes. |
| | This amounted to 605 million DNA tokens (≈ 3.6 billion bases) and 273 million English tokens, sampled uniformly over tasks for 2 billion instruction tokens. |
| | Examples of questions and sequences for each task, as well as additional task information, can be found in [Datasets_overview.csv](Datasets_overview.csv). |
| |
|
| | ### Tokenization |
| | DNA inputs are broken into overlapping 6‑mer tokens and padded or truncated to 2048 tokens (~ 12 kb). English prompts and |
| | outputs use the LLaMA tokenizer, augmented with `<DNA>` as a special token to mark sequence insertion points. |
| |
|
| | ### Limitations and Disclaimer |
| | ChatNT can only handle questions related to the 27 tasks it has been trained on, including the same format of DNA sequences. ChatNT is **not** a clinical or diagnostic tool. |
| | It can produce incorrect or “hallucinated” answers, particularly on out‑of‑distribution inputs, and its numeric predictions may suffer digit‑level errors. Confidence |
| | estimates require post‑hoc calibration. Users should always validate critical outputs against experiments or specialized bioinformatics |
| | pipelines. |
| |
|
| | ### Other notes |
| | We also provide the params for the ChatNT jax model in `jax_params`. |
| |
|
| | ## How to use |
| |
|
| | Until its next release, the transformers library needs to be installed from source with the following command in order to use the models. |
| | PyTorch should also be installed. |
| |
|
| | ``` |
| | pip install --upgrade git+https://github.com/huggingface/transformers.git |
| | pip install torch sentencepiece |
| | ``` |
| |
|
| | A small snippet of code is given here in order to **generate ChatNT answers from a pipeline (high-level)**. |
| | - The prompt used for training ChatNT is already incorporated inside the pipeline and is the following: |
| | "A chat between a curious user and an artificial intelligence assistant that can handle bio sequences. The assistant gives helpful, |
| | detailed, and polite answers to the user's questions." |
| |
|
| | ``` |
| | # Load pipeline |
| | from transformers import pipeline |
| | pipe = pipeline(model="InstaDeepAI/ChatNT", trust_remote_code=True) |
| | |
| | # Define custom inputs (note that the number of <DNA> token in the english sequence must be equal to len(dna_sequences)) |
| | english_sequence = "Is there any evidence of an acceptor splice site in this sequence <DNA> ?" |
| | dna_sequences = ["ATCGGAAAAAGATCCAGAAAGTTATACCAGGCCAATGGGAATCACCTATTACGTGGATAATAGCGATAGTATGTTACCTATAAATTTAACTACGTGGATATCAGGCAGTTACGTTACCAGTCAAGGAGCACCCAAAACTGTCCAGCAACAAGTTAATTTACCCATGAAGATGTACTGCAAGCCTTGCCAACCAGTTAAAGTAGCTACTCATAAGGTAATAAACAGTAATATCGACTTTTTATCCATTTTGATAATTGATTTATAACAGTCTATAACTGATCGCTCTACATAATCTCTATCAGATTACTATTGACACAAACAGAAACCCCGTTAATTTGTATGATATATTTCCCGGTAAGCTTCGATTTTTAATCCTATCGTGACAATTTGGAATGTAACTTATTTCGTATAGGATAAACTAATTTACACGTTTGAATTCCTAGAATATGGAGAATCTAAAGGTCCTGGCAATGCCATCGGCTTTCAATATTATAATGGACCAAAAGTTACTCTATTAGCTTCCAAAACTTCGCGTGAGTACATTAGAACAGAAGAATAACCTTCAATATCGAGAGAGTTACTATCACTAACTATCCTATG"] |
| | |
| | # Generate sequence |
| | generated_english_sequence = pipe( |
| | inputs={ |
| | "english_sequence": english_sequence, |
| | "dna_sequences": dna_sequences |
| | } |
| | ) |
| | |
| | # Expected output: "Yes, an acceptor splice site is without question present in the sequence." |
| | ``` |
| |
|
| | A small snippet of code is given here in order to **infer with the model without any abstraction (low-level)**. |
| |
|
| | ``` |
| | import numpy as np |
| | from transformers import AutoModel, AutoTokenizer |
| | |
| | # Load model and tokenizers |
| | model = AutoModel.from_pretrained("InstaDeepAI/ChatNT", trust_remote_code=True) |
| | english_tokenizer = AutoTokenizer.from_pretrained("InstaDeepAI/ChatNT", subfolder="english_tokenizer") |
| | bio_tokenizer = AutoTokenizer.from_pretrained("InstaDeepAI/ChatNT", subfolder="bio_tokenizer") |
| | |
| | # Define custom inputs (note that the number of <DNA> token in the english sequence must be equal to len(dna_sequences)) |
| | # Here the english sequence should include the prompt |
| | english_sequence = "A chat between a curious user and an artificial intelligence assistant that can handle bio sequences. The assistant gives helpful, detailed, and polite answers to the user's questions. USER: Is there any evidence of an acceptor splice site in this sequence <DNA> ?" |
| | dna_sequences = ["ATCGGAAAAAGATCCAGAAAGTTATACCAGGCCAATGGGAATCACCTATTACGTGGATAATAGCGATAGTATGTTACCTATAAATTTAACTACGTGGATATCAGGCAGTTACGTTACCAGTCAAGGAGCACCCAAAACTGTCCAGCAACAAGTTAATTTACCCATGAAGATGTACTGCAAGCCTTGCCAACCAGTTAAAGTAGCTACTCATAAGGTAATAAACAGTAATATCGACTTTTTATCCATTTTGATAATTGATTTATAACAGTCTATAACTGATCGCTCTACATAATCTCTATCAGATTACTATTGACACAAACAGAAACCCCGTTAATTTGTATGATATATTTCCCGGTAAGCTTCGATTTTTAATCCTATCGTGACAATTTGGAATGTAACTTATTTCGTATAGGATAAACTAATTTACACGTTTGAATTCCTAGAATATGGAGAATCTAAAGGTCCTGGCAATGCCATCGGCTTTCAATATTATAATGGACCAAAAGTTACTCTATTAGCTTCCAAAACTTCGCGTGAGTACATTAGAACAGAAGAATAACCTTCAATATCGAGAGAGTTACTATCACTAACTATCCTATG"] |
| | |
| | # Tokenize |
| | english_tokens = english_tokenizer(english_sequence, return_tensors="pt", padding="max_length", truncation=True, max_length=512).input_ids |
| | bio_tokens = bio_tokenizer(dna_sequences, return_tensors="pt", padding="max_length", max_length=512, truncation=True).input_ids.unsqueeze(0) # unsqueeze to simulate batch_size = 1 |
| | |
| | # Predict |
| | outs = model( |
| | multi_omics_tokens_ids=(english_tokens, bio_tokens), |
| | projection_english_tokens_ids=english_tokens, |
| | projected_bio_embeddings=None, |
| | ) |
| | |
| | # Expected output: Dictionary of logits and projected_bio_embeddings |
| | ``` |