Model Card for Model ID

This model is a fine-tuned version of TinyLlama-1.1B-Chat-v1.0, specialized for question-answering and summarization tasks related to the topic of DNA data storage. It was trained using the PEFT (Parameter-Efficient Fine-Tuning) method with LoRA adapters on a custom dataset tatsu-lab/alpaca.

Model Details

Model Description

This is a fine-tuned language model based on the TinyLlama-1.1B-Chat-v1.0 architecture. The model was trained to improve its ability to understand, summarize, and answer questions from text related to DNA data storage technology. It utilizes LoRA adapters, which makes the fine-tuned checkpoint small and efficient. This model is intended for research and educational purposes to explore the application of LLMs in niche, domain-specific tasks.

  • Developed by: Abhishek Singh
  • Model type: Causal language model
  • Language(s) (NLP): English
  • License: Apache 2.0 (Inherits the license from the base model)
  • Finetuned from model: TinyLlama/TinyLlama-1.1B-Chat-v1.0

Model Sources [optional]

Uses

Direct Use

This model is intended to be used for the following purposes:

  • Summarizing key points from new documents or texts about DNA data storage.
  • Answering specific questions based on provided context regarding DNA data storage.
  • Generating short, informative explanations on the topic.

Out-of-Scope Use

This model is not suitable for:

  • General-purpose chat or conversational tasks on topics outside of DNA data storage.
  • Generating creative writing or essays.
  • Factual question-answering on general knowledge, as its knowledge is constrained to the fine-tuning data.

Bias, Risks, and Limitations

This model has several limitations due to its specialized nature and small size:

  • Domain-Specific Knowledge: The model's knowledge is highly specialized. It may provide incorrect or nonsensical information (hallucinate) when asked about topics outside of DNA data storage.
  • Potential for Bias: The model inherits the biases of its base model, TinyLlama.
  • Simplicity: The model is not a substitute for expert advice or comprehensive research. It should be used as a supplementary tool for text analysis.

Recommendations

Users should be aware of the model's limitations and verify any critical information it provides. It is recommended to use the model with a clear, specific prompt that includes relevant context for the best results.

How to Get Started with the Model

You can get started with the model by loading it directly from the Hugging Face Hub using the transformers library.

from transformers import AutoTokenizer, AutoModelForCausalLM

model_id = "learn-abc/tatsu-lab_alpaca-tinyllama-chat-bnb-4bit"

# Load the tokenizer and model
tokenizer = AutoTokenizer.from_pretrained(model_id)
model = AutoModelForCausalLM.from_pretrained(model_id)

# Example usage
prompt = """### Instruction:
Summarize the key findings from the provided text about DNA data storage.

### Input:
Deoxyribonucleic acid (DNA) has been successfully proposed as an advanced archival storage medium, due to its extraordinary data capacity and robust stability. ... (rest of your text here)

### Response:
"""

inputs = tokenizer(prompt, return_tensors="pt")
outputs = model.generate(**inputs, max_new_tokens=256)
response = tokenizer.decode(outputs[0], skip_special_tokens=True)

print(response)

Training Details

Training Data

The model was fine-tuned on a custom dataset containing text, summaries, and Q&A pairs related to the topic of DNA data storage. The dataset was formatted into a chat-like template with ### Instruction:, ### Input:, and ### Response: sections.

Training Procedure

1. Preprocessing The training data was pre-processed into a chat template format to prepare it for the model. The tokenizer's pad_token was set to the eos_token to handle variable-length sequences.

2. Training Hyperparameters

  • Fine-tuning Method: PEFT (LoRA)
  • Training regime: bf16 mixed precision

Evaluation

No formal evaluation metrics were calculated for this model. Its performance was qualitatively assessed by generating responses to prompts and checking for relevance and accuracy with respect to the fine-tuning data.

Downloads last month
10
Safetensors
Model size
1.1B params
Tensor type
F32
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support