deepkaria commited on
Commit
b9d8559
·
verified ·
1 Parent(s): 47a9273

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +70 -6
README.md CHANGED
@@ -12,12 +12,76 @@ language:
12
  - en
13
  ---
14
 
15
- # Uploaded model
16
 
17
- - **Developed by:** deepkaria
18
- - **License:** apache-2.0
19
- - **Finetuned from model :** unsloth/deepseek-r1-distill-qwen-1.5b-unsloth-bnb-4bit
20
 
21
- This qwen2 model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
22
 
23
- [<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
 
12
  - en
13
  ---
14
 
15
+ # Model: deepkaria/deepseek-r1-1.5b-indian-culture
16
 
17
+ ## Language
18
+ **en**
 
19
 
20
+ ## Tags
21
+ - deepseek
22
+ - indian-culture
23
+ - cultural-heritage
24
+ - lora
25
+ - fine-tuned
26
+
27
+ ## Datasets
28
+ - [deepkaria/indian-culture-dataset](https://huggingface.co/datasets/deepkaria/indian-culture-dataset)
29
+
30
+ ## License
31
+ **Apache-2.0**
32
+
33
+ ## Base Model
34
+ [deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B](https://huggingface.co/deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B)
35
+
36
+ ## Model Description
37
+ This model has been fine-tuned on the Indian Culture Dataset to provide detailed and accurate information about various aspects of Indian culture, including festivals, performing arts, architecture, rituals, traditional medicine, and more.
38
+
39
+ ## Training Details
40
+ **Base Model:** deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B
41
+
42
+ ### Training Method
43
+ LoRA fine-tuning with the following parameters:
44
+ - **LoRA rank:** 16
45
+ - **LoRA alpha:** 32
46
+ - **Target modules:** Attention and MLP layers
47
+ - **Training epochs:** 3
48
+ - **Learning rate:** 2e-4 with cosine scheduler
49
+
50
+ ## Usage Example
51
+
52
+ ```python
53
+ from transformers import AutoModelForCausalLM, AutoTokenizer
54
+
55
+ model_name = "deepkaria/deepseek-r1-1.5b-indian-culture"
56
+ tokenizer = AutoTokenizer.from_pretrained(model_name)
57
+ model = AutoModelForCausalLM.from_pretrained(model_name)
58
+
59
+ prompt = """Below is an instruction that describes a task, paired with an input that provides further context.
60
+ Write a response that appropriately addresses the instruction.
61
+
62
+ ### Instruction:
63
+ You are an expert on Indian culture, traditions, and heritage. Provide detailed and accurate information about the following aspect of Indian culture.
64
+
65
+ ### Input:
66
+ Tell me about Kathakali from Kerala.
67
+
68
+ ### Response:
69
+ """
70
+
71
+ inputs = tokenizer(prompt, return_tensors="pt").to(model.device)
72
+ outputs = model.generate(
73
+ **inputs,
74
+ max_new_tokens=512,
75
+ temperature=0.7,
76
+ top_p=0.9,
77
+ do_sample=True
78
+ )
79
+ print(tokenizer.decode(outputs[0], skip_special_tokens=True))
80
+ ```
81
+
82
+ ## Limitations
83
+ The model's knowledge is limited to the information contained in the training dataset. While it covers a wide range of Indian cultural topics, it may not have comprehensive information about very specific or regional cultural practices.
84
+
85
+ ## Intended Use
86
+ This model is designed for educational purposes, cultural research, and to promote understanding of India's diverse cultural landscape.
87