Satyam66 commited on
Commit
fda2f7e
·
verified ·
1 Parent(s): 2d7ae4d

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +40 -18
README.md CHANGED
@@ -1,37 +1,54 @@
1
  ---
2
  base_model: microsoft/Phi-3-mini-4k-instruct
3
- library_name: peft
 
 
 
 
 
 
 
 
 
 
 
 
4
  ---
5
 
6
  # Model Card for Model ID
7
 
8
- <!-- Provide a quick summary of what the model is/does. -->
9
-
10
 
11
 
12
  ## Model Details
13
-
 
 
 
 
14
  ### Model Description
15
 
16
- <!-- Provide a longer summary of what this model is. -->
17
-
 
 
 
 
 
18
 
19
 
20
- - **Developed by:** [More Information Needed]
21
- - **Funded by [optional]:** [More Information Needed]
22
- - **Shared by [optional]:** [More Information Needed]
23
- - **Model type:** [More Information Needed]
24
- - **Language(s) (NLP):** [More Information Needed]
25
- - **License:** [More Information Needed]
26
- - **Finetuned from model [optional]:** [More Information Needed]
27
 
28
- ### Model Sources [optional]
29
 
30
  <!-- Provide the basic links for the model. -->
31
 
32
- - **Repository:** [More Information Needed]
33
- - **Paper [optional]:** [More Information Needed]
34
- - **Demo [optional]:** [More Information Needed]
35
 
36
  ## Uses
37
 
@@ -93,7 +110,12 @@ Use the code below to get started with the model.
93
  #### Training Hyperparameters
94
 
95
  - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
96
-
 
 
 
 
 
97
  #### Speeds, Sizes, Times [optional]
98
 
99
  <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
 
1
  ---
2
  base_model: microsoft/Phi-3-mini-4k-instruct
3
+ library_name: transformers
4
+ license: apache-2.0
5
+ language:
6
+ - en
7
+ tags:
8
+ - Finetuning
9
+ - PEFT
10
+ - NLP
11
+ - LLM
12
+ - text-generation-inference
13
+ - transformers
14
+ - QLoRA
15
+ - LoRA
16
  ---
17
 
18
  # Model Card for Model ID
19
 
20
+ This is a finetuned model trained on agricultural datasets for crop disease remedies.
 
21
 
22
 
23
  ## Model Details
24
+ phi3-finetuned-20250414-0740 can be used for crop disease remedies.
25
+ Peft ,QLoRA ,LoRA ,
26
+ transformers are used and supervised finetuning is done for training this model.
27
+ LoRA_dropout was taken 0.1, Lora_r=16.
28
+ This model was trained on Google Colab free tier giving T4 GPU of 15 GB vRAM and can be used for 12 hours.
29
  ### Model Description
30
 
31
+ Finetuned Agricultural Chatbot (Phi-3-mini-4k-instruct)
32
+ fine-tuned Microsoft’s Phi-3-mini-4k-instruct, a compact yet powerful instruction-tuned LLM (~3.8B parameters),
33
+ specifically for agriculture-related tasks using curated domain-specific datasets.
34
+ Built on top of Microsoft’s Phi-3-mini-4k-instruct, a lightweight but capable open-source language model,
35
+ this chatbot has been carefully trained using thousands of real-world examples from the agricultural domain.
36
+ From crop disease symptoms and soil health tips to pesticide usage and sustainable farming practices,
37
+ it has absorbed knowledge from curated, high-quality datasets.
38
 
39
 
40
+ - **Developed by:** Satyam Kahali (reach me out on LinkedIN "https://www.linkedin.com/in/satyam-kahali-883098235/")
41
+ - **Model type:** Causal Language Model (CausalLM)
42
+ - **Language(s) (NLP):** Python
43
+ - **License:** apache-2.0
44
+ - **Finetuned from model :** microsoft/Phi-3-mini-4k-instruct
 
 
45
 
46
+ ### Model Sources
47
 
48
  <!-- Provide the basic links for the model. -->
49
 
50
+ - **Repository:** Satyam66/phi3-finetuned-20250414-0740
51
+
 
52
 
53
  ## Uses
54
 
 
110
  #### Training Hyperparameters
111
 
112
  - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
113
+ lora_alpha: 32,
114
+ lora_bias: false,
115
+ lora_dropout: 0.05,
116
+ r: 16,
117
+ fp16 = True
118
+ bf16 = False
119
  #### Speeds, Sizes, Times [optional]
120
 
121
  <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->