Update README.md
Browse files
README.md
CHANGED
@@ -1,37 +1,54 @@
|
|
1 |
---
|
2 |
base_model: microsoft/Phi-3-mini-4k-instruct
|
3 |
-
library_name:
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
4 |
---
|
5 |
|
6 |
# Model Card for Model ID
|
7 |
|
8 |
-
|
9 |
-
|
10 |
|
11 |
|
12 |
## Model Details
|
13 |
-
|
|
|
|
|
|
|
|
|
14 |
### Model Description
|
15 |
|
16 |
-
|
17 |
-
|
|
|
|
|
|
|
|
|
|
|
18 |
|
19 |
|
20 |
-
- **Developed by:**
|
21 |
-
- **
|
22 |
-
- **
|
23 |
-
- **
|
24 |
-
- **
|
25 |
-
- **License:** [More Information Needed]
|
26 |
-
- **Finetuned from model [optional]:** [More Information Needed]
|
27 |
|
28 |
-
### Model Sources
|
29 |
|
30 |
<!-- Provide the basic links for the model. -->
|
31 |
|
32 |
-
- **Repository:**
|
33 |
-
|
34 |
-
- **Demo [optional]:** [More Information Needed]
|
35 |
|
36 |
## Uses
|
37 |
|
@@ -93,7 +110,12 @@ Use the code below to get started with the model.
|
|
93 |
#### Training Hyperparameters
|
94 |
|
95 |
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
|
96 |
-
|
|
|
|
|
|
|
|
|
|
|
97 |
#### Speeds, Sizes, Times [optional]
|
98 |
|
99 |
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
|
|
|
1 |
---
|
2 |
base_model: microsoft/Phi-3-mini-4k-instruct
|
3 |
+
library_name: transformers
|
4 |
+
license: apache-2.0
|
5 |
+
language:
|
6 |
+
- en
|
7 |
+
tags:
|
8 |
+
- Finetuning
|
9 |
+
- PEFT
|
10 |
+
- NLP
|
11 |
+
- LLM
|
12 |
+
- text-generation-inference
|
13 |
+
- transformers
|
14 |
+
- QLoRA
|
15 |
+
- LoRA
|
16 |
---
|
17 |
|
18 |
# Model Card for Model ID
|
19 |
|
20 |
+
This is a finetuned model trained on agricultural datasets for crop disease remedies.
|
|
|
21 |
|
22 |
|
23 |
## Model Details
|
24 |
+
phi3-finetuned-20250414-0740 can be used for crop disease remedies.
|
25 |
+
Peft ,QLoRA ,LoRA ,
|
26 |
+
transformers are used and supervised finetuning is done for training this model.
|
27 |
+
LoRA_dropout was taken 0.1, Lora_r=16.
|
28 |
+
This model was trained on Google Colab free tier giving T4 GPU of 15 GB vRAM and can be used for 12 hours.
|
29 |
### Model Description
|
30 |
|
31 |
+
Finetuned Agricultural Chatbot (Phi-3-mini-4k-instruct)
|
32 |
+
fine-tuned Microsoft’s Phi-3-mini-4k-instruct, a compact yet powerful instruction-tuned LLM (~3.8B parameters),
|
33 |
+
specifically for agriculture-related tasks using curated domain-specific datasets.
|
34 |
+
Built on top of Microsoft’s Phi-3-mini-4k-instruct, a lightweight but capable open-source language model,
|
35 |
+
this chatbot has been carefully trained using thousands of real-world examples from the agricultural domain.
|
36 |
+
From crop disease symptoms and soil health tips to pesticide usage and sustainable farming practices,
|
37 |
+
it has absorbed knowledge from curated, high-quality datasets.
|
38 |
|
39 |
|
40 |
+
- **Developed by:** Satyam Kahali (reach me out on LinkedIN "https://www.linkedin.com/in/satyam-kahali-883098235/")
|
41 |
+
- **Model type:** Causal Language Model (CausalLM)
|
42 |
+
- **Language(s) (NLP):** Python
|
43 |
+
- **License:** apache-2.0
|
44 |
+
- **Finetuned from model :** microsoft/Phi-3-mini-4k-instruct
|
|
|
|
|
45 |
|
46 |
+
### Model Sources
|
47 |
|
48 |
<!-- Provide the basic links for the model. -->
|
49 |
|
50 |
+
- **Repository:** Satyam66/phi3-finetuned-20250414-0740
|
51 |
+
|
|
|
52 |
|
53 |
## Uses
|
54 |
|
|
|
110 |
#### Training Hyperparameters
|
111 |
|
112 |
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
|
113 |
+
lora_alpha: 32,
|
114 |
+
lora_bias: false,
|
115 |
+
lora_dropout: 0.05,
|
116 |
+
r: 16,
|
117 |
+
fp16 = True
|
118 |
+
bf16 = False
|
119 |
#### Speeds, Sizes, Times [optional]
|
120 |
|
121 |
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
|