Update README.md
Browse files
README.md
CHANGED
|
@@ -9,8 +9,18 @@ language:
|
|
| 9 |
---
|
| 10 |
|
| 11 |
# Model Card for Model ID
|
| 12 |
-
A 4-bits quantized version of [ernestoBocini/Phi3-mini-DPO-Tuned](https://huggingface.co/ernestoBocini/Phi3-mini-DPO-Tuned).
|
| 13 |
|
| 14 |
## Model Details
|
| 15 |
|
| 16 |
-
This is a Phi-3-mini-4k-instruct fine-tuned with SFT and DPO on STEM domains, and finally quantized to a 4 bits precision, to serve the purpose of being an AI university tutor.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 9 |
---
|
| 10 |
|
| 11 |
# Model Card for Model ID
|
| 12 |
+
A 4-bits double-quantized version of [ernestoBocini/Phi3-mini-DPO-Tuned](https://huggingface.co/ernestoBocini/Phi3-mini-DPO-Tuned).
|
| 13 |
|
| 14 |
## Model Details
|
| 15 |
|
| 16 |
+
This is a Phi-3-mini-4k-instruct fine-tuned with SFT and DPO on STEM domains, and finally quantized to a 4 bits precision, to serve the purpose of being an AI university tutor.
|
| 17 |
+
|
| 18 |
+
Quantization config used:
|
| 19 |
+
```python
|
| 20 |
+
BitsAndBytesConfig(
|
| 21 |
+
load_in_4bit=True,
|
| 22 |
+
bnb_4bit_quant_type="nf4",
|
| 23 |
+
bnb_4bit_use_double_quant=True,
|
| 24 |
+
bnb_4bit_compute_dtype=torch.bfloat16
|
| 25 |
+
)
|
| 26 |
+
```
|