Model Card: llm-course-hw3-dora

This model was fine-tuned as part of Homework 3 in the HSE LLM Course.
It applies a custom DoRA implementation for sentiment classification, using standard causal language modeling training.

The model predicts a sentiment label (negative, neutral, or positive) by generating a short textual output conditioned on the input text.

Model Sources

Training Method

Training follows a standard causal LM setup with frozen backbone weights and trainable DoRA adapters inserted into attention projection layers.

Training Hyperparameters

  • PEFT method: DoRA (custom implementation)
  • Rank: 8
  • DoRA alpha: 16
  • Target modules: attention projections (k_proj, v_proj)
  • Batch size: 32
  • Learning rate: 5e-4
  • Optimizer: AdamW
  • Precision: FP16 (mixed precision)
  • Epochs: 3

Trainable parameters: ~0.14% of total model parameters.

Results

  • Macro F1: ~0.5 on the test set
Downloads last month
1
Safetensors
Model size
0.3B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for sodeniZz/llm-course-hw3-dora

Finetuned
(29)
this model

Dataset used to train sodeniZz/llm-course-hw3-dora

Collection including sodeniZz/llm-course-hw3-dora