Parameter-Efficient Fine-Tuning (LoRA & DoRa & QLoRA)
Collection
A collection of parameter-efficient fine-tuning experiments for sentiment classification using chat-based instruction tuning • 4 items • Updated
This model was fine-tuned as part of Homework 3 in the HSE LLM Course.
It applies a custom DoRA implementation for sentiment classification, using standard causal language modeling training.
The model predicts a sentiment label (negative, neutral, or positive) by generating a short textual output conditioned on the input text.
Training follows a standard causal LM setup with frozen backbone weights and trainable DoRA adapters inserted into attention projection layers.
k_proj, v_proj)Trainable parameters: ~0.14% of total model parameters.
Base model
OuteAI/Lite-Oute-1-300M-Instruct