Henyo-70M

Henyo is a custom Tagalog LLM trained on a subset of Wikipedia.

Model Architecture

  • Parameter Count: 70M
  • Architecture: Decoder-only Transformer (Custom)
  • Features:
    • SwiGLU Activation
    • Grouped Query Attention (GQA)
    • Rotary Positional Embeddings (RoPE)
    • RMSNorm

Usage

This model uses a custom architecture code. You can load it using the AutoModel class with trust_remote_code=True (if code is uploaded) or by defining the class manually.

from transformers import AutoTokenizer, AutoModelForCausalLM

tokenizer = AutoTokenizer.from_pretrained("marcuscedricridia/Henyo-70M")
# Note: Since this is a custom model, you may need the inference script provided in the repo.
Downloads last month
11
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support