drawing

Table of Contents

  1. TL;DR
  2. Model Details
  3. Training Details
  4. Usage
  5. Evaluation
  6. Citation

TL;DR

Model Details

Model Description

  • Developed by: https://www.tii.ae
  • Model type: Causal decoder-only
  • Architecture: Hybrid Transformers + Mamba architecture
  • Language(s) (NLP): English
  • Number of Parameters: 90M
  • License: Falcon-LLM License

Training details

For more details about the training protocol of this model, please refer to the Falcon-H1-Tiny technical blogpost.

Usage

Currently to use this model you can either rely on Hugging Face transformers, vLLM, sglang, llama.cpp, ollama or mlx library.

Inference

πŸ€— transformers

Refer to the snippet below to run H1 models using πŸ€— transformers:

import torch
from transformers import AutoModelForCausalLM, AutoTokenizer

model_id = "tiiuae/Tiny-H1-SFT"

model = AutoModelForCausalLM.from_pretrained(
  model_id,
  torch_dtype=torch.bfloat16,
  device_map="auto"
)

# Perform text generation

or

transformers serve tiiuae/Tiny-H1-SFT

llama.cpp

You can find all GGUF files compatible with llama.cpp under our official collection - an example setup could be:

brew install llama.cpp 
pip install huggingface_hub 
hf download tiiuae/Tiny-H1-SFT tiny-h1-sft-pretrain-Q8_0.gguf --local-dir ./ 
llama-cli ./ Tiny-H1-SFT-Q8_0.gguf -cnv 

ollama

ollama run hf.co/tiiuae/Tiny-H1-SFT:Q8_0 

Apple mlx

mlx_lm.chat --model tiiuae/Tiny-H1-SF 

vLLM

For vLLM, simply start a server by executing the command below:

# pip install vllm>=0.9.0
vllm serve tiiuae/Tiny-H1-SFT --tensor-parallel-size 2 --data-parallel-size 1

sglang

python -m sglang.launch_server \
  --model ttiiuae/Tiny-H1-SFT \
  --tensor-parallel-size 1 

Evaluation

For detailed evaluation of Falcon-H1-Tiny series, please refer to our technical blogpost

Useful links

Citation

If the Falcon-H1-Tiny family of models were helpful to your work, feel free to give us a cite.

@misc{falcon_h1_tiny,
  title={Falcon-H1-Tiny: A series of extremely small, yet powerful language models redefining capabilities at small scale},
  author={Falcon-LLM Team},
  year={2026}, 
}
Downloads last month
15
Safetensors
Model size
91.1M params
Tensor type
BF16
Β·
Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support

Model tree for tiiuae/Falcon-H1-Tiny-90M-Base

Quantizations
1 model

Collection including tiiuae/Falcon-H1-Tiny-90M-Base