YAML Metadata Warning: empty or missing yaml metadata in repo card
Check out the documentation for more information.
AnCoder-Base-2.1B
Anchored bidirectional diffusion language model built on Qwen3. Current checkpoint: 37k steps.
Usage
from transformers import AutoModel, AutoTokenizer
model = AutoModel.from_pretrained("EER6/AnCoder-Base-2.1B", trust_remote_code=True)
tokenizer = AutoTokenizer.from_pretrained("EER6/AnCoder-Base-2.1B")
# Pad with mask tokens for diffusion inference
tokenizer.pad_token_id = 151660 # <|fim_middle|>
inputs = tokenizer("def fibonacci(n):", return_tensors="pt", padding="max_length", max_length=1024)
outputs = model(**inputs)
outputs.logits # (B, L, V) denoiser predictions
outputs.anchor_logits # (B, L, V) anchor predictions
- Downloads last month
- 14
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support