Post
186
GLM-Image from Z.ai is out π₯
It was fully trained on Ascend Atlas 800T A2 with MindSpore, probably the first SOTA multimodal model fully trained on domestic chips π
zai-org/GLM-Image
β¨ Hybrid Architecture: combined autoregressive + diffusion design delivers strong semantic alignment with high-fidelity details
β¨ Strong performance in long, dense, and multilingual text rendering
β¨ MIT licensed (VQ tokenizer & ViT weights under Apache 2.0)
β¨ Now live on Hugging Face inference provider π€
It was fully trained on Ascend Atlas 800T A2 with MindSpore, probably the first SOTA multimodal model fully trained on domestic chips π
zai-org/GLM-Image
β¨ Hybrid Architecture: combined autoregressive + diffusion design delivers strong semantic alignment with high-fidelity details
β¨ Strong performance in long, dense, and multilingual text rendering
β¨ MIT licensed (VQ tokenizer & ViT weights under Apache 2.0)
β¨ Now live on Hugging Face inference provider π€