Recipe ·
GLM-5.1
GLM-5.1, developed by Zhipu AI, is a next-generation foundation model featuring a 1 million token context window and support for 128K output tokens.
Ingredient list
Invented by Nous Research · 2023-08 · Velocity 3y
“GLM-5.1 extends context length to 1M tokens using YaRN (Yet another RoPE extensioN) method.”
architecturehighInvented by Google · 2023-05 · Velocity 3y
“GLM-5.1 architecture uses Grouped-Query Attention (GQA) to reduce KV cache memory.”
architecturehighInvented by Stanford · 2022-05 · Velocity 4y
“GLM-5.1 implements FlashAttention-2 for efficient attention computation.”
inferencehighInvented by Google · 2022-03 · Velocity 4y
“GLM-5.1 can use self-consistency by sampling multiple reasoning paths.”
reasoningmediumInvented by Google · 2022-01 · Velocity 4y
“GLM-5.1 demonstrates chain-of-thought reasoning capabilities in examples.”
reasoningmediumInvented by Google · 2021-09 · Velocity 5y
“GLM-5.1 is instruction-tuned on diverse tasks following FLAN methodology.”
trainingmediumInvented by Zhuiyi Technology · 2021-04 · Velocity 5y
“GLM-5.1 uses Rotary Position Embedding (RoPE) for positional encoding.”
architecturehigh
This recipe is part of the gentic.news Deployment Atlas. Every ingredient has an origin paper + evidence. Methodology is public. Dataset is CC BY 4.0.