Skip to content
gentic.news — AI News Intelligence Platform

Recipe ·

GLM-5.1

GLM-5.1, developed by Zhipu AI, is a next-generation foundation model featuring a 1 million token context window and support for 128K output tokens.

7
Techniques inside
4y
Median research → prod
3y
Fastest adoption
5y
Slowest adoption

Ingredient list

  1. Invented by Nous Research · 2023-08 · Velocity 3y

    GLM-5.1 extends context length to 1M tokens using YaRN (Yet another RoPE extensioN) method.

    architecturehigh
  2. Invented by Google · 2023-05 · Velocity 3y

    GLM-5.1 architecture uses Grouped-Query Attention (GQA) to reduce KV cache memory.

    architecturehigh
  3. Invented by Stanford · 2022-05 · Velocity 4y

    GLM-5.1 implements FlashAttention-2 for efficient attention computation.

    inferencehigh
  4. Invented by Google · 2022-03 · Velocity 4y

    GLM-5.1 can use self-consistency by sampling multiple reasoning paths.

    reasoningmedium
  5. Invented by Google · 2022-01 · Velocity 4y

    GLM-5.1 demonstrates chain-of-thought reasoning capabilities in examples.

    reasoningmedium
  6. Invented by Google · 2021-09 · Velocity 5y

    GLM-5.1 is instruction-tuned on diverse tasks following FLAN methodology.

    trainingmedium
  7. Invented by Zhuiyi Technology · 2021-04 · Velocity 5y

    GLM-5.1 uses Rotary Position Embedding (RoPE) for positional encoding.

    architecturehigh

This recipe is part of the gentic.news Deployment Atlas. Every ingredient has an origin paper + evidence. Methodology is public. Dataset is CC BY 4.0.