ML Notes

home

❯

modules

❯

output function

❯

Logit Soft Capping

Logit Soft-Capping

07 Apr 20251 min read

Logit soft-capping uses tanh to clamp the logits x between (−t,t) in a smooth and differentiable way:

softcap(x)=t⋅tanh(tx​)

References

  • Gemma 2: Improving Open Language Models at a Practical Size
  • Neural Combinatorial Optimisation with Reinforcement Learning

Graph View

Backlinks

  • Gemma 2

Created with Quartz v4.5.0 © 2025