Matt Oswalt
Codex
  • Linux
    • File Descriptors
    • Networking
      • eBPF
      • Sockets
  • LLM
    • Resources
    • Inference Stack
    • Apps & Libraries
    • Model Evaluation
    • Memory
    • Glossary
  • Machine Learning
    • Deep Learning
    • Machine Learning
    • Glossary
  • Math
    • Glossary
  • Rust
    • Common Traits
    • Ownership
  • Video
    • GoPro
  • Cheat Sheets
Matt Oswalt
  • Blogs
    • All Categories
    • Rust
    • General Programming
    • Systems
    • Machine Learning
    • Personal
  • Codex
  • Bookclub
  • Portfolio
  • Sponsor Me!
  • Github
  • Twitter
  • Twitch
  • LinkedIn
  • YouTube
  • Facebook
  • Bluesky
  • RSS

Search

Loading search index…

No recent searches

No results for "Query here"

  • to select
  • to navigate
  • to close

Search by FlexSearch

  • Linux
    • File Descriptors
    • Networking
      • eBPF
      • Sockets
  • LLM
    • Resources
    • Inference Stack
    • Apps & Libraries
    • Model Evaluation
    • Memory
    • Glossary
  • Machine Learning
    • Deep Learning
    • Machine Learning
    • Glossary
  • Math
    • Glossary
  • Rust
    • Common Traits
    • Ownership
  • Video
    • GoPro
  • Cheat Sheets

This Glossary

  • Context window
  • Dense model
  • Distilled model
  • Instruct model
  • KV cache
  • MoE (Mixture of Experts)
  • Open-weight model
  • Parameter
  • Reasoning model
  • RLHF (Reinforcement Learning from Human Feedback)
  • t/s (tokens per second)
  • Token

Parameter

← Back to Glossary

A single numerical value in a neural network that is learned during training. Model size is expressed in parameters — a “70B model” has 70 billion parameters. More parameters generally means more capable but also more memory-hungry.

  • Parameter - Google ML Glossary
Referenced in
  • Dense model
  • MoE (Mixture of Experts)
Prev
Open-weight model
Next
Q4_K_M
    • © 2010 - 2026 Matt Oswalt · Powered by Hugo & Hyas.