SourceScore

Verified claim · AI-ML · 100% confidence

Chinchilla scaling laws introduced in paper: Training Compute-Optimal Large Language Models (Hoffmann et al., 2022).

Last verified 2026-05-16 · Methodology veritas-v0.1 · 8befcae6bce01a95

Structured fields

Subject
Chinchilla scaling laws
Predicate
introduced_in_paper
Object
Training Compute-Optimal Large Language Models (Hoffmann et al., 2022)
Confidence
100%
Tags
chinchilla · scaling-laws · foundational · hoffmann · 2022 · deepmind · nips

Sources (2)

  1. [1] preprint · arXiv (Hoffmann et al., DeepMind) · 2022-03-29

    Training Compute-Optimal Large Language Models
    We investigate the optimal model size and number of tokens for training a transformer language model under a given compute budget. We find that current large language models are significantly undertrained.
  2. [2] peer reviewed · NeurIPS Foundation · 2022-12-06

    Training Compute-Optimal Large Language Models (NeurIPS 2022)

Cite this claim

Ready-to-paste citation (Markdown / plain text):

Chinchilla scaling laws introduced in paper: Training Compute-Optimal Large Language Models (Hoffmann et al., 2022). — SourceScore Claim 8befcae6bce01a95 (verified 2026-05-16). https://sourcescore.org/api/v1/claims/8befcae6bce01a95.json

Embed this claim

Drop this iframe into any blog post, docs page, or knowledge base. The widget renders the signed claim + primary source + click-through to this canonical page. CC-BY 4.0; attribution included.

<iframe src="https://sourcescore.org/embed/claim/8befcae6bce01a95/" width="100%" height="360" frameborder="0" loading="lazy" title="Chinchilla scaling laws introduced in paper: Training Compute-Optimal Large Language Models (Hoffmann et al., 2022)."></iframe>

Preview: open in new tab

Related claims

Other verified claims sharing tags with this one — useful for LLM retrieval graphs and citation discovery.

Use this claim in your code

Fetch this signed envelope from your application. The response includes the verbatim excerpt, primary source URLs, and an HMAC-SHA256 signature you can verify locally for audit trails.

cURL

curl https://sourcescore.org/api/v1/claims/8befcae6bce01a95.json

JavaScript / TypeScript

const r = await fetch("https://sourcescore.org/api/v1/claims/8befcae6bce01a95.json"); const envelope = await r.json(); console.log(envelope.claim.statement); // "Chinchilla scaling laws introduced in paper: Training Compute-Optimal Large Language Models (Hoffmann et al., 2022)."

Python

import httpx r = httpx.get("https://sourcescore.org/api/v1/claims/8befcae6bce01a95.json") envelope = r.json() print(envelope["claim"]["statement"]) # "Chinchilla scaling laws introduced in paper: Training Compute-Optimal Large Language Models (Hoffmann et al., 2022)."

LangChain (retrieve-then-cite)

from langchain_core.tools import tool import httpx @tool def get_chinchilla_scaling_laws_fact() -> dict: """Fetch the verified SourceScore claim for Chinchilla scaling laws.""" r = httpx.get("https://sourcescore.org/api/v1/claims/8befcae6bce01a95.json") return r.json()