WikiContributor · Talk · Preferences · Watchlist · Contributions · Log out
Article Talk 2 Read Edit source View history

Editing Large language model

Section: History and development  ·  Edit the full article instead

Before saving: Verify all facts against reliable sources. Do not copy copyrighted text. All new claims require inline citations. See the content assessment talk for current improvement priorities.
Active talk page discussions 2
▲ collapse
Proposed restructure of "Training data" section 2 comments · last edited 6 hours ago
ResearchContrib88 03:14, 4 April 2026 (UTC)

The current "Training data" section feels fragmented — it mixes pre-training corpus details with RLHF specifics in a way that's hard to follow. I'd suggest splitting into two subsections: one for pre-training corpus composition and one for alignment techniques. This would match the structure used on the BERT and GPT-3 articles. Any objections?

NLPeditor_K 05:52, 4 April 2026 (UTC)

Agree with the split. The RLHF subsection should also reference the Constitutional AI paper (Bai et al., 2022) — it's currently missing from this section. Happy to draft the subsection if nobody else is actively editing it.

Fact check: founding date and parameter count in lead section 1 comment · posted 1 hour ago
FactualEditor2024 22:08, 4 April 2026 (UTC)

{fill}

/* History and development */

Briefly describe what you changed and why. How to write a good edit summary

By publishing changes, you agree to the Terms of Use and irrevocably release your contribution under the CC BY-SA 4.0 License and the GFDL. A hyperlink or URL is sufficient attribution under the Creative Commons license.

Editing help · Wikipedia policies · Manual of Style · Citing sources · Accessibility guidelines