Skip to content

docs: clarify 100B refers to training tokens, not model size#443

Open
JasonOA888 wants to merge 1 commit intomicrosoft:mainfrom
JasonOA888:docs/clarify-100b-training-tokens
Open

docs: clarify 100B refers to training tokens, not model size#443
JasonOA888 wants to merge 1 commit intomicrosoft:mainfrom
JasonOA888:docs/clarify-100b-training-tokens

Conversation

@JasonOA888
Copy link

Summary

Fixes #391 - Clarifies that the 100B mentioned in the README refers to training tokens, not model parameters.

Problem

The original text run a 100B BitNet b1.58 model is frequently misinterpreted as referring to a 100B parameter model, when it actually means a model trained on 100B tokens.

This has become a yearly rite of passage for engagement farming accounts to misinterpret this section.

Change

  • Before: run a 100B BitNet b1.58 model
  • After: run a BitNet b1.58 model trained on 100B tokens

Testing

  • Verified README.md renders correctly
  • No code changes, documentation only

Closes #391

This fixes confusion where readers misinterpret '100B model'
as referring to parameter count, when it actually refers to
training tokens. The change makes this distinction clear:
- Before: 'run a 100B BitNet b1.58 model'
- After: 'run a BitNet b1.58 model trained on 100B tokens'

Closes microsoft#391

Co-authored-by: Jason L <jason@outland.art>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

Please consider rewording section about 100B training tokens

1 participant