Skip to content

data: Llama 4 Scout BF16-direct shards (full downloads, 78.4 MB total)#69

Merged
AdaWorldAPI merged 1 commit into
masterfrom
claude/transcode-deepnsm-rust-oNa1Z
Mar 31, 2026
Merged

data: Llama 4 Scout BF16-direct shards (full downloads, 78.4 MB total)#69
AdaWorldAPI merged 1 commit into
masterfrom
claude/transcode-deepnsm-rust-oNa1Z

Conversation

@AdaWorldAPI
Copy link
Copy Markdown
Owner

Replaces partial old-pipeline shards with complete BF16-direct + F64x8 runs. All 5 shards streamed from HuggingFace with segment cache + retry.

Shard 1: 22 MB (layers 0-10 + embeddings, 117 tensors)
Shard 2: 12 MB (layers 11-21, ~110 tensors)
Shard 3: 24 MB (layers 22-32, 126 tensors)
Shard 4: 13 MB (layers 33-43)
Shard 5: 7.4 MB (layers 44-47 + output, 40 tensors)
OpenChat: 41 MB (7B Q8_0, 226 tensors)

Peak RAM: 134 MB. Compression: Attention 361×, FFN 5939×, Embedding 23770×.

https://claude.ai/code/session_01Y69Vnw751w75iVSBRws7o7

Replaces partial old-pipeline shards with complete BF16-direct + F64x8 runs.
All 5 shards streamed from HuggingFace with segment cache + retry.

Shard 1: 22 MB (layers 0-10 + embeddings, 117 tensors)
Shard 2: 12 MB (layers 11-21, ~110 tensors)
Shard 3: 24 MB (layers 22-32, 126 tensors)
Shard 4: 13 MB (layers 33-43)
Shard 5: 7.4 MB (layers 44-47 + output, 40 tensors)
OpenChat: 41 MB (7B Q8_0, 226 tensors)

Peak RAM: 134 MB. Compression: Attention 361×, FFN 5939×, Embedding 23770×.

https://claude.ai/code/session_01Y69Vnw751w75iVSBRws7o7
@chatgpt-codex-connector
Copy link
Copy Markdown

You have reached your Codex usage limits for code reviews. You can see your limits in the Codex usage dashboard.
To continue using code reviews, you can upgrade your account or add credits to your account and enable them for code reviews in your settings.

@AdaWorldAPI AdaWorldAPI merged commit 95e9c78 into master Mar 31, 2026
4 of 10 checks passed
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants