Patrick

pbarker

AI & ML interests

None yet

Recent Activity

new activity about 1 month ago
Qwen/Qwen3-32B:Is this multimodal?
updated a dataset 2 months ago
agentsea/Capybara-slim
published a dataset 2 months ago
agentsea/Capybara-slim
View all activity

Organizations

AgentSea's profile picture Hugging Face Discord Community's profile picture

pbarker's activity

New activity in Qwen/Qwen3-32B about 1 month ago

Is this multimodal?

1
#2 opened about 1 month ago by
pbarker
reacted to Jaward's post with ❤️ 4 months ago
view post
Post
3156
nanoBLT: Simplified lightweight implementation of a character-level Byte Latent Transformer model (under 500 lines of code). The model is 2x4x2 (n_layers_encoder, n_layers_latent, n_layers_decoder) layer deep trained on ~1M bytes of tiny Shakespeare with a patch size of 4.

Code: https://github.com/Jaykef/ai-algorithms/blob/main/byte_latent_transformer.ipynb