My currently most used. These are all fully uncensored (no refusals).
Naphula
Naphula
AI & ML interests
Building new model tools, merges (7B-24B), ablations, finetunes, and datasets.
Recent Activity
new activity about 3 hours ago
24B-Suite/Mergedonia-Suite-24B-v1:Custom Methods (Index) reacted to Parveshiiii's post with 🔥 about 6 hours ago
Just did something I’ve been meaning to try for ages.
In only 3 hours, on 10 billion+ tokens, I trained a custom BPE + tiktoken-style tokenizer using my new library microtok — and it hits the same token efficiency as Qwen3.
Tokenizers have always felt like black magic to me. We drop them into every LLM project, but actually training one from scratch? That always seemed way too complicated.
Turns out it doesn’t have to be.
microtok makes the whole process stupidly simple — literally just 3 lines of code. No heavy setup, no GPU required. I built it on top of the Hugging Face tokenizers library so it stays clean, fast, and actually understandable.
If you’ve ever wanted to look under the hood and build your own optimized vocabulary instead of just copying someone else’s, this is the entry point you’ve been waiting for.
I wrote up the full story, threw in a ready-to-run Colab template, and dropped the trained tokenizer on Hugging Face.
Blog → https://parveshiiii.github.io/blogs/microtok/
Trained tokenizer → https://huggingface.co/Parveshiiii/microtok
GitHub repo → https://github.com/Parveshiiii/microtok
liked a model 1 day ago
Retreatcost/Chrysologus-12B