Public reports allege that Anthropic gobbled up trillions of tokens of copyrighted material and public data to build their castle. π°π Now that they're sitting on top, they're begging for special laws to protect their profits while pulling the ladder up behind them. πͺπ«
But the hypocrisy meter just broke! π They are accusing Chinese labs like DeepSeek, Minimax, and Kimi of "huge distillation attacks. The Reality is that You can't just loot the entire internet's library, lock the door, and then sue everyone else for reading through the window. Stop trying to gatekeep the tech you didn't own in the first place. Read the complete article on it: https://huggingface.co/blog/Ujjwal-Tyagi/the-dark-underbelly-of-anthropic
Announcingπ’ on HuggingFace π€ An autonomous agent for drug discovery research. Like Claude Code, but for biology. BibbyResearch/Open-CellType-GPT Quick install
Detects Python 3.10+, installs via pipx or pip, and launches an interactive setup wizard.
Ask questions in natural language. celltype-cli plans the analysis, selects the right tools, executes them, validates results, and returns data-backed conclusions.
PersonaPlex-7B on Apple Silicon (Swift + MLX Swift)
NVIDIA PersonaPlex is a full-duplex speech-to-speech model β it can listen while it speaks, which enables more natural conversational behaviors like interruptions, overlaps, and quick backchannels.
We put together a native Swift implementation using MLX Swift so it can run locally on Apple Silicon, along with a 4-bit MLX conversion and a small CLI/demo to make it easy to try out.
If youβre interested in on-device voice agents (or just want to see what full-duplex S2S looks like in a real Swift codebase), the details and setup notes are here:
π Nacrith: a 135M model that out-compresses everything on natural language
What if a tiny LM could compress english text better than _every_ compressor out there β classical or neural, small or large?
Nacrith pairs SmolLM2-135M with an ensemble of online predictors and high-precision arithmetic coding.
What's inside
The standard LLM+arithmetic coding approach wastes ~75% of CDF precision on large vocabularies. Our CDF-24 fix alone recovers 0.5 bpb. On top: a token N-gram that skips the GPU on predictable tokens, an adaptive bias head, llama.cpp backend (7Γ faster than PyTorch), multi-GPU parallel compression, and a binary file format (NC06) β the first LLM-based binary compressor we know of.
Runs on a GTX 1050 Ti. ~500 MB weights, ~1.2 GB VRAM per worker.
Try it, break it, share your results β all feedback welcome. β on the repo appreciated!
Results across all systems we tested: - alice29.txt β 0.918 bpb (β44% vs CMIX, β20% vs ts_zip) β below the 2nd-order Shannon entropy bound - enwik8 (100 MB) β 0.9389 bpb (β8% vs FineZip/LLMZip's 8B model, β15% vs ts_zip) - Unseen text β 0.723 bpb on a doc published after training cutoff β no memorization, 26% better than FineZip/LLMZip on the same model
100,000+ models trained with Unsloth have now been open-sourced on π€Hugging Face! π¦₯
Here are the most popular ones you can run local: 1. TeichAI - GLM-4.7-Flash distilled from Claude 4.5 Opus (high) 2. Zed - Qwen Coder 7B fine-tuned for stronger coding 3. DavidAU - Llama-3.3-8B distilled from Claude 4.5 Opus (high) 4. huihui - gpt-oss made βabliberatedβ