Expert Parallelism Bookmarks | William Callahan - Bookmarks
Discover More
dev.synthetic.new
Synthetic LLM Hosted Models
Chat with open-source models privately
z.ai
GLM-5: From Vibe Coding to Agentic Engineering
GLM-5 is a 744B-parameter MoE model (40B active) from Zhipu AI, scaled up from GLM-4.5's 355B with 28.5T pre-training tokens and DeepSeek Sparse Atten...
openrouter.ai
Trinity Mini (free) - API, Providers, Stats
Trinity Mini is a 26B-parameter (3B active) sparse mixture-of-experts language model featuring 128 experts with 8 active per token. Engineered for eff...