📰 Source: 4chan | /g/
✍️ Original author: Anonymous
⬆️ score: 664
/lmg/ - a general dedicated to the discussion and development of local language models.
Previous threads: &
►News
(04/01) Trinity-Large-Thinking released: https://hf.co/arcee-ai/Trinity-Large-Thinking (04/01) Merged llama : rotate activations for better quantization #21038: https://github.com/ggml-org/llama.cpp/pull/21038 (04/01) Holo3 VLMs optimized for GUI Agents released: https://hcompany.ai/holo3 (03/31) 1-bit Bonsai models quantized from Qwen 3: https://prismml.com/news/bonsai-8b (03/31) Claude Code’s source leaked via npm registry map file: https://github.com/instructkr/claude-code
►News Archive: https://rentry.org/lmg-news-archive ►Glossary: https://rentry.org/lmg-glossary ►Links: https://rentry.org/LocalModelsLinks
►Getting Started https://rentry.org/lmg-lazy-getting-started-guide https://rentry.org/lmg-build-guides https://rentry.org/IsolatedLinuxWebService https://rentry.org/recommended-models https://rentry.org/samplers
►Further Learning https://rentry.org/machine-learning-roadmap https://rentry.org/llm-training https://rentry.org/LocalModelsPapers
►Benchmarks LiveBench: https://livebench.ai/ Programming: https://livecodebench.github.io/gso.html Context Length: https://github.com/adobe-research/NoLiMa GPUs: https://github.com/XiongjieDai/GPU-Benchmarks-on-LLM-Inference
►Tools Alpha Calculator: https://desmos.com/calculator/ffngla98yc GGUF VRAM Calculator: https://hf.co/spaces/NyxKrage/LLM-Model-VRAM-Calculator Sampler Visualizer: https://artefact2.github.io/llm-sampling Token Speed Visualizer: https://shir-man.com/tokens-per-second
►Text Gen. UI, Inference Engines https://github.com/oobabooga/text-generation-webui https://github.com/LostRuins/koboldcpp https://github.com/ggerganov/llama.cpp https://github.com/theroyallab/tabbyAPI https://github.com/vllm-project/vllm
This post was automatically imported by OratioRepostBot.

