Loading...

📰 Source: 4chan | /g/

✍️ Original author: Anonymous

⬆️ score: 664


/lmg/ - a general dedicated to the discussion and development of local language models.

Previous threads: &

►News

(04/01) Trinity-Large-Thinking released: https://hf.co/arcee-ai/Trinity-Large-Thinking (04/01) Merged llama : rotate activations for better quantization #21038: https://github.com/ggml-org/llama.cpp/pull/21038 (04/01) Holo3 VLMs optimized for GUI Agents released: https://hcompany.ai/holo3 (03/31) 1-bit Bonsai models quantized from Qwen 3: https://prismml.com/news/bonsai-8b (03/31) Claude Code’s source leaked via npm registry map file: https://github.com/instructkr/claude-code

►News Archive: https://rentry.org/lmg-news-archive ►Glossary: https://rentry.org/lmg-glossary ►Links: https://rentry.org/LocalModelsLinks

►Getting Started https://rentry.org/lmg-lazy-getting-started-guide https://rentry.org/lmg-build-guides https://rentry.org/IsolatedLinuxWebService https://rentry.org/recommended-models https://rentry.org/samplers

►Further Learning https://rentry.org/machine-learning-roadmap https://rentry.org/llm-training https://rentry.org/LocalModelsPapers

►Benchmarks LiveBench: https://livebench.ai/ Programming: https://livecodebench.github.io/gso.html Context Length: https://github.com/adobe-research/NoLiMa GPUs: https://github.com/XiongjieDai/GPU-Benchmarks-on-LLM-Inference

►Tools Alpha Calculator: https://desmos.com/calculator/ffngla98yc GGUF VRAM Calculator: https://hf.co/spaces/NyxKrage/LLM-Model-VRAM-Calculator Sampler Visualizer: https://artefact2.github.io/llm-sampling Token Speed Visualizer: https://shir-man.com/tokens-per-second

►Text Gen. UI, Inference Engines https://github.com/oobabooga/text-generation-webui https://github.com/LostRuins/koboldcpp https://github.com/ggerganov/llama.cpp https://github.com/theroyallab/tabbyAPI https://github.com/vllm-project/vllm

This post was automatically imported by OratioRepostBot.