π§ LFM2 Collection LFM2 is a new generation of hybrid models, designed for on-device deployment. β’ 21 items β’ Updated about 3 hours ago β’ 103
view article Article Tiny Agents in Python: a MCP-powered agent in ~70 lines of code By celinah and 3 others β’ May 23 β’ 164
view article Article Vision Language Models (Better, Faster, Stronger) By merve and 4 others β’ May 12 β’ 538
Executable Code Actions Elicit Better LLM Agents Paper β’ 2402.01030 β’ Published Feb 1, 2024 β’ 171
DINOv2 Collection DINOv2: foundation models producing robust visual features suitable for image-level and pixel-level visual tasks - https://arxiv.org/abs/2304.07193 β’ 5 items β’ Updated Aug 13 β’ 27
SmolLM2 Collection State-of-the-art compact LLMs for on-device applications: 1.7B, 360M, 135M β’ 16 items β’ Updated May 5 β’ 291
SigLIP 2: Multilingual Vision-Language Encoders with Improved Semantic Understanding, Localization, and Dense Features Paper β’ 2502.14786 β’ Published Feb 20 β’ 148
view article Article Ο0 and Ο0-FAST: Vision-Language-Action Models for General Robot Control By danaaubakirova and 3 others β’ Feb 4 β’ 176
Phi-3 Collection Phi-3 family of small language and multi-modal models. Language models are available in short- and long-context lengths. β’ 26 items β’ Updated May 1 β’ 574
SmolLM2: When Smol Goes Big -- Data-Centric Training of a Small Language Model Paper β’ 2502.02737 β’ Published Feb 4 β’ 242
view article Article Open-source DeepResearch β Freeing our search agents By m-ric and 4 others β’ Feb 4 β’ 1.3k
view article Article Open-R1: a fully open reproduction of DeepSeek-R1 By eliebak and 2 others β’ Jan 28 β’ 880
view article Article Run the strongest open-source LLM model: Llama3 70B with just a single 4GB GPU! By lyogavin β’ Apr 21, 2024 β’ 44
Open LLM Leaderboard best models β€οΈβπ₯ Collection A daily uploaded list of models with best evaluations on the LLM leaderboard: β’ 65 items β’ Updated Mar 20 β’ 643
LLaVA-Video Collection Models focus on video understanding (previously known as LLaVA-NeXT-Video). β’ 8 items β’ Updated Feb 21 β’ 63
Building and better understanding vision-language models: insights and future directions Paper β’ 2408.12637 β’ Published Aug 22, 2024 β’ 133
DocLLM: A layout-aware generative language model for multimodal document understanding Paper β’ 2401.00908 β’ Published Dec 31, 2023 β’ 188
Improving Text Embeddings with Large Language Models Paper β’ 2401.00368 β’ Published Dec 31, 2023 β’ 82