SigLIP 2: Multilingual Vision-Language Encoders with Improved Semantic Understanding, Localization, and Dense Features Paper • 2502.14786 • Published Feb 20, 2025 • 156
SmolLM2: When Smol Goes Big -- Data-Centric Training of a Small Language Model Paper • 2502.02737 • Published Feb 4, 2025 • 253
view article Article KV Caching Explained: Optimizing Transformer Inference Efficiency Jan 30, 2025 • 209
multilingual vision models Collection Some papers I read for understanding vision models and also adding multilingual capabilities to them • 14 items • Updated Dec 11, 2024 • 2
Maya: An Instruction Finetuned Multilingual Multimodal Model Paper • 2412.07112 • Published Dec 10, 2024 • 28
M-RewardBench: Evaluating Reward Models in Multilingual Settings Paper • 2410.15522 • Published Oct 20, 2024 • 12
Transformer Explainer: Interactive Learning of Text-Generative Models Paper • 2408.04619 • Published Aug 8, 2024 • 174
Federated Learning driven Large Language Models for Swarm Intelligence: A Survey Paper • 2406.09831 • Published Jun 14, 2024 • 1
Cohere Labs Aya 23 Collection Aya 23 is an open weights research release of an instruction fine-tuned model with highly advanced multilingual capabilities. • 3 items • Updated Jul 31, 2025 • 56