Common Pitfalls in Sharing Open Source Models on Hugging Face (and How to Dodge Them) By FriendliAI and 2 others • 6 days ago • 21
Should We Still Pretrain Encoders with Masked Language Modeling? By Nicolas-BZRD and 3 others • 5 days ago • 17
Why We Built the OpenMDW License: A Comprehensive License for ML Models By linuxfoundation • 5 days ago • 12
LLMs recognise bias but also reproduce harmful stereotypes: an analysis of bias in leading LLMs By davidberenstein1957 and 3 others • 6 days ago • 7
DeepSeek-R1 Dissection: Understanding PPO & GRPO Without Any Prior Reinforcement Learning Knowledge By NormalUhr • Feb 7 • 177
Welcome the NVIDIA Llama Nemotron Nano VLM to Hugging Face Hub By nvidia and 10 others • 10 days ago • 25
Announcing NeurIPS 2025 E2LM Competition: Early Training Evaluation of Language Models By tiiuae and 8 others • 3 days ago • 6
🅰️ℹ️ 1️⃣0️⃣1️⃣ **What is HtmlRAG, Multimodal RAG and Agentic RAG?** By Kseniase and 1 other • Jan 9 • 11
Common Pitfalls in Sharing Open Source Models on Hugging Face (and How to Dodge Them) By FriendliAI and 2 others • 6 days ago • 21
Should We Still Pretrain Encoders with Masked Language Modeling? By Nicolas-BZRD and 3 others • 5 days ago • 17
Why We Built the OpenMDW License: A Comprehensive License for ML Models By linuxfoundation • 5 days ago • 12
LLMs recognise bias but also reproduce harmful stereotypes: an analysis of bias in leading LLMs By davidberenstein1957 and 3 others • 6 days ago • 7
DeepSeek-R1 Dissection: Understanding PPO & GRPO Without Any Prior Reinforcement Learning Knowledge By NormalUhr • Feb 7 • 177
Welcome the NVIDIA Llama Nemotron Nano VLM to Hugging Face Hub By nvidia and 10 others • 10 days ago • 25
Announcing NeurIPS 2025 E2LM Competition: Early Training Evaluation of Language Models By tiiuae and 8 others • 3 days ago • 6
🅰️ℹ️ 1️⃣0️⃣1️⃣ **What is HtmlRAG, Multimodal RAG and Agentic RAG?** By Kseniase and 1 other • Jan 9 • 11