Introducing HalluMix: A Task-Agnostic, Multi-Domain Benchmark for Detecting Hallucinations in Real-World Scenarios By quotientai and 3 others • 3 days ago • 18
Mixture of Tunable Experts - Behavior Modification of DeepSeek-R1 at Inference Time By rbrt and 4 others • Feb 18 • 31
Building Multimodal RAG Systems: Supercharging Retrieval with MultiModal Embeddings and LLMs By Omartificial-Intelligence-Space • 4 days ago • 6
Navigating the RLHF Landscape: From Policy Gradients to PPO, GAE, and DPO for LLM Alignment By NormalUhr • Feb 11 • 27
What is MoE 2.0? Update Your Knowledge about Mixture-of-experts By Kseniase and 1 other • 8 days ago • 5
Agent2Agent and MCP: An End-to-End Tutorial for a complete Agentic Pipeline By tsadoq • 6 days ago • 4
Introducing HalluMix: A Task-Agnostic, Multi-Domain Benchmark for Detecting Hallucinations in Real-World Scenarios By quotientai and 3 others • 3 days ago • 18
Mixture of Tunable Experts - Behavior Modification of DeepSeek-R1 at Inference Time By rbrt and 4 others • Feb 18 • 31
Building Multimodal RAG Systems: Supercharging Retrieval with MultiModal Embeddings and LLMs By Omartificial-Intelligence-Space • 4 days ago • 6
Navigating the RLHF Landscape: From Policy Gradients to PPO, GAE, and DPO for LLM Alignment By NormalUhr • Feb 11 • 27
What is MoE 2.0? Update Your Knowledge about Mixture-of-experts By Kseniase and 1 other • 8 days ago • 5
Agent2Agent and MCP: An End-to-End Tutorial for a complete Agentic Pipeline By tsadoq • 6 days ago • 4