How Reliable Are LLM Judges? Lessons from DataRobot's Evaluation Framework Relying on automated judges powered by Large Language Models (LLMs) to assess AI output may seem efficient, but it comes with hidden risks. LLM judges can be impressively confident even when they're w... AI benchmarking AI trust LLM evaluation machine learning open-source tools prompt engineering RAG systems
QuArch Puts AI Agents to the Test on Computer Architecture Computer architecture is having an AI moment. Yet despite rapid progress in agentic tooling for coding and verification, hardware-centric knowledge remains stubbornly hard for language models to maste... AI benchmarking Artificial Intelligence computer architecture Computer Science
Introducing LiveMCPBench: Evaluating Models on Large Tool Set Usage A new arXiv preprint, LiveMCPBench: Can Agents Navigate an Ocean of MCP Tools , from the Chinese Academy of Sciences and UCAS, introduces a benchmark to test AI agents in realistic tool-rich environme... AI benchmarking AI tools Artificial Intelligence MCP MCP Server