PaTH Attention: The Next Leap in Context-Aware Language Models Large language models have transformed artificial intelligence, powering everything from chatbots to automated code generation. Yet, even the most advanced models often struggle to follow evolving sta... AI research context awareness large language models machine learning PaTH Attention position encoding reasoning transformers
How DiscoRL Is Changing the Rules: AI That Discovers Its Own Learning Algorithms What if artificial intelligence could not only learn from experience but also invent the very rules that govern its learning, outpacing even the best human-crafted algorithms? Google DeepMind has take... AI research algorithm discovery automation DeepMind generalization meta-learning neural networks reinforcement learning
Edison Scientific Kosmos: Accelerating Six Months of Science in a Single Day Looking to get more out your research efforts without the timeline? Introducing Kosmos , Edison Scientific’s groundbreaking AI Scientist. Building on lessons from its predecessor, Robin, Kosmos ... AI research automation Edison Scientific Kosmos machine learning materials science neuroscience scientific discovery
Microsoft and Princeton University Spark a New Era of AI Innovation in New Jersey New Jersey is set to become a national leader in artificial intelligence, thanks to a groundbreaking partnership between Microsoft and Princeton University. As a founding partner of the New Jersey Art... AI research Discovery platform Innovation hub Microsoft New Jersey Princeton University Public-private partnership Workforce development
Apple Unveils FS-DFM: A Leap Forward in Fast, High-Quality AI Text Generation AI-generated writing has always faced a trade-off between speed and quality—until now. Apple, collaborating with Ohio State University, has introduced a transformative language model that can rapidly ... AI research Apple diffusion models FS-DFM language models machine learning natural language processing text generation
MIT is Making Large Language Model Training Affordable: Insights from AI Scaling Laws Training large language models (LLMs) requires immense computational resources and significant financial investment. For many AI researchers and organizations, predicting model performance while keepi... AI efficiency AI research budget optimization LLM training machine learning model evaluation scaling laws
Neural Networks Are Transforming 3D Rendering: Inside Microsoft's RenderFormer 3D rendering powers our most captivating digital experiences, from blockbuster movies to cutting-edge virtual reality. Traditionally, this field has relied on physics-based methods to recreate the int... 3D rendering AI research computer graphics deep learning machine learning neural networks RenderFormer SIGGRAPH
mmBERT: How Johns Hopkins Built a 1,833-Language AI That Outperforms XLM-R mmBERT: How Johns Hopkins Built a 1,833-Language AI That Outperforms XLM-R Imagine trying to build an AI system that truly understands human language not just in English, but in over 7,000 languages s... AI research annealed language learning cross-lingual digital inclusion encoder-only FlashAttention Gemma tokenizer GLUE inverse masking language model mmBERT ModernBERT MTEB multilingual NLP XLM-R XTREME
Test-Time Diffusion Deep Researcher: Ushering in a Human-Like AI Research Paradigm Introducing Test-Time Diffusion Deep Researcher (TTD-DR) framework an AI assistant that doesn't just gather information, but actively thinks, revises, and refines its work, much like a skilled human r... AI research draft refinement large language models multihop reasoning research automation retrieval augmentation self-evolution test-time diffusion