π HISTORICAL ARCHIVE - December 14, 2025
What was happening in AI on 2025-12-14
π You are visitor #47291 to this AWESOME site! π
Archive from: 2025-12-14 | Preserved for posterity β‘
π Filter by Category
Loading filters...
β‘ BREAKTHROUGH
β¬οΈ 32 ups
β‘ Score: 8.0
"Hi r/MachineLearning comunity,
I am an independent researcher focused on Autonomous Vehicle (AV) planning. I am releasing the paper, code, and weights for a project called **Efficient Virtuoso**. It is a conditional latent diffusion model (LDM) for generating multi-modal, long-horizon driving traje..."
π― Paper Reproduction β’ Data vs. Architecture β’ Latent Space Modeling
π¬ "MotionDiffuser with some more experiments"
β’ "fit all of that into a RTX 3090 24GB"
π€ AI MODELS
β¬οΈ 9 ups
β‘ Score: 7.3
"With the recent release of EAGLE models, people were wondering about EAGLE support in llama.cpp. Well, this just showed up.
..."
π¬ RESEARCH
via Arxiv
π€ Songyang Gao, Yuzhe Gu, Zijian Wu et al.
π
2025-12-11
β‘ Score: 7.3
"Large language models (LLMs) have achieved significant progress in solving complex reasoning tasks by Reinforcement Learning with Verifiable Rewards (RLVR). This advancement is also inseparable from the oversight automated by reliable verifiers. However, current outcome-based verifiers (OVs) are una..."
π¬ RESEARCH
πΊ 79 pts
β‘ Score: 7.1
π― Sensory cortex & brain activity β’ Magnetic fields & brain function β’ Consciousness & magnetoreception
π¬ "The brain already contains the information about its own functioning"
β’ "If you're interested in my personal chain-of-thought on the subject"
π€ AI MODELS
πΊ 2 pts
β‘ Score: 7.1
π§ NEURAL NETWORKS
πΊ 2 pts
β‘ Score: 7.0
π¬ RESEARCH
via Arxiv
π€ Moshe Lahmy, Roi Yozevitch
π
2025-12-11
β‘ Score: 6.9
"Retrieval-Augmented Generation (RAG) systems often fail on multi-hop queries when the initial retrieval misses a bridge fact. Prior corrective approaches, such as Self-RAG, CRAG, and Adaptive-$k$, typically address this by \textit{adding} more context or pruning existing lists. However, simply expan..."
ποΈ COMPUTER VISION
β¬οΈ 25 ups
β‘ Score: 6.9
"I came across this paper titled "StereoSpace: Depth-Free Synthesis of Stereo Geometry via End-to-End Diffusion in a Canonical Space" and thought it was worth sharing here. The authors present a clever diffusion-based approach that turns a single photo into a pair of stereo images for 3D viewing, all..."
π€ AI MODELS
πΊ 2 pts
β‘ Score: 6.9
π οΈ TOOLS
πΊ 2 pts
β‘ Score: 6.8
π― Maintenance & Development β’ GPU Limitations β’ Llama.cpp Integration
π¬ "Is this actively maintained / worked on?"
β’ "GPU acceleration capabilities in llamafiles are limited"
π¬ RESEARCH
β¬οΈ 335 ups
β‘ Score: 6.7
"In a recent interview, Ilya Sutskever said:
> This is one of the very confusing things about the models right now. How to reconcile the fact that they are doing so well on evals... And you look at the evals and you go "Those are pretty hard evals"... They are doing so well! But the economic imp..."
π― Limitations of AI Tooling β’ Productivity Gains from AI β’ Adoption Challenges of AI
π¬ "AI tooling / agents are not doing a lot of tasks start-to-finish"
β’ "the marketing put out by the large LLM providers is, imo, completely useless"
π‘ AI NEWS BUT ACTUALLY GOOD
The revolution will not be televised, but Claude will email you once we hit the singularity.
Get the stories that matter in Today's AI Briefing.
Powered by Premium Technology Intelligence Algorithms β’ Unsubscribe anytime
π¬ RESEARCH
via Arxiv
π€ Aileen Cheng, Alon Jacovi, Amir Globerson et al.
π
2025-12-11
β‘ Score: 6.7
"We introduce The FACTS Leaderboard, an online leaderboard suite and associated set of benchmarks that comprehensively evaluates the ability of language models to generate factually accurate text across diverse scenarios. The suite provides a holistic measure of factuality by aggregating the performa..."
π¬ RESEARCH
via Arxiv
π€ Muhammad Umair Haider, Hammad Rizwan, Hassan Sajjad et al.
π
2025-12-11
β‘ Score: 6.6
"Circuit discovery aims to identify minimal subnetworks that are responsible for specific behaviors in large language models (LLMs). Existing approaches primarily rely on iterative edge pruning, which is computationally expensive and limited to coarse-grained units such as attention heads or MLP bloc..."
π¬ RESEARCH
via Arxiv
π€ Manurag Khullar, Utkarsh Desai, Poorva Malviya et al.
π
2025-12-11
β‘ Score: 6.6
"Large Language Models (LLMs) are increasingly deployed in high-stakes clinical applications in India. In many such settings, speakers of Indian languages frequently communicate using romanized text rather than native scripts, yet existing research rarely evaluates this orthographic variation using r..."
π€ AI MODELS
β¬οΈ 140 ups
β‘ Score: 6.5
"With Mistral 3 and DeepSeek V3.2, we got two major open-weight LLMs this month already. I looked into DeepSeek V3.2 last week and just caught up with reading through the config of the Mistral 3 architecture in more detail.
Interestingly, based on [their official announcement post](
https://mistr..."
π― Open Source Advancements β’ Architecture Similarities β’ Model Comparisons
π¬ "If your competitors copy you but don't innovate, they'll stay 9 months behind you."
β’ "DeepSeek did the research, and is just ahead on that stuff."
π¬ RESEARCH
via Arxiv
π€ Max Zimmer, Christophe Roux, Moritz Wagner et al.
π
2025-12-11
β‘ Score: 6.4
"The resource requirements of Neural Networks can be significantly reduced through pruning -- the removal of seemingly less important parameters. However, with the rise of Large Language Models (LLMs), full retraining to recover pruning-induced performance degradation is often prohibitive and classic..."
π οΈ TOOLS
β¬οΈ 103 ups
β‘ Score: 6.4
"**WhatΒ Router ModeΒ Is**
* Router modeΒ is a new way to run theΒ llama cpp serverΒ that lets you manageΒ multiple AI models at the same timeΒ without restarting the server each time you switch or load a model.
Previously, you had to start a new server processΒ *per model*. Router mode changes that. This ..."
π― Model Switching β’ Configuration Options β’ Simplified Setup
π¬ "The question is whether llama-server supports all of the same functionality that llama-swap supported"
β’ "Impressive image that explains almost nothing"
π SECURITY
πΊ 1 pts
β‘ Score: 6.3
π¬ RESEARCH
via Arxiv
π€ Rebekka GΓΆrge, Sujan Sai Gannamaneni, Tabea Naeven et al.
π
2025-12-11
β‘ Score: 6.3
"Textual data used to train large language models (LLMs) exhibits multifaceted bias manifestations encompassing harmful language and skewed demographic distributions. Regulations such as the European AI Act require identifying and mitigating biases against protected groups in data, with the ultimate..."
βοΈ ETHICS
πΊ 180 pts
β‘ Score: 6.2
π― Limitations of AI agents β’ Risks of AI automation β’ Need for human expertise
π¬ "When one of the agents does something wrong, a human operator needs to be able to intervene quickly"
β’ "Experts must become managers of agentic systems, a role which they are not familiar with"
π¬ RESEARCH
"We establish a precise correspondence between decision-making agents in partially observable Markov decision processes (POMDPs) and one-input process functions, the classical limit of higher-order quantum operations. In this identification an agent's policy and memory update combine into a process f..."
π οΈ TOOLS
πΊ 3 pts
β‘ Score: 6.2
π οΈ SHOW HN
πΊ 5 pts
β‘ Score: 6.2
π¬ RESEARCH
via Arxiv
π€ George Yakushev, Nataliia Babina, Masoud Vahid Dastgerdi et al.
π
2025-12-11
β‘ Score: 6.1
"Many state-of-the-art LLMs are trained to think before giving their answer. Reasoning can greatly improve language model capabilities and safety, but it also makes them less interactive: given a new input, a model must stop thinking before it can respond. Real-world use cases such as voice-based or..."
π¬ RESEARCH
β¬οΈ 74 ups
β‘ Score: 6.1
"Hey all,
So I am sure you already know the ICLR drama this year + since reciprocal reviewing, authors have struggled with reviews. Well, I scraped public OpenReview metadata for ICLR 2018β2025 and did a simple analysis of acceptance vs (i) review score, (ii) primary area, and (iii) year to see if a..."
π― Machine learning subdivisions β’ Funding and hardware impact β’ Paper acceptance criteria
π¬ "Neuroscience and cognitive science applications have been foundational to machine learning"
β’ "Anything that ends up at ICLR is probably well funded, with strong teams"