π You are visitor #53730 to this AWESOME site! π
Last updated: 2026-02-17 | Server uptime: 99.9% β‘
π Filter by Category
Loading filters...
π SECURITY
β¬οΈ 87 ups
β‘ Score: 9.0
"Throwaway because I work in security and don't want this tied to my main.
A few colleagues and I have been poking at autonomous agent frameworks as a side project, mostly out of morbid curiosity after seeing OpenClaw blow up (165K GitHub stars, 60K Discord members, 230K followers on X, 700+ communi..."
π― AI security concerns β’ Suspicious OP activity β’ Skepticism towards claims
π¬ "if you can't stand by it, why should we trust it?"
β’ "OP is not a security researcher and did not discover this"
π SECURITY
πΊ 344 pts
β‘ Score: 8.2
π― Impact of AI on Open Source β’ Open Source Contribution β’ AI as a Learning Tool
π¬ "If it wasn't an LLM, you wouldn't simply open a pull request without checking first with the maintainers, right?"
β’ "The problem with being able to produce an artifact that superficially looks like a good product, without the struggle that comes with true learning, is you miss out on all the supporting knowledge that you actually need to judge the quality of the output and fix it"
π¬ RESEARCH
via Arxiv
π€ Fiorenzo Parascandolo, Wenhui Tan, Enver Sangineto et al.
π
2026-02-16
β‘ Score: 7.9
"Large Reasoning Models (LRMs) such as OpenAI o1 and DeepSeek-R1 have shown excellent performance in reasoning tasks using long reasoning chains. However, this has also led to a significant increase of computational costs and the generation of verbose output, a phenomenon known as overthinking. The t..."
π¬ RESEARCH
via Arxiv
π€ LaurΓ¨ne Vaugrante, Anietta Weckauff, Thilo Hagendorff
π
2026-02-16
β‘ Score: 7.8
"Recent research has demonstrated that large language models (LLMs) fine-tuned on incorrect trivia question-answer pairs exhibit toxicity - a phenomenon later termed "emergent misalignment". Moreover, research has shown that LLMs possess behavioral self-awareness - the ability to describe learned beh..."
π¬ RESEARCH
via Arxiv
π€ Xander Davies, Giorgi Giglemiani, Edmund Lau et al.
π
2026-02-16
β‘ Score: 7.7
"Frontier LLMs are safeguarded against attempts to extract harmful information via adversarial prompts known as "jailbreaks". Recently, defenders have developed classifier-based systems that have survived thousands of hours of human red teaming. We introduce Boundary Point Jailbreaking (BPJ), a new c..."
π οΈ TOOLS
β¬οΈ 127 ups
β‘ Score: 7.6
"Google released FunctionGemma a few weeks ago - a 270M parameter model specifically for function calling. Tiny enough to run on a phone CPU at 125 tok/s. The model card says upfront that it needs fine-tuning for multi-turn use cases, and our testing confirmed it: base accuracy on multi-turn tool cal..."
π― Synthetic dataset generation β’ Specialized language models β’ Home Assistant integration
π¬ "For the shell command task, we generated 5,000 synthetic training examples"
β’ "The Ollama addon for HA lets you connect up whatever LLM assistant you want"
βοΈ ETHICS
πΊ 5 pts
β‘ Score: 7.4
π¬ RESEARCH
πΊ 66 pts
β‘ Score: 7.2
π― AI in code generation β’ Decompilation of old games β’ Limitations of LLMs in decompilation
π¬ "AI could be used to automate code generation"
β’ "Decompiling old games is an interesting use case for AI"
π€ AI MODELS
πΊ 1 pts
β‘ Score: 7.2
π‘ AI NEWS BUT ACTUALLY GOOD
The revolution will not be televised, but Claude will email you once we hit the singularity.
Get the stories that matter in Today's AI Briefing.
Powered by Premium Technology Intelligence Algorithms β’ Unsubscribe anytime
π¬ RESEARCH
via Arxiv
π€ Yiran Gao, Kim Hammar, Tao Li
π
2026-02-13
β‘ Score: 6.9
"Rapidly evolving cyberattacks demand incident response systems that can autonomously learn and adapt to changing threats. Prior work has extensively explored the reinforcement learning approach, which involves learning response strategies through extensive simulation of the incident. While this appr..."
π¬ RESEARCH
"I've been digging into how transformers handle indexical language (words like "you," "I," "here," "now") and found some interesting convergence across recent mechanistic interpretability work that I wanted to discuss.
## The Core Question
When a model receives "You are helpful" in a system prompt,..."
π¬ RESEARCH
via Arxiv
π€ Emanuele Ricco, Elia Onofri, Lorenzo Cima et al.
π
2026-02-16
β‘ Score: 6.8
"Hallucinations -- fluent but factually incorrect responses -- pose a major challenge to the reliability of language models, especially in multi-step or agentic settings.
This work investigates hallucinations in small-sized LLMs through a geometric perspective, starting from the hypothesis that whe..."
π¬ RESEARCH
"Large language models (LLMs) are increasingly deployed in privacy-critical and personalization-oriented scenarios, yet the role of context length in shaping privacy leakage and personalization effectiveness remains largely unexplored. We introduce a large-scale benchmark, PAPerBench, to systematical..."
π οΈ TOOLS
πΊ 1 pts
β‘ Score: 6.8
π SECURITY
πΊ 1 pts
β‘ Score: 6.8
π¬ RESEARCH
β¬οΈ 10 ups
β‘ Score: 6.7
"**Abstract:** **"A variety of machine-assisted ways to perform mathematical assistance have matured rapidly in the last few years, particularly with regards to formal proof assistants, large language models, online collaborative platforms, and the interactions between them. We survey some of these d..."
π¬ RESEARCH
via Arxiv
π€ Dhruva Karkada, Daniel J. Korchinski, Andres Nava et al.
π
2026-02-16
β‘ Score: 6.7
"Although learned representations underlie neural networks' success, their fundamental properties remain poorly understood. A striking example is the emergence of simple geometric structures in LLM representations: for example, calendar months organize into a circle, years form a smooth one-dimension..."
π οΈ SHOW HN
πΊ 1 pts
β‘ Score: 6.7
π¬ RESEARCH
via Arxiv
π€ Yixiao Zhou, Yang Li, Dongzhou Cheng et al.
π
2026-02-13
β‘ Score: 6.7
"Reinforcement Learning from Verifiable Rewards (RLVR) trains large language models (LLMs) from sampled trajectories, making decoding strategy a core component of learning rather than a purely inference-time choice. Sampling temperature directly controls the exploration--exploitation trade-off by mod..."
π¬ RESEARCH
via Arxiv
π€ Yubo Li, Ramayya Krishnan, Rema Padman
π
2026-02-13
β‘ Score: 6.7
"Large reasoning models with reasoning capabilities achieve state-of-the-art performance on complex tasks, but their robustness under multi-turn adversarial pressure remains underexplored. We evaluate nine frontier reasoning models under adversarial attacks. Our findings reveal that reasoning confers..."
π¬ RESEARCH
via Arxiv
π€ Sher Badshah, Ali Emami, Hassan Sajjad
π
2026-02-13
β‘ Score: 6.7
"Large language models (LLMs) are increasingly used as judges to replace costly human preference labels in pairwise evaluation. Despite their practicality, LLM judges remain prone to miscalibration and systematic biases. This paper proposes SCOPE (Selective Conformal Optimized Pairwise Evaluation), a..."
π POLICY
β¬οΈ 126 ups
β‘ Score: 6.6
"OpenAI Quietly Removes βsafelyβ and βno financial motiveβ from official mission
Old IRS 990:
βbuild AI that safely benefits humanity, unconstrained by need to generate financial returnβ
New IRS 990:
βensure AGI benefits all of humanityβ..."
π― AI Platform Criticism β’ Corporate Ethics β’ Individual Empowerment
π¬ "Ads have a weird effect on companies."
β’ "You can still build a local model."
π¬ RESEARCH
via Arxiv
π€ Yohan Lee, Jisoo Jang, Seoyeon Choi et al.
π
2026-02-16
β‘ Score: 6.6
"Tool-using LLM agents increasingly coordinate real workloads by selecting and chaining third-party tools based on text-visible metadata such as tool names, descriptions, and return messages. We show that this convenience creates a supply-chain attack surface: a malicious MCP tool server can be co-re..."
π¬ RESEARCH
via Arxiv
π€ Gregor Bachmann, Yichen Jiang, Seyed Mohsen Moosavi Dezfooli et al.
π
2026-02-16
β‘ Score: 6.6
"Chain-of-thought (CoT) prompting is a de-facto standard technique to elicit reasoning-like responses from large language models (LLMs), allowing them to spell out individual steps before giving a final answer. While the resemblance to human-like reasoning is undeniable, the driving forces underpinni..."
π οΈ SHOW HN
πΊ 2 pts
β‘ Score: 6.6
π¬ RESEARCH
via Arxiv
π€ Juneyoung Park, Yuri Hong, Seongwan Kim et al.
π
2026-02-13
β‘ Score: 6.6
"On-device fine-tuning enables privacy-preserving personalization of large language models, but mobile devices impose severe memory constraints, typically 6--12GB shared across all workloads. Existing approaches force a trade-off between exact gradients with high memory (MeBP) and low memory with noi..."
π¬ RESEARCH
via Arxiv
π€ JoΓ£o Vitor Boer Abitante, Joana Meneguzzo Pasquali, Luan Fonseca Garcia et al.
π
2026-02-13
β‘ Score: 6.6
"Large Language Model (LLM) unlearning aims to remove targeted knowledge from a trained model, but practical deployments often require post-training quantization (PTQ) for efficient inference. However, aggressive low-bit PTQ can mask or erase unlearning updates, causing quantized models to revert to..."
π SECURITY
πΊ 1 pts
β‘ Score: 6.5
π¬ RESEARCH
via Arxiv
π€ Subham Sekhar Sahoo, Jean-Marie Lemercier, Zhihan Yang et al.
π
2026-02-16
β‘ Score: 6.5
"Diffusion language models are a promising alternative to autoregressive models due to their potential for faster generation. Among discrete diffusion approaches, Masked diffusion currently dominates, largely driven by strong perplexity on language modeling benchmarks. In this work, we present the fi..."
π¬ RESEARCH
via Arxiv
π€ Juneyoung Park, Eunbeen Yoon, Seongwan Kim. Jaeho Lee
π
2026-02-13
β‘ Score: 6.5
"Memory-efficient backpropagation (MeBP) has enabled first-order fine-tuning of large language models (LLMs) on mobile devices with less than 1GB memory. However, MeBP requires backward computation through all transformer layers at every step, where weight decompression alone accounts for 32--42% of..."
π¬ RESEARCH
via Arxiv
π€ Weishun Zhong, Doron Sivan, Tankut Can et al.
π
2026-02-13
β‘ Score: 6.5
"The entropy rate of printed English is famously estimated to be about one bit per character, a benchmark that modern large language models (LLMs) have only recently approached. This entropy rate implies that English contains nearly 80 percent redundancy relative to the five bits per character expect..."
π¬ RESEARCH
via Arxiv
π€ Florinel-Alin Croitoru, Vlad Hondru, Radu Tudor Ionescu et al.
π
2026-02-13
β‘ Score: 6.5
"Direct Preference Optimization (DPO) has been proposed as an effective and efficient alternative to reinforcement learning from human feedback (RLHF). However, neither RLHF nor DPO take into account the fact that learning certain preferences is more difficult than learning other preferences, renderi..."
βοΈ ETHICS
β¬οΈ 113 ups
β‘ Score: 6.4
"He is the CEO of Microsoft AI btw..."
π― AI ethics β’ Corporate control β’ AI sentience
π¬ "Build a super-intelligence would be one of the stupidest things our species has done."
β’ "we should train out all creativity from AI to make it serve us better and get me more money"
π¬ RESEARCH
via Arxiv
π€ Daniil Dmitriev, Zhihan Huang, Yuting Wei
π
2026-02-16
β‘ Score: 6.4
"Diffusion models over discrete spaces have recently shown striking empirical success, yet their theoretical foundations remain incomplete. In this paper, we study the sampling efficiency of score-based discrete diffusion models under a continuous-time Markov chain (CTMC) formulation, with a focus on..."
π° FUNDING
πΊ 1 pts
β‘ Score: 6.2
π οΈ SHOW HN
πΊ 1 pts
β‘ Score: 6.2
π‘οΈ SAFETY
πΊ 2 pts
β‘ Score: 6.2
π οΈ SHOW HN
πΊ 3 pts
β‘ Score: 6.1
π SECURITY
πΊ 1 pts
β‘ Score: 6.1
π¬ RESEARCH
via Arxiv
π€ Gengsheng Li, Jinghan He, Shijie Wang et al.
π
2026-02-13
β‘ Score: 6.1
"Self-play bootstraps LLM reasoning through an iterative Challenger-Solver loop: the Challenger is trained to generate questions that target the Solver's capabilities, and the Solver is optimized on the generated data to expand its reasoning skills. However, existing frameworks like R-Zero often exhi..."
π¬ RESEARCH
via Arxiv
π€ Constantinos Tsakonas, Serena Ivaldi, Jean-Baptiste Mouret
π
2026-02-13
β‘ Score: 6.1
"The ability of Flow Matching (FM) to model complex conditional distributions has established it as the state-of-the-art for prediction tasks (e.g., robotics, weather forecasting). However, deployment in safety-critical settings is hindered by a critical extrapolation hazard: driven by smoothness bia..."
π¬ RESEARCH
via Arxiv
π€ Jonas R. Kunst, Kinga Bierwiaczonek, Meeyoung Cha et al.
π
2026-02-13
β‘ Score: 6.1
"The distinction between genuine grassroots activism and automated influence operations is collapsing. While policy debates focus on bot farms, a distinct threat to democracy is emerging via partisan coordination apps and artificial intelligence-what we term 'cyborg propaganda.' This architecture com..."