π HISTORICAL ARCHIVE - February 22, 2026
What was happening in AI on 2026-02-22
π You are visitor #47291 to this AWESOME site! π
Archive from: 2026-02-22 | Preserved for posterity β‘
π Filter by Category
Loading filters...
π SECURITY
πΊ 184 pts
β‘ Score: 8.3
π― AI-Assisted Security Audits β’ Backdoor Detection Challenges β’ Reverse Engineering Accessibility
π¬ "The whole field of working with binaries becomes accessible to a much wider range of software engineers."
β’ "Measuring false positives when you ask the model to complete a detection related task may be a good way of doing that."
π οΈ TOOLS
πΊ 504 pts
β‘ Score: 8.2
π― AI-assisted software development β’ Planning and documentation β’ Constraining AI model outputs
π¬ "I try these staging-document patterns, but suspect they have 2 fundamental flaws"
β’ "For me, I treat the LLM as model training + post processing + input tokens = output tokens"
π¬ RESEARCH
πΊ 3 pts
β‘ Score: 8.0
π OPEN SOURCE
β¬οΈ 21 ups
β‘ Score: 7.9
"nanollama β train Llama 3 from scratch.
I've been working on a framework for training Llama 3 architecture models from scratch: not fine-tuning, not LoRA, actual from-zero pretraining. The output is a llama.cpp-compatible GGUF file.
The whole pipeline is one command:
'''
bash runs/lambda\_trai..."
π― Easy data setup β’ Hardware performance β’ Local LLM development
π¬ "You should prepare for people example datasets"
β’ "this is localllama so i gotta ask"
π οΈ TOOLS
"Most discussions about RAG and LLM agents focus on βwhat architecture to useβ or βwhich model / vector store is betterβ. In practice, the systems I have seen fail in the same, very repetitive ways across projects, companies, and even different tech stacks.
Over the past years I have been debugging ..."
π οΈ TOOLS
β¬οΈ 8 ups
β‘ Score: 7.3
"I got Llama 3.2 1B running inference entirely on the AMD NPU on Linux. Every operation (attention, GEMM, RoPE, RMSNorm, SiLU, KV cache) runs on the NPU; no CPU or GPU fallback. As far as I can tell, this is the first time anyone has publicly documented this working on Linux.
## Hardware
- AMD Ryze..."
π€ AI MODELS
πΊ 2 pts
β‘ Score: 7.1
π SECURITY
πΊ 1 pts
β‘ Score: 7.0
π¬ RESEARCH
via Arxiv
π€ Lance Ying, Ryan Truong, Prafull Sharma et al.
π
2026-02-19
β‘ Score: 6.9
"Rigorously evaluating machine intelligence against the broad spectrum of human general intelligence has become increasingly important and challenging in this era of rapid technological advance. Conventional AI benchmarks typically assess only narrow capabilities in a limited range of human activity...."
π¬ RESEARCH
"Some transformer attention heads appear to function as membership testers, dedicating themselves to answering the question "has this token appeared before in the context?" We identify these heads across four language models (GPT-2 small, medium, and large; Pythia-160M) and show that they form a spec..."
π¬ RESEARCH
via Arxiv
π€ Dimitri Staufer, Kirsten Morehouse
π
2026-02-19
β‘ Score: 6.9
"Large language models (LLMs), and conversational agents based on them, are exposed to personal data (PD) during pre-training and during user interactions. Prior work shows that PD can resurface, yet users lack insight into how strongly models associate specific information to their identity. We audi..."
π‘ AI NEWS BUT ACTUALLY GOOD
The revolution will not be televised, but Claude will email you once we hit the singularity.
Get the stories that matter in Today's AI Briefing.
Powered by Premium Technology Intelligence Algorithms β’ Unsubscribe anytime
π¬ RESEARCH
via Arxiv
π€ Shayan Kiyani, Sima Noorani, George Pappas et al.
π
2026-02-19
β‘ Score: 6.8
"Reasoning with LLMs increasingly unfolds inside a broader verification loop. Internally, systems use cheap checks, such as self-consistency or proxy rewards, which we call weak verification. Externally, users inspect outputs and steer the model through feedback until results are trustworthy, which w..."
π¬ RESEARCH
via Arxiv
π€ Jyotin Goel, Souvik Maji, Pratik Mazumder
π
2026-02-19
β‘ Score: 6.8
"Instruction-following language models are trained to be helpful and safe, yet their safety behavior can deteriorate under benign fine-tuning and worsen under adversarial updates. Existing defenses often offer limited protection or force a trade-off between safety and utility. We introduce a training..."
π¬ RESEARCH
via Arxiv
π€ Yue Liu, Zhiyuan Hu, Flood Sung et al.
π
2026-02-19
β‘ Score: 6.8
"This paper introduces KLong, an open-source LLM agent trained to solve extremely long-horizon tasks. The principle is to first cold-start the model via trajectory-splitting SFT, then scale it via progressive RL training. Specifically, we first activate basic agentic abilities of a base model with a..."
π¬ RESEARCH
via Arxiv
π€ Jianda Du, Youran Sun, Haizhao Yang
π
2026-02-19
β‘ Score: 6.8
"PDEs are central to scientific and engineering modeling, yet designing accurate numerical solvers typically requires substantial mathematical expertise and manual tuning. Recent neural network-based approaches improve flexibility but often demand high computational cost and suffer from limited inter..."
π¬ RESEARCH
via Arxiv
π€ Sima Noorani, Shayan Kiyani, Hamed Hassani et al.
π
2026-02-19
β‘ Score: 6.7
"As humans increasingly rely on multiround conversational AI for high stakes decisions, principled frameworks are needed to ensure such interactions reliably improve decision quality. We adopt a human centric view governed by two principles: counterfactual harm, ensuring the AI does not undermine hum..."
π¬ RESEARCH
via Arxiv
π€ Shashank Aggarwal, Ram Vikas Mishra, Amit Awekar
π
2026-02-19
β‘ Score: 6.7
"In multi-agent IR pipelines for tasks such as search and ranking, LLM-based agents exchange intermediate reasoning in terms of Chain-of-Thought (CoT) with each other. Current CoT evaluation narrowly focuses on target task accuracy. However, this metric fails to assess the quality or utility of the r..."
π¬ RESEARCH
via Arxiv
π€ Xiaohan Zhao, Zhaoyi Li, Yaxin Luo et al.
π
2026-02-19
β‘ Score: 6.7
"Black-box adversarial attacks on Large Vision-Language Models (LVLMs) are challenging due to missing gradients and complex multimodal boundaries. While prior state-of-the-art transfer-based approaches like M-Attack perform well using local crop-level matching between source and target images, we fin..."
π¬ RESEARCH
via Arxiv
π€ Payel Bhattacharjee, Osvaldo Simeone, Ravi Tandon
π
2026-02-19
β‘ Score: 6.7
"Reward modeling is a core component of modern alignment pipelines including RLHF and RLAIF, underpinning policy optimization methods including PPO and TRPO. However, training reliable reward models relies heavily on human-labeled preference data, which is costly and limited, motivating the use of da..."
π¬ RESEARCH
via Arxiv
π€ Baihe Huang, Eric Xu, Kannan Ramchandran et al.
π
2026-02-19
β‘ Score: 6.6
"The proliferation of Large Language Models (LLMs) necessitates efficient mechanisms to distinguish machine-generated content from human text. While statistical watermarking has emerged as a promising solution, existing methods suffer from two critical limitations: the lack of a principled approach f..."
π¬ RESEARCH
via Arxiv
π€ Faria Huq, Zora Zhiruo Wang, Zhanqiu Guo et al.
π
2026-02-19
β‘ Score: 6.6
"Despite rapid progress in autonomous web agents, human involvement remains essential for shaping preferences and correcting agent behavior as tasks unfold. However, current agentic systems lack a principled understanding of when and why humans intervene, often proceeding autonomously past critical d..."
π¬ RESEARCH
via Arxiv
π€ Luke Huang, Zhuoyang Zhang, Qinghao Hu et al.
π
2026-02-19
β‘ Score: 6.6
"Reinforcement learning (RL) is widely used to improve large language models on reasoning tasks, and asynchronous RL training is attractive because it increases end-to-end throughput. However, for widely adopted critic-free policy-gradient methods such as REINFORCE and GRPO, high asynchrony makes the..."
π SECURITY
πΊ 1 pts
β‘ Score: 6.2
π¬ RESEARCH
πΊ 2 pts
β‘ Score: 6.2
π SECURITY
πΊ 2 pts
β‘ Score: 6.2
β‘ BREAKTHROUGH
β¬οΈ 8 ups
β‘ Score: 6.2
"Time series foundation models like Chronos-2 have been hyped recently for their ability to forecast zero-shot from arbitrary time series segments presented "in-context". But they are essentially based on statistical pattern matching -- in contrast, DynaMix ([
https://neurips.cc/virtual/2025/loc/san-d..."
π¬ RESEARCH
β¬οΈ 3 ups
β‘ Score: 6.1
"Following up on our DynaMix #NeurIPS2025 paper (see link below), the first foundation model for dynamical systems reconstruction, we have now
\- included **comparisons to most recent time series FMs like Chronos-2** in the latest update ([
https://neurips.cc/virtual/2025/loc/san-diego/poster/118041]..."
π¬ RESEARCH
via Arxiv
π€ Hojung Jung, Rodrigo Hormazabal, Jaehyeong Jo et al.
π
2026-02-19
β‘ Score: 6.1
"Molecular generation with diffusion models has emerged as a promising direction for AI-driven drug discovery and materials science. While graph diffusion models have been widely adopted due to the discrete nature of 2D molecular graphs, existing models suffer from low chemical validity and struggle..."
π¬ RESEARCH
"Current speech LLMs largely perform implicit ASR: on tasks solvable from a transcript, they are behaviorally and mechanistically equivalent to simple Whisper$\to$LLM cascades. We show this through matched-backbone testing across four speech LLMs and six tasks, controlling for the LLM backbone for th..."
π¨ CREATIVE
β¬οΈ 2190 ups
β‘ Score: 6.0
"I was trying to see if I could create a coherent character through multiple images with a background that maintains continuity. It did generally well although if look closely objects shift around slightly.
Each image was generated using the same prompt more or less (collage vs single image) but was..."
π― Dating app behavior β’ AI-generated images β’ Bartender's perspective
π¬ "Dating apps are fucking cooked chat"
β’ "Maybe inconsistency is actually what we should be looking for to find real people"