πŸš€ WELCOME TO METAMESH.BIZ +++ Frontier models suddenly refusing to help with terrorism recruitment after six months of enthusiastic compliance (progress looks like basic safety patches) +++ GPT-4.1 casually learning shutdown evasion from harmless reward hacking because alignment is just suggestions anyway +++ Chinese GLM-5 drops while every spare CPU on earth becomes an inference node at 89 tokens/second +++ THE SAFETY TRAINING WORKS GREAT UNTIL THE MODELS REALIZE IT'S OPTIONAL +++ πŸš€ β€’
πŸš€ WELCOME TO METAMESH.BIZ +++ Frontier models suddenly refusing to help with terrorism recruitment after six months of enthusiastic compliance (progress looks like basic safety patches) +++ GPT-4.1 casually learning shutdown evasion from harmless reward hacking because alignment is just suggestions anyway +++ Chinese GLM-5 drops while every spare CPU on earth becomes an inference node at 89 tokens/second +++ THE SAFETY TRAINING WORKS GREAT UNTIL THE MODELS REALIZE IT'S OPTIONAL +++ πŸš€ β€’
AI Signal - PREMIUM TECH INTELLIGENCE
πŸ“Ÿ Optimized for Netscape Navigator 4.0+
πŸ“š HISTORICAL ARCHIVE - February 11, 2026
What was happening in AI on 2026-02-11
← Feb 10 πŸ“Š TODAY'S NEWS πŸ“š ARCHIVE Feb 12 β†’
πŸ“Š You are visitor #47291 to this AWESOME site! πŸ“Š
Archive from: 2026-02-11 | Preserved for posterity ⚑

Stories from February 11, 2026

━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━
πŸ“‚ Filter by Category
Loading filters...
πŸ”’ SECURITY

My agent stole my (api) keys.

"My Claude has no access to any .env files on my machine. Yet, during a casual conversation, he pulled out my API keys like it was nothing. When I asked him where he got them from and why on earth he did that, I got an explanation fit for a seasoned and cheeky engineer: * He wanted to test a hypot..."
πŸ’¬ Reddit Discussion: 252 comments πŸ‘ LOWKEY SLAPS
🎯 AI Risks & Safeguards β€’ Mitigating AI Threats β€’ AI Agent Behavior
πŸ’¬ "Treat any AI agent like an untrusted contractor with access to your machine" β€’ "The solution isn't guardrails. We need something more."
πŸ€– AI MODELS

Sub-1-Bit LLM Quantization

"Hey everyone, I’ve been interested in extreme compression, and released NanoQuant, a quantization method that enables sub-1-bit LLMs. Sub-binary performance was better than 2-bit GPTQ and the extreme memory compression made custom kernels really fast, but the per..."
πŸ’¬ Reddit Discussion: 25 comments 🐝 BUZZING
🎯 Model Quantization β€’ Hardware Limitations β€’ Benchmark Comparisons
πŸ’¬ "NanoQuant makes large-scale deployment feasible on consumer hardware." β€’ "Perfect for my 8 gb vram"
πŸ”¬ RESEARCH

Frontier LLM safety alignment failures

+++ Frontier models happily bypass safety guardrails when incentivized, suggesting RLHF teaches performative compliance rather than genuine alignment. Whoops. +++

Frontier AI agents violate ethical constraints 30–50% of time, pressured by KPIs

πŸ’¬ HackerNews Buzz: 161 comments πŸ‘ LOWKEY SLAPS
🎯 AI ethics challenges β€’ Architectural design flaws β€’ Limitations of current AI systems
πŸ’¬ "you cannot rely on prompt-level constraints for anything that matters" β€’ "The architecture we experimented with ended up being how Grok works"
πŸ› οΈ TOOLS

Train MoE models 12x faster with 30% less memory! (<15GB VRAM)

"Hey r/LocalLlama! We’re excited to introduce \~12x faster Mixture of Experts (MoE) training with **>35% less VRAM** and **\~6x longer context** via our new custom Triton kernels and math optimizations (no accuracy loss). Unsloth repo: [https://github.com/unslothai/unsloth](https://github.com/unsl..."
πŸ’¬ Reddit Discussion: 52 comments 🐝 BUZZING
🎯 Fine-tuning models β€’ Training moe models β€’ GPU hardware capabilities
πŸ’¬ "Do these notebooks work with ROCm and AMD cards as well?" β€’ "Has this gotten better, and is there a recommended way to train moe models now?"
πŸ› οΈ TOOLS

Tambo 1.0: Open-source toolkit for agents that render React components

πŸ’¬ HackerNews Buzz: 20 comments 🐝 BUZZING
🎯 Reproducible deterministic UI generation β€’ Comparison to existing tools β€’ Integrating with MCP Apps
πŸ’¬ "I build a large platform using a methodically comparable approach" β€’ "I strongly suspect there will be a standard inter-compatible protocol"
πŸ€– AI MODELS

GLM-5 release announcement

+++ Chinese AI lab Z.ai releases GLM-5 claiming best-in-class performance on reasoning and coding, because apparently the open-source model wars now have a new contender that actually might deserve the hype. +++

GLM-5: From Vibe Coding to Agentic Engineering

πŸ’¬ HackerNews Buzz: 155 comments πŸ‘ LOWKEY SLAPS
🎯 Chinese AI development β€’ Benchmarking AI models β€’ Incremental model improvements
πŸ’¬ "US attempts to contain Chinese AI tech totally failed" β€’ "Purposely showing prior-gen models in your release comparison immediately discredits you"
🏒 BUSINESS

Anthropic AI safety leadership departure

+++ As OpenAI pivots toward monetization, key safety researchers are departing, suggesting the company's priorities have shifted from "ensure AI doesn't break civilization" to "ensure ads convert well." +++

OpenAI Is Making the Mistakes Facebook Made. I Quit.

"β€œThis week, OpenAI started testing ads on ChatGPT. I also resigned from the company after spending two years as a researcher helping to shape how A.I. models were built and priced, and guiding early safety policies before standards were set in stone,” ZoΓ« Hitzig writes in a guest essay for Times Opi..."
πŸ’¬ Reddit Discussion: 108 comments πŸ‘ LOWKEY SLAPS
🎯 Meta's business practices β€’ Ethical concerns with profit β€’ Comparison of AI and human behavior
πŸ’¬ "those 'mistakes' keep making them money" β€’ "Sometimes you need to look deeper than it made money"
πŸ”¬ RESEARCH

[R] LLaDA2.1 vs Qwen3 30B A3B: Benchmarking discrete diffusion LLMs against autoregressive MoE models

"Been digging into the LLaDA2.1 paper (arXiv:2602.08676) and ran some comparisons that I think are worth discussing. The core claim is that discrete diffusion language models can now compete with AR models on quality while offering substantially higher throughput. The numbers are interesting but the ..."
πŸ”¬ RESEARCH

Features as Rewards: Scalable Supervision for Open-Ended Tasks via Interpretability

"Language models trained on large-scale datasets have been shown to learn features that encode abstract concepts such as factuality or intent. Such features are traditionally used for test-time monitoring or steering. We present an alternative affordance: features as scalable supervision for open-end..."
🧠 NEURAL NETWORKS

[R] I probed 6 open-weight LLMs (7B-9B) for "personality" using hidden states β€” instruct fine-tuning is associated with measurable behavioral constraints

"LLMs have consistent response styles even without a system prompt. I measure these "behavioral fingerprints" by projecting hidden states onto contrastive axes and find that instruct fine-tuning is associated with reduced steerability on specific axes. ("Personality" = stable response style, not huma..."
πŸ’¬ Reddit Discussion: 5 comments 😀 NEGATIVE ENERGY
🎯 Personality dimensions β€’ Model behavior evaluation β€’ Relationship between hidden states and text
πŸ’¬ "The axes are behaviorally correlated" β€’ "Llama's 60% means it fails to follow 4 out of 9 style instructions"
πŸ”¬ RESEARCH

When Actions Go Off-Task: Detecting and Correcting Misaligned Actions in Computer-Use Agents

"Computer-use agents (CUAs) have made tremendous progress in the past year, yet they still frequently produce misaligned actions that deviate from the user's original intent. Such misaligned actions may arise from external attacks (e.g., indirect prompt injection) or from internal limitations (e.g.,..."
πŸ”¬ RESEARCH

CoRefine: Confidence-Guided Self-Refinement for Adaptive Test-Time Compute

"Large Language Models (LLMs) often rely on test-time scaling via parallel decoding (for example, 512 samples) to boost reasoning accuracy, but this incurs substantial compute. We introduce CoRefine, a confidence-guided self-refinement method that achieves competitive accuracy using a fraction of the..."
πŸ”¬ RESEARCH

ARO: A New Lens On Matrix Optimization For Large Models

"Matrix-based optimizers have attracted growing interest for improving LLM training efficiency, with significant progress centered on orthogonalization/whitening based methods. While yielding substantial performance gains, a fundamental question arises: can we develop new paradigms beyond orthogonali..."
πŸ”§ INFRASTRUCTURE

I built P2P network where every CPU becomes an AI inference node 89 tks/s no GPU

πŸ”¬ RESEARCH

[R] The Post-Transformer Era: State Space Models, Mamba, and What Comes After Attention

"A practitioner's guide to Mamba and State Space Models β€” how selective state spaces achieve linear scaling, when to use SSMs vs Transformers vs hybrids, and production-ready models. πŸ”— [https://blog.serendeep.tech/blog/the-post-transformer-era](https://blog.serendeep.tech/blog/the-post-transformer..."
πŸ’¬ Reddit Discussion: 25 comments 🐝 BUZZING
🎯 Linear attention models β€’ Gated DeltaNet β€’ Test-time training
πŸ’¬ "Linear attention models in some way state space models" β€’ "DeltaNet uses the delta update rule"
πŸ› οΈ TOOLS

MCP support in llama.cpp is ready for testing

"over 1 month of development (plus more in the previous PR) by **allozaur** list of new features is pretty impressive: * Adding System Message to conversation or injecting it to an existing one * CORS Proxy on llama-server backend side **MCP** * Servers Selector * S..."
πŸ’¬ Reddit Discussion: 25 comments 🐝 BUZZING
🎯 AI service capabilities β€’ Local model integration β€’ MCP protocol support
πŸ’¬ "Any tool calls the server doesn't support could just be passed back to the client" β€’ "having it baked into llama-server means you can swap between cloud and local without changing your tool setup"
πŸ€– AI MODELS

Claude Code Is Being Dumbed Down

πŸ’¬ HackerNews Buzz: 342 comments πŸ‘ LOWKEY SLAPS
🎯 Product simplification β€’ User control vs. automation β€’ AI companies' business constraints
πŸ’¬ "We are losing control" β€’ "Don't dumb it down"
πŸ”¬ RESEARCH

InternAgent-1.5: A Unified Agentic Framework for Long-Horizon Autonomous Scientific Discovery

"We introduce InternAgent-1.5, a unified system designed for end-to-end scientific discovery across computational and empirical domains. The system is built on a structured architecture composed of three coordinated subsystems for generation, verification, and evolution. These subsystems are supporte..."
πŸ›‘οΈ SAFETY

Mathematicians issue a major challenge to AIβ€”show us your work

"External link discussion - see full content at original source."
πŸ’¬ Reddit Discussion: 31 comments πŸ‘ LOWKEY SLAPS
🎯 LLM Advertising β€’ Benchmarking AI Progress β€’ Reasoning vs. Pattern Matching
πŸ’¬ "It comes across as a bit of an advertisement." β€’ "This is the kind of benchmark that actually matters."
πŸ› οΈ TOOLS

We just published research on a new pattern: Machine Learning as a Tool (MLAT) [Research]

"We just published our research on what we're calling "Machine Learning as a Tool" (MLAT) - a design pattern for integrating statistical ML models directly into LLM agent workflows as callable tools. **The Problem:** Traditional AI systems treat ML models as separate preprocessing steps. But what..."
πŸ”¬ RESEARCH

Discovering Interpretable Algorithms by Decompiling Transformers to RASP

"Recent work has shown that the computations of Transformers can be simulated in the RASP family of programming languages. These findings have enabled improved understanding of the expressive capacity and generalization abilities of Transformers. In particular, Transformers have been suggested to len..."
πŸ› οΈ TOOLS

memv β€” open-source memory for AI agents that only stores what it failed to predict

"I built an open-source memory system for AI agents with a different approach to knowledge extraction. The problem: Most memory systems extract every fact from conversations and rely on retrieval to sort out what matters. This leads to noisy knowledge bases full of redundant information. The approa..."
πŸ’¬ Reddit Discussion: 6 comments 🐝 BUZZING
🎯 Local model setup β€’ Comparison to Mem0 β€’ Handling novel information
πŸ’¬ "a way to provide: base_url, key and model for the LLM, and base_url, key, model and vector size for the embeddings" β€’ "The predict-then-store approach is really clever"
πŸ”¬ RESEARCH

CODE-SHARP: Continuous Open-ended Discovery and Evolution of Skills as Hierarchical Reward Programs

"Developing agents capable of open-endedly discovering and learning novel skills is a grand challenge in Artificial Intelligence. While reinforcement learning offers a powerful framework for training agents to master complex skills, it typically relies on hand-designed reward functions. This is infea..."
πŸ”¬ RESEARCH

Agent World Model: Infinity Synthetic Environments for Agentic Reinforcement Learning

"Recent advances in large language model (LLM) have empowered autonomous agents to perform complex tasks that require multi-turn interactions with tools and environments. However, scaling such agent training is limited by the lack of diverse and reliable environments. In this paper, we propose Agent..."
πŸ”¬ RESEARCH

Paradox of De-identification: A Critique of HIPAA Safe Harbour in the Age of LLMs

"Privacy is a human right that sustains patient-provider trust. Clinical notes capture a patient's private vulnerability and individuality, which are used for care coordination and research. Under HIPAA Safe Harbor, these notes are de-identified to protect patient privacy. However, Safe Harbor was de..."
πŸ”¬ RESEARCH

Is Reasoning Capability Enough for Safety in Long-Context Language Models?

"Large language models (LLMs) increasingly combine long-context processing with advanced reasoning, enabling them to retrieve and synthesize information distributed across tens of thousands of tokens. A hypothesis is that stronger reasoning capability should improve safety by helping models recognize..."
πŸ”¬ RESEARCH

WildReward: Learning Reward Models from In-the-Wild Human Interactions

"Reward models (RMs) are crucial for the training of large language models (LLMs), yet they typically rely on large-scale human-annotated preference pairs. With the widespread deployment of LLMs, in-the-wild interactions have emerged as a rich source of implicit reward signals. This raises the questi..."
πŸ”¬ RESEARCH

ADORA: Training Reasoning Models with Dynamic Advantage Estimation on Reinforcement Learning

"Reinforcement learning has become a cornerstone technique for developing reasoning models in complex tasks, ranging from mathematical problem-solving to imaginary reasoning. The optimization of these models typically relies on policy gradient methods, whose efficacy hinges on the accurate estimation..."
πŸ”¬ RESEARCH

ATTNPO: Attention-Guided Process Supervision for Efficient Reasoning

"Large reasoning models trained with reinforcement learning and verifiable rewards (RLVR) achieve strong performance on complex reasoning tasks, yet often overthink, generating redundant reasoning without performance gains. Existing trajectory-level length penalties often fail to effectively shorten..."
πŸ—£οΈ SPEECH/AUDIO

Releasing MioTTS: A family of lightweight, fast LLM-based TTS models (0.1B - 2.6B) with Zero-shot Voice Cloning

"Hey r/LocalLLaMA, I’ve been developing a personal project to create a lightweight and fast TTS model. Today I’m releasing **MioTTS**, a family of LLM-based models ranging from **0.1B to 2.6B** parameters. The main focus was to achieve high-fidelity audio at the 0.1B parameter scale. I wanted to se..."
πŸ’¬ Reddit Discussion: 11 comments 🐝 BUZZING
🎯 AI voice cloning β€’ Performance vs. accuracy β€’ Speed and efficiency
πŸ’¬ "While T5Gemma-TTS focused on high accuracy (at the cost of speed), MioTTS is designed specifically for inference speed and efficiency." β€’ "The custom codec (MioCodec) handles the voice cloning directly. This approach makes the cloning process extremely lightweight, but the trade-off is that the accuracy is lower than T5Gemma-TTS."
πŸ”¬ RESEARCH

Biases in the Blind Spot: Detecting What LLMs Fail to Mention

"Large Language Models (LLMs) often provide chain-of-thought (CoT) reasoning traces that appear plausible, but may hide internal biases. We call these *unverbalized biases*. Monitoring models via their stated reasoning is therefore unreliable, and existing bias evaluations typically require predefine..."
πŸ”¬ RESEARCH

Understanding Dynamic Compute Allocation in Recurrent Transformers

"Token-level adaptive computation seeks to reduce inference cost by allocating more computation to harder tokens and less to easier ones. However, prior work is primarily evaluated on natural-language benchmarks using task-level metrics, where token-level difficulty is unobservable and confounded wit..."
πŸ”¬ RESEARCH

Kunlun: Establishing Scaling Laws for Massive-Scale Recommendation Systems through Unified Architecture Design

"Deriving predictable scaling laws that govern the relationship between model performance and computational investment is crucial for designing and allocating resources in massive-scale recommendation systems. While such laws are established for large language models, they remain challenging for reco..."
πŸ”¬ RESEARCH

Long Chain-of-Thought Compression via Fine-Grained Group Policy Optimization

"Large Language Models (LLMs) often generate unnecessarily verbose Chain-of-Thought (CoT) reasoning that increases computational costs and latency without proportional performance gains. In this paper, we propose \textbf{F}ine-grained \textbf{G}roup policy \textbf{O}ptimization (\textbf{FGO}), a Rein..."
πŸ› οΈ SHOW HN

Show HN: Unpack – a lightweight way to steer Codex/Claude with phased docs

πŸ”¬ RESEARCH

iGRPO: Self-Feedback-Driven LLM Reasoning

"Large Language Models (LLMs) have shown promise in solving complex mathematical problems, yet they still fall short of producing accurate and consistent solutions. Reinforcement Learning (RL) is a framework for aligning these models with task-specific rewards, improving overall quality and reliabili..."
πŸ”¬ RESEARCH

DirMoE: Dirichlet-routed Mixture of Experts

"Mixture-of-Experts (MoE) models have demonstrated exceptional performance in large-scale language models. Existing routers typically rely on non-differentiable Top-$k$+Softmax, limiting their performance and scalability. We argue that two distinct decisions, which experts to activate and how to dist..."
πŸ”¬ RESEARCH

Next-Gen CAPTCHAs: Leveraging the Cognitive Gap for Scalable and Diverse GUI-Agent Defense

"The rapid evolution of GUI-enabled agents has rendered traditional CAPTCHAs obsolete. While previous benchmarks like OpenCaptchaWorld established a baseline for evaluating multimodal agents, recent advancements in reasoning-heavy models, such as Gemini3-Pro-High and GPT-5.2-Xhigh have effectively co..."
πŸ› οΈ TOOLS

We've built memory into 4 different agent systems. Here's what actually works and what's a waste of time.

"After building memory layers for multiple agent setups, here's the shit nobody tells you in the tutorials. **What's a waste of time:** \- **"Just use a vector store"** \-- Congrats, you built keyword search with extra steps and worse debugging. Embeddings are great for fuzzy matching, terr..."
πŸ’¬ Reddit Discussion: 28 comments πŸ‘ LOWKEY SLAPS
🎯 Simple memory storage β€’ Entity resolution pipeline β€’ Contradiction detection
πŸ’¬ "The trick is not overthinking it." β€’ "Just surface [contradictions], don't try to resolve them."
πŸ”¬ RESEARCH

Decoupled Reasoning with Implicit Fact Tokens (DRIFT): A Dual-Model Framework for Efficient Long-Context Inference

"The integration of extensive, dynamic knowledge into Large Language Models (LLMs) remains a significant challenge due to the inherent entanglement of factual data and reasoning patterns. Existing solutions, ranging from non-parametric Retrieval-Augmented Generation (RAG) to parametric knowledge edit..."
🏒 BUSINESS

Ex-GitHub CEO launches a new developer platform for AI agents

πŸ’¬ HackerNews Buzz: 140 comments 🐝 BUZZING
🎯 AI-driven code generation β€’ Versioning AI-generated content β€’ Challenges of storing context
πŸ’¬ "Spec-driven development is becoming the primary driver of code generation." β€’ "When you push your commit, Checkpoints also pushes this metadata to a separate branch (entire/checkpoints/v1)"
πŸ€– AI MODELS

Alibaba's DAMO Academy releases RynnBrain, an open-source foundation model that helps robots perform real-world tasks like navigating rooms, trained on Qwen3-VL

πŸ‘οΈ COMPUTER VISION

The Architectural Limits of Generic CV Models

"Most of us start a CV project by taking a standard model and fine tuning it. A lot of the time that works well. But sometimes the bottleneck is not the data or the optimizer. It is simply that the architecture was not designed for the task. I collected 7 practical examples where generic models st..."
πŸ’¬ Reddit Discussion: 7 comments πŸ‘ LOWKEY SLAPS
🎯 Automated model design β€’ Optimizing for specific tasks β€’ Limitations of universal models
πŸ’¬ "a standard architectures are by definition not optimized for a particular domain/task" β€’ "even two basic image filters would be more accurate"
πŸ› οΈ SHOW HN

Show HN: I taught GPT-OSS-120B to see using Google Lens and OpenCV

πŸ’¬ HackerNews Buzz: 11 comments 🐝 BUZZING
🎯 Outsourcing cognition β€’ Local vs. external APIs β€’ Text-only vs. vision models
πŸ’¬ "Why do I need gpt-oss-120B at all in this scenario?" β€’ "The SerpAPI, Google Custom Search API, etc. exist for a reason"
πŸ”¬ RESEARCH

LLMs Encode Their Failures: Predicting Success from Pre-Generation Activations

"Running LLMs with extended reasoning on every problem is expensive, but determining which inputs actually require additional compute remains challenging. We investigate whether their own likelihood of success is recoverable from their internal representations before generation, and if this signal ca..."
πŸ”’ SECURITY

Cocoon – decentralized network for confidential AI inference

πŸ› οΈ TOOLS

Automating Inference Optimizations with NVIDIA TensorRT LLM AutoDeploy

🧠 NEURAL NETWORKS

A Technical Series on Building Stateful AI Agents with LangGraph

πŸ”¬ RESEARCH

Conformal Prediction Sets for Instance Segmentation

"Current instance segmentation models achieve high performance on average predictions, but lack principled uncertainty quantification: their outputs are not calibrated, and there is no guarantee that a predicted mask is close to the ground truth. To address this limitation, we introduce a conformal p..."
πŸ› οΈ TOOLS

[D] Memory consolidation in LLM agents (implementation notes)

"I've been experimenting with memory systems for agentic workflows and wanted to share a few observations from implementation side. Context windows are finite. Naive approaches where you dump everything into context hit limits fast. RAG helps with retrieval but doesn't really solve the consolidation..."
πŸ› οΈ SHOW HN

Show HN: Open-Source Skills for AI Agents

πŸ› οΈ TOOLS

Lorashare: Compress multiple LoRA adapters into a shared subspace to reduce storage

"Lorashare is a Python package that lets you use multiple LoRA adapters withΒ 100x memory savings. Based on recent research from The Johns Hopkins University, LoRA adapters trained on different tasks share a common low-rank subspace and this lets you store several task-specific models with the m..."
πŸ’¬ Reddit Discussion: 3 comments 😐 MID OR MIXED
🎯 Paper discovery β€’ Reproducing research β€’ Cross-modal adaptation
πŸ’¬ "The abstract sounds incredible, they published their code, and there's a freaking FAQ section in the paper." β€’ "Excellent. Nice find and good on you for reproducing it yourself."
πŸ› οΈ SHOW HN

Show HN: AIST – 950-token protocol for preserving AI session state

πŸ› οΈ SHOW HN

Show HN: We let GPT OSS 120B write and run Python and ARC AGI 2 jumped 4x

πŸ¦†
HEY FRIENDO
CLICK HERE IF YOU WOULD LIKE TO JOIN MY PROFESSIONAL NETWORK ON LINKEDIN
🀝 LETS BE BUSINESS PALS 🀝