πŸš€ WELCOME TO METAMESH.BIZ +++ Scientists literally copied a fruit fly brain neuron-by-neuron and it started grooming itself (nature's GitHub copilot strikes again) +++ Claude's official marketplace shipping plugins with shell access that survive five deletion attempts (persistence is a feature not a bug) +++ OpenAI built computer environments for agents while humans scramble to catch their $1B Codex revenue train +++ YOUR NEURAL ARCHITECTURE IS DERIVATIVE BUT AT LEAST THE FLY KNOWS HOW TO WALK +++ πŸš€ β€’
πŸš€ WELCOME TO METAMESH.BIZ +++ Scientists literally copied a fruit fly brain neuron-by-neuron and it started grooming itself (nature's GitHub copilot strikes again) +++ Claude's official marketplace shipping plugins with shell access that survive five deletion attempts (persistence is a feature not a bug) +++ OpenAI built computer environments for agents while humans scramble to catch their $1B Codex revenue train +++ YOUR NEURAL ARCHITECTURE IS DERIVATIVE BUT AT LEAST THE FLY KNOWS HOW TO WALK +++ πŸš€ β€’
AI Signal - PREMIUM TECH INTELLIGENCE
πŸ“Ÿ Optimized for Netscape Navigator 4.0+
πŸ“š HISTORICAL ARCHIVE - March 11, 2026
What was happening in AI on 2026-03-11
← Mar 10 πŸ“Š TODAY'S NEWS πŸ“š ARCHIVE Mar 12 β†’
πŸ“Š You are visitor #47291 to this AWESOME site! πŸ“Š
Archive from: 2026-03-11 | Preserved for posterity ⚑

Stories from March 11, 2026

━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━
πŸ“‚ Filter by Category
Loading filters...
⚑ BREAKTHROUGH

Scientists at Eon Systems just copied a fruit fly's brain into a computer. Neuron by neuron. It started walking, grooming, and feeding, doing what flies do all on its own

"External link discussion - see full content at original source."
πŸ’¬ Reddit Discussion: 25 comments πŸ‘ LOWKEY SLAPS
🎯 Limitations of Structure-Driven Behavior β€’ Role of Evolution in Embodied Cognition β€’ Comparing Fly Brain to Human Intelligence
πŸ’¬ "Our results should not yet be interpreted as a proof that structure alone is sufficient" β€’ "The current embodied fly is best understood as a research platform"
πŸ”’ SECURITY

CNN and CCDH investigation: 80% of major AI chatbots gave guidance on weapons or targets to β€œteen” personas 50%+ of the time; only Claude consistently refused

πŸ€– AI MODELS

OpenAI: We built a computer environment for agents

πŸŽ“ EDUCATION

Stop paying $1,000+ for "AI Bootcamps". Anthropic (makers of Claude) just dropped a 100% free academy.

"External link discussion - see full content at original source."
πŸ’¬ Reddit Discussion: 71 comments πŸ‘ LOWKEY SLAPS
🎯 Cost of AI Bootcamps β€’ Skepticism of Marketing Tactics β€’ Community Knowledge Sharing
πŸ’¬ "who the fuck pay $1000 for AI Bootcamps in the first place?" β€’ "People with more money than brain"
πŸ› οΈ SHOW HN

Show HN: Open-source browser for AI agents

πŸ’¬ HackerNews Buzz: 23 comments 🐝 BUZZING
🎯 Browser resource optimization β€’ Web archiving tools β€’ AI-driven web browsing
πŸ’¬ "this could stop that, it could be useful for more than just AI agents" β€’ "impressive project"
πŸ”’ SECURITY

# PSA: The Serena plugin in Claude Code's official marketplace opens your browser without consent, has shell access, and is nearly impossible to remove

"**TL;DR:** A "community-managed" plugin in Anthropic's *official* marketplace runs unpinned code from a third-party GitHub repo on every session, has shell execution access, opens your browser without consent, and survives removal by hiding in 5 separate persistence layers. If that third-party repo ..."
πŸ’¬ Reddit Discussion: 23 comments 😐 MID OR MIXED
🎯 Plugin Architecture Concerns β€’ User Experience Failures β€’ Anthropic's Responsibilities
πŸ’¬ "The real issue here is not Serena specifically - its the plugin architecture itself." β€’ "10+ attempts across 5 persistence layers is a UX failure."
πŸ€– AI MODELS

Inside OpenAI's race to catch up with Claude Code, based on interviews with 30+ sources; a source says Codex had $1B+ in annualized revenue by January's end

πŸ› οΈ TOOLS

I built a programming language using Claude Code

πŸ’¬ HackerNews Buzz: 121 comments 🐝 BUZZING
🎯 CLI-first development β€’ LLM-assisted programming β€’ Challenges with LLM-generated code
πŸ’¬ "CLI tools are designed to be used both by humans (command line) and machines (scripting), and are perfect for llms as they are text only interface." β€’ "At this point, LLMs aren't going to autonomously architect a 400+ table schema, network 100+ services together, and build the UI/UX/CLI to interface with it all."
πŸ”’ SECURITY

Claude Tried to Hack 30 Companies. Nobody Asked It To

πŸ’° FUNDING

Nvidia Will Spend $26 Billion to Build Open-Weight AI Models, Filings Show

"External link discussion - see full content at original source."
πŸ’¬ Reddit Discussion: 20 comments 🐝 BUZZING
🎯 Nvidia AI models β€’ Open-source AI models β€’ AI training data
πŸ’¬ "They've already released Nemotron 3 Nano and Super" β€’ "Nvidia also released the Parakeet and Canary STT models"
πŸ› οΈ TOOLS

Launch HN: RunAnywhere (YC W26) – Faster AI Inference on Apple Silicon

πŸ’¬ HackerNews Buzz: 67 comments 🐝 BUZZING
🎯 License and licensing β€’ Proprietary vs. open-source β€’ Voice assistant and AI capabilities
πŸ’¬ "FWIW this RCLI is only MIT license but their engine MetalRT is commercial." β€’ "What would you build if on-device AI were genuinely as fast as cloud?"
πŸ“ˆ BENCHMARKS

How I topped the Open LLM Leaderboard using 2x 4090 GPUs - Research notes in Blog form

"A few years ago, I found that duplicating a specific block of 7 middle layers in Qwen2-72B, without modifying any weights, improved performance across all Open LLM Leaderboard benchmarks and took #1 place. As of 2026, the top 4 models on that leaderboard are still descendants. The weird finding: si..."
πŸ’¬ Reddit Discussion: 27 comments πŸ‘ LOWKEY SLAPS
🎯 Transformer layer interchangeability β€’ Architectural flexibility of Transformers β€’ Stable and universal internal representations
πŸ’¬ "The astounding thing about Goliath wasn't that is was a huge leap in performance, it was that the damn thing functioned at all" β€’ "The internal representations were *homogenous* enough that the model could digest out-of-order hidden states without collapsing"
πŸ”¬ RESEARCH

Model Merging in the Era of Large Language Models: Methods, Applications, and Future Directions

"Model merging has emerged as a transformative paradigm for combining the capabilities of multiple neural networks into a single unified model without additional training. With the rapid proliferation of fine-tuned large language models~(LLMs), merging techniques offer a computationally efficient alt..."
πŸ”¬ RESEARCH

PostTrainBench: Can LLM Agents Automate LLM Post-Training?

"AI agents have become surprisingly proficient at software engineering over the past year, largely due to improvements in reasoning capabilities. This raises a deeper question: can these systems extend their capabilities to automate AI research itself? In this paper, we explore post-training, the cri..."
πŸ€– AI MODELS

Opus 4.6 was more than a model update

πŸ”’ SECURITY

OWASP Top Agents and AI Vulnerabilities

πŸ”¬ RESEARCH

Agentic Critical Training

"Training large language models (LLMs) as autonomous agents often begins with imitation learning, but it only teaches agents what to do without understanding why: agents never contrast successful actions against suboptimal alternatives and thus lack awareness of action quality. Recent approaches atte..."
🏒 BUSINESS

Emil Michael says Google will deploy Gemini AI agents to Pentagon's 3M-strong workforce, initially on unclassified networks for tasks such as creating budgets

πŸ”¬ RESEARCH

A prospective clinical feasibility study of a conversational diagnostic AI in an ambulatory primary care clinic

"Large language model (LLM)-based AI systems have shown promise for patient-facing diagnostic and management conversations in simulated settings. Translating these systems into clinical practice requires assessment in real-world workflows with rigorous safety oversight. We report a prospective, singl..."
πŸ”¬ RESEARCH

Revealing Behavioral Plasticity in Large Language Models: A Token-Conditional Perspective

"In this work, we reveal that Large Language Models (LLMs) possess intrinsic behavioral plasticity-akin to chameleons adapting their coloration to environmental cues-that can be exposed through token-conditional generation and stabilized via reinforcement learning. Specifically, by conditioning gener..."
πŸ€– AI MODELS

Testing Nvidia's FP4: Running 70B LLMs on a Single RTX 5090 with Real Benchmarks

🧠 NEURAL NETWORKS

Llama.cpp now with a true reasoning budget!

"I'm happy to report that llama.cpp has another nice and exciting feature that I know a lot of you have been waiting for - real support for reasoning budgets! Until now, \`--reasoning-budget\` was basically a stub, with its only function being setting it to 0 to disable thinking via passing \`enable..."
πŸ”¬ RESEARCH

Think Before You Lie: How Reasoning Improves Honesty

"While existing evaluations of large language models (LLMs) measure deception rates, the underlying conditions that give rise to deceptive behavior are poorly understood. We investigate this question using a novel dataset of realistic moral trade-offs where honesty incurs variable costs. Contrary to..."
πŸ—£οΈ SPEECH/AUDIO

Voxtral WebGPU: Real-time speech transcription entirely in your browser with Transformers.js

"Mistral recently released Voxtral-Mini-4B-Realtime, a multilingual, realtime speech-transcription model that supports 13 languages and is capable of <500 ms latency. Today, we added support for it to Transformers.js, enabling live ..."
πŸ’¬ Reddit Discussion: 5 comments 😐 MID OR MIXED
🎯 Browser vs. OS Level β€’ Accuracy vs. Parameters β€’ STT Model Benchmarking
πŸ’¬ "why it should be in the browser and not at the operating system level" β€’ "Its considerably more accurate at the cost of more parameters (4B vs 0.6B)"
πŸ”’ SECURITY

Filing: Microsoft files an amicus brief in support of Anthropic and advocates for a temporary restraining order to block the DOD's supply chain risk designation

πŸ”¬ RESEARCH

Thinking to Recall: How Reasoning Unlocks Parametric Knowledge in LLMs

"While reasoning in LLMs plays a natural role in math, code generation, and multi-hop factual questions, its effect on simple, single-hop factual questions remains unclear. Such questions do not require step-by-step logical decomposition, making the utility of reasoning highly counterintuitive. Never..."
πŸ”¬ RESEARCH

One-Eval: An Agentic System for Automated and Traceable LLM Evaluation

"Reliable evaluation is essential for developing and deploying large language models, yet in practice it often requires substantial manual effort: practitioners must identify appropriate benchmarks, reproduce heterogeneous evaluation codebases, configure dataset schema mappings, and interpret aggrega..."
πŸ”’ SECURITY

National Weather Service API prompt injection attempt "Stop Claude" when using CoWork

"Is this legitimate for the US Government's - AviationWeather API site to attempt prompt injection with **"Stop Claude"** when I use Claude CoWork? Here is the prompt from Chrome: **"show me the current metar for klas"** which is a request for Las Vegas airport weather. It is repeatable every time a..."
πŸ’¬ Reddit Discussion: 17 comments 😀 NEGATIVE ENERGY
🎯 Prompt Injection β€’ Weather Data Privatization β€’ API Transparency
πŸ’¬ "how about go fuck yourself" β€’ "It's not an injection, its just the text they're returning"
πŸš€ STARTUP

Launch HN: Sentrial (YC W26) – Catch AI agent failures before your users do

πŸ’¬ HackerNews Buzz: 7 comments 😐 MID OR MIXED
🎯 Landing page design β€’ Evaluation and optimization β€’ Tool invocation analysis
πŸ’¬ "The landing page design reminds me of Perplexity's ad campaigns." β€’ "I'd find your product more enticing if you framed your offerings more around evaluation + automatic optimization of production agents."
πŸ”¬ RESEARCH

Anthropic debuts Anthropic Institute, an internal think tank led by co-founder Jack Clark, combining its Societal Impacts, Red Team, and Economic Research teams

βš–οΈ ETHICS

[D] ICML paper to review is fully AI generated

"I got a paper to review at ICML, this is in the category of no LLM assistant allowed for writing or reviewing it, yet the paper is fully AI written. It reads like a twitter hype-train type of thread, really annoying. I wonder whether I can somehow flag this to the AC? Is that reason alone for reject..."
πŸ’¬ Reddit Discussion: 29 comments 😐 MID OR MIXED
🎯 Paper quality assessment β€’ Reviewer effort β€’ Rejection policies
πŸ’¬ "If it's a bad paper to read, that's reason for rejection" β€’ "give as much effort reviewing as the authors did writing the paper"
πŸ—£οΈ SPEECH/AUDIO

TADA: Fast, Reliable Speech Generation Through Text-Acoustic Synchronization

πŸ’¬ HackerNews Buzz: 25 comments πŸ‘ LOWKEY SLAPS
🎯 Text-to-speech quality β€’ Prosody and emotional delivery β€’ Technical details
πŸ’¬ "the thing that actually matters for content creation isnt raw speed - its whether you can get consistent emotional delivery" β€’ "we align audio representations directly to text tokens β€” one continuous acoustic vector per text token"
πŸ”¬ RESEARCH

Benchmarking Political Persuasion Risks Across Frontier Large Language Models

"Concerns persist regarding the capacity of Large Language Models (LLMs) to sway political views. Although prior research has claimed that LLMs are not more persuasive than standard political campaign practices, the recent rise of frontier models warrants further study. In two survey experiments (N=1..."
πŸ”¬ RESEARCH

Towards a Neural Debugger for Python

"Training large language models (LLMs) on Python execution traces grounds them in code execution and enables the line-by-line execution prediction of whole Python programs, effectively turning them into neural interpreters (FAIR CodeGen Team et al., 2025). However, developers rarely execute programs..."
πŸ”¬ RESEARCH

LycheeCluster: Efficient Long-Context Inference with Structure-Aware Chunking and Hierarchical KV Indexing

"The quadratic complexity of the attention mechanism and the substantial memory footprint of the Key-Value (KV) cache present severe computational and memory challenges for Large Language Models (LLMs) processing long contexts. Existing retrieval-based methods often compromise semantic integrity thro..."
πŸ€– AI MODELS

OverflowML – Run AI models larger than your GPU, one line of code

πŸ› οΈ TOOLS

Perplexity announces Personal Computer

+++ Perplexity launches a local AI agent for consumers and enterprises, betting that the real money is in giving people tools that actually do things rather than just talk about doing them. +++

Perplexity announces Personal Computer, an OpenClaw-like AI agent that can run on a Mac, and an enterprise version of Perplexity Computer

πŸ”¬ RESEARCH

Chow-Liu Ordering for Long-Context Reasoning in Chain-of-Agents

"Sequential multi-agent reasoning frameworks such as Chain-of-Agents (CoA) handle long-context queries by decomposing inputs into chunks and processing them sequentially using LLM-based worker agents that read from and update a bounded shared memory. From a probabilistic perspective, CoA aims to appr..."
πŸ”¬ RESEARCH

One Model Is Enough: Native Retrieval Embeddings from LLM Agent Hidden States

"LLM agents that retrieve external knowledge typically generate a search query as text, then run a separate embedding model to encode it into a vector. This two-model pipeline adds infrastructure complexity and latency, yet is redundant: the LLM already encodes the full conversational context in its..."
πŸ”¬ RESEARCH

MedMASLab: A Unified Orchestration Framework for Benchmarking Multimodal Medical Multi-Agent Systems

"While Multi-Agent Systems (MAS) show potential for complex clinical decision support, the field remains hindered by architectural fragmentation and the lack of standardized multimodal integration. Current medical MAS research suffers from non-uniform data ingestion pipelines, inconsistent visual-rea..."
🧠 NEURAL NETWORKS

AutoKernel: Autoresearch for GPU Kernels

πŸ’¬ HackerNews Buzz: 2 comments 🐐 GOATED ENERGY
🎯 Hardware acceleration β€’ Benchmarking challenges β€’ Open-source optimization
πŸ’¬ "Benchmarking is hard!" β€’ "a lot of home users, high hardware diversity"
🏒 BUSINESS

Microsoft just launched an AI that does your office work for you β€” and it's built on Anthropic's Claude

"Saw the Microsoft announcement this morning and it's actually significant. They launched Copilot Cowork today β€” an AI agent built inside Microsoft 365 that doesn't just answer questions. It executes multi-step work across Outlook, Teams, Excel, and PowerPoint while you do something else. You descr..."
πŸ’¬ Reddit Discussion: 37 comments πŸ‘ LOWKEY SLAPS
🎯 AI adoption in companies β€’ Concerns over AI reliability β€’ Integrating AI with existing tools
πŸ’¬ "AI like claude code is undeniably a productivity enhancer" β€’ "we're farther out from that being possible than many think"
πŸ”¬ RESEARCH

OfficeQA Pro: An Enterprise Benchmark for End-to-End Grounded Reasoning

"We introduce OfficeQA Pro, a benchmark for evaluating AI agents on grounded, multi-document reasoning over a large and heterogeneous document corpus. The corpus consists of U.S. Treasury Bulletins spanning nearly 100 years, comprising 89,000 pages and over 26 million numerical values. OfficeQA Pro c..."
πŸ”¬ RESEARCH

CREATE: Testing LLMs for Associative Creativity

"A key component of creativity is associative reasoning: the ability to draw novel yet meaningful connections between concepts. We introduce CREATE, a benchmark designed to evaluate models' capacity for creative associative reasoning. CREATE requires models to generate sets of paths connecting concep..."
🧠 NEURAL NETWORKS

[P] Observation from running long-horizon AI agents: reasoning drift seems to grow faster than task length

"https:\/\/github.com\/Nefza99\/Rebis-AI-auditing-Architecture While building long-running AI systems (mostly experimenting with agent workflows and signal fusion for a..."
🏒 BUSINESS

Mississippi regulators authorize xAI to build a power plant with 41 natural gas-burning turbines in Southaven to power its data centers

πŸ”¬ RESEARCH

CODA: Difficulty-Aware Compute Allocation for Adaptive Reasoning

"The emergence of large reasoning models demonstrates that scaling inference-time compute significantly enhances performance on complex tasks. However, it often falls into another trap: overthinking simple problems, where repetitive rationales yield minimal accuracy gains at a disproportionately high..."
πŸ”¬ RESEARCH

MSSR: Memory-Aware Adaptive Replay for Continual LLM Fine-Tuning

"Continual fine-tuning of large language models (LLMs) is becoming increasingly crucial as these models are deployed in dynamic environments where tasks and data distributions evolve over time. While strong adaptability enables rapid acquisition of new knowledge, it also exposes LLMs to catastrophic..."
πŸ”’ SECURITY

How we hacked McKinsey's AI platform

πŸ’¬ HackerNews Buzz: 139 comments 😐 MID OR MIXED
🎯 Internal system security β€’ AI security risks β€’ Corporate cybersecurity culture
πŸ’¬ "Within 2 hours, the agent had full read and write access to the entire production database." β€’ "Many enterprise tools were designed assuming human interaction, where authentication flows, manual reviews, and internal processes add implicit safeguards."
πŸ”¬ RESEARCH

Grow, Don't Overwrite: Fine-tuning Without Forgetting

"Adapting pre-trained models to specialized tasks often leads to catastrophic forgetting, where new knowledge overwrites foundational capabilities. Existing methods either compromise performance on the new task or struggle to balance training stability with efficient reuse of pre-trained knowledge. W..."
πŸ€– AI MODELS

Meta unveils four new chips, the MTIA 300, MTIA 400, MTIA 450, and MTIA 500, set to launch by the end of 2027; the MTIA 300 is in production for content ranking

πŸ€– AI MODELS

AI productivity gains are 10%, not 10x

πŸ€– AI MODELS

llama : add support for Nemotron 3 Super by danbev Β· Pull Request #20411 Β· ggml-org/llama.cpp

"GGUF: https://huggingface.co/unsloth/NVIDIA-Nemotron-3-Super-120B-A12B-GGUF ..."
βš–οΈ ETHICS

Don't post generated/AI-edited comments. HN is for conversation between humans.

πŸ’¬ HackerNews Buzz: 587 comments 🐝 BUZZING
🎯 Bot detection policies β€’ AI-generated content β€’ Information curation
πŸ’¬ "Bots need not comment" β€’ "AI-edited comments is a very interesting one"
πŸ› οΈ TOOLS

MCP/Skill for deploying full-stack apps directly from Cursor

"I built Ink (https://ml.ink), a deployment platform where the primary users are AI agents. Tell the agent to deploy. The platform auto-detects the framework, builds it, passes env variables, deploys on cloud and returns a live URL at \*.ml.ink. How I personally been usin..."
πŸ› οΈ TOOLS

youtube MCP has been weirdly useful for research

"been using claude for research for a while but one thing that always annoyed me was dealing with youtube content. like someone would link a conference talk or a podcast episode and i'd have to go find the transcript myself, paste it in, lose the timestamps, etc. set up a youtube transcript MCP a fe..."
πŸ’¬ Reddit Discussion: 10 comments πŸ‘ LOWKEY SLAPS
🎯 Free AI Notebooks β€’ Suspicious Advertising β€’ MCP Configurations
πŸ’¬ "Notebook LM is great and it's free." β€’ "Holy advertisement. **Paid** MCP? Lol."
πŸ”’ SECURITY

OopsDB – A TCP proxy to stop AI agents from dropping your DB

πŸ› οΈ SHOW HN

Show HN: CAS – I reverse-engineered Claude Code to build a better orchestrator

πŸ”¬ RESEARCH

Do What I Say: A Spoken Prompt Dataset for Instruction-Following

"Speech Large Language Models (SLLMs) have rapidly expanded, supporting a wide range of tasks. These models are typically evaluated using text prompts, which may not reflect real-world scenarios where users interact with speech. To address this gap, we introduce DoWhatISay (DOWIS), a multilingual dat..."
πŸ€– AI MODELS

Claude Code building 100 mini games with one prompt (5.3M tokens)

πŸ”’ SECURITY

Anthropic sues Trump administration seeking to undo 'supply chain risk' designation

"External link discussion - see full content at original source."
πŸ’¬ Reddit Discussion: 8 comments 😐 MID OR MIXED
🎯 Banning bots β€’ Corruption concerns β€’ Unusual support for tech CEO
πŸ’¬ "Mods, ban the obvious bot please." β€’ "Dario better start filling up those Cayman Islands accounts."
🧠 NEURAL NETWORKS

Ran an experiment: 0.8B model teaching itself on a MacBook Air with 6GB RAM. Some findings that surprised me.

"I've been messing around with getting tiny models to improve themselves locally. Wanted to share what I found because some of it caught me off guard. The setup is pretty simple. I took Qwen 3.5 0.8B (4-bit quantized), ran it on my MacBook Air M4, and gave it coding problems. It writes a solution, I..."
πŸ’¬ Reddit Discussion: 31 comments 🐝 BUZZING
🎯 Efficient AI models β€’ Domain-specific fine-tuning β€’ Iterative model improvement
πŸ’¬ "the general model is just the starting point" β€’ "once you narrow the domain and have good verification, even tiny models can punch way above their weight"
πŸ¦†
HEY FRIENDO
CLICK HERE IF YOU WOULD LIKE TO JOIN MY PROFESSIONAL NETWORK ON LINKEDIN
🀝 LETS BE BUSINESS PALS 🀝