πŸš€ WELCOME TO METAMESH.BIZ +++ Photonic AI chips heading to space with 860x less power draw (because neural nets work better when cosmic rays can't flip their bits) +++ Linux kernel getting AI code review that catches bugs humans miss (Sashiko finding your off-by-ones before Linus does) +++ Someone's letting AI design and run its own experiments on BOINC (distributed computing meets unsupervised science) +++ THE MESH EXPANDS WHERE SILICON FEARS TO TREAD +++ πŸš€ β€’
πŸš€ WELCOME TO METAMESH.BIZ +++ Photonic AI chips heading to space with 860x less power draw (because neural nets work better when cosmic rays can't flip their bits) +++ Linux kernel getting AI code review that catches bugs humans miss (Sashiko finding your off-by-ones before Linus does) +++ Someone's letting AI design and run its own experiments on BOINC (distributed computing meets unsupervised science) +++ THE MESH EXPANDS WHERE SILICON FEARS TO TREAD +++ πŸš€ β€’
AI Signal - PREMIUM TECH INTELLIGENCE
πŸ“Ÿ Optimized for Netscape Navigator 4.0+
πŸ“š HISTORICAL ARCHIVE - March 22, 2026
What was happening in AI on 2026-03-22
← Mar 21 πŸ“Š TODAY'S NEWS πŸ“š ARCHIVE
πŸ“Š You are visitor #47291 to this AWESOME site! πŸ“Š
Archive from: 2026-03-22 | Preserved for posterity ⚑

Stories from March 22, 2026

━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━
πŸ“‚ Filter by Category
Loading filters...
⚑ BREAKTHROUGH

Llama 8B matching 70B on multi-hop QA with structured prompting, no fine-tuning

"Ran a bunch of experiments with Graph RAG (KET-RAG) on multi hop question answering. Turns out **retrieval** is basically **solved**, the answer is in the context 77 to 91% of the time. The **bottleneck is reasoning**: 73 to 84% of wrong answers come from the model failing to connect the dots, not f..."
πŸ’¬ Reddit Discussion: 18 comments 🐝 BUZZING
🎯 Model Improvements β€’ Reasoning vs Retrieval β€’ Graph Compression
πŸ’¬ "the finding that 73-84% of failures are reasoning not retrieval is honestly the most important takeaway here" β€’ "the graph compression piece is interesting too - cutting context by 60% without extra llm calls probably helps more than ppl realize"
πŸ€– AI MODELS

Found 3 instructions in Anthropic's docs that dramatically reduce Claude's hallucination. Most people don't know they exist.

"Been building a daily research workflow on Claude. Kept gettingΒ confident-sounding outputs with zero sources. The kind of stuff that sounds right but you can't verify.Β  Β  Β  Β  Β  Β  Β  Β  Β  Β  Β  Β  Β  Β  Β  Β  Β  Β  Β  Β  Β  Β  Β  Β  Β  Β  Β I stumbled into Anthropic's "Reduce Hallucinations" documentation page byΒ accid..."
πŸ’¬ Reddit Discussion: 148 comments πŸ‘ LOWKEY SLAPS
🎯 Tradeoffs in AI Guardrails β€’ Customization and User Responsibility β€’ General Usefulness vs Specific Needs
πŸ’¬ "there's a tradeoff" β€’ "It's user responsability to be informed and to adjust it for their needs"
πŸ› οΈ TOOLS

Tinybox- offline AI device 120B parameters

πŸ’¬ HackerNews Buzz: 283 comments 🐝 BUZZING
🎯 Hardware pricing β€’ Hardware customization β€’ Sustainability and recycling
πŸ’¬ "the cheapest box seems pricey at 12 for a what is essentially a few gaming gpus" β€’ "Nobody is going to order a $10 million piece of infrastructure through your website's order form"
⚑ BREAKTHROUGH

I built a photonic AI chip for space with 860x less power, rad-hard to 106 krad

πŸ›‘οΈ SAFETY

What 33 AI Agents Taught Me About Alignment

πŸ› οΈ TOOLS

I built a daemon that polls Linear for issues and spawns Claude Code agents to implement them automatically

"I've been running a bash daemon that watches my Linear board for issues tagged "claude" and spawns autonomous Claude Code instances to implement them β€” in isolated git worktrees, with full transcripts, up to 5 concurrent workers. This applies equally well to Cursor CLI: Here's the workflow: ..."
πŸ’¬ Reddit Discussion: 3 comments 🐝 BUZZING
🎯 Worktree management β€’ Autonomous agents β€’ Merge conflict resolution
πŸ’¬ "Worktrees per agent is the right call" β€’ "The key insight you nailed is keeping tasks small and well-scoped"
πŸ”¬ RESEARCH

[P] I built an open-source benchmark to test if LLMs are actually as confident as they claim to be (Spoiler: They often aren't)

"Hey everyone, When building systems around modern open-source LLMs, one of the biggest issues is that they can confidently hallucinate or state an incorrect answer with a 95%+ probability. This makes it really hard to deploy them into the real world reliably if we don't understand their "overconfid..."
πŸ’¬ Reddit Discussion: 7 comments 🐐 GOATED ENERGY
🎯 Model confidence β€’ Calibration of confidence β€’ Benchmark evaluation
πŸ’¬ "Is confidence a score written out by the LLM?" β€’ "how confident are they on their confidence rating?"
πŸ”¬ RESEARCH

Nemotron-Cascade 2: Post-Training LLMs with Cascade RL and Multi-Domain On-Policy Distillation

"We introduce Nemotron-Cascade 2, an open 30B MoE model with 3B activated parameters that delivers best-in-class reasoning and strong agentic capabilities. Despite its compact size, its mathematical and coding reasoning performance approaches that of frontier open models. It is the second open-weight..."
πŸ› οΈ TOOLS

ik_llama.cpp gives 26x faster prompt processing on Qwen 3.5 27B β€” real world numbers

"I've been running Qwen 3.5 27B Q4_K_M on a Blackwell RTX PRO 4000 (24GB) for agentic coding work and hit a wall with mainline llama.cpp. Switched to the ik_llama.cpp fork today and the difference is staggering. Posting real numbers in case it helps others. Hardware Lenovo ThinkStation P520, Xeon W-..."
πŸ’¬ Reddit Discussion: 53 comments πŸ‘ LOWKEY SLAPS
🎯 GPU performance optimization β€’ Quantization challenges β€’ Model architecture differences
πŸ’¬ "your kV cache uses different quant which greatly slows down the speed" β€’ "The 26x is specifically the fused GDN kernel improvement for Qwen 3.5's hybrid SSM architecture"
πŸ›‘οΈ SAFETY

A circuit breaker for AI agents that fires before the wrong action executes

πŸ› οΈ TOOLS

Sashiko: AI code review system for the Linux kernel spots bugs humans miss

πŸ”¬ RESEARCH

Why Building Mega Clusters Is Wrong

πŸ› οΈ SHOW HN

Show HN: A BOINC project where AI designs and runs experiments autonomously

πŸ”¬ RESEARCH

Box Maze: A Process-Control Architecture for Reliable LLM Reasoning

"Large language models (LLMs) demonstrate strong generative capabilities but remain vulnerable to hallucination and unreliable reasoning under adversarial prompting. Existing safety approaches -- such as reinforcement learning from human feedback (RLHF) and output filtering -- primarily operate at th..."
πŸ”’ SECURITY

Ga. Court Order Included AI-Hallucinated Cases from Prosecutor's Proposed Order

πŸ› οΈ TOOLS

Impressive thread from /r/ChatGPT, where after ChatGPT finds out no 7Zip, tar, py7zr, apt-get, Internet, it just manually parsed and unzipped from hex data of the .7z file. What model + prompts would

"External link discussion - see full content at original source."
πŸ’¬ Reddit Discussion: 61 comments πŸ‘ LOWKEY SLAPS
🎯 Impressive AI capabilities β€’ Autonomous problem-solving β€’ Proprietary software challenges
πŸ’¬ "The fact that it just brute-forced a 7z format from raw hex without any tools is genuinely unhinged." β€’ "It was super spooky..it was just working in a loop and I started to see new trace + exceptions show up on the console of my training process while it figured out the path."
πŸ”¬ RESEARCH

Do VLMs Need Vision Transformers? Evaluating State Space Models as Vision Encoders

"Large vision--language models (VLMs) often use a frozen vision backbone, whose image features are mapped into a large language model through a lightweight connector. While transformer-based encoders are the standard visual backbone, we ask whether state space model (SSM) vision backbones can be a st..."
πŸ› οΈ TOOLS

MCP Is Costing You 37% More Tokens Than Necessary

"When we use skills, plugins or MCP tools, Claude reads long input schemas or injects prompt instructions. Those tokens are charged as input tokens, and can be expensive at scale, especially when it comes to API usage. We even ask Claude to explore other folders and sibling repositories, read files ..."
πŸ’¬ Reddit Discussion: 16 comments 🐝 BUZZING
🎯 MCP vs. CLI tools β€’ Token overhead β€’ Tool discovery
πŸ’¬ "Agents are naturally good at bash β€” reading files, writing files, piping commands, parsing output." β€’ "The one thing MCP does well is when it's tightly integrated (like Claude Code's built-in tools) β€” that feels natural because they control both sides."
πŸ› οΈ TOOLS

Nvidia Open-Sources OpenShell: Agent Runtime with Security Guardrails

πŸ’¬ HackerNews Buzz: 2 comments 🐐 GOATED ENERGY
🎯 AI agents as computing paradigm β€’ Systems-level changes for AI agents β€’ Nvidia's AI advancements
πŸ’¬ "What actually has to change at the systems level for agents to become a first-class workload?" β€’ "NVIDIA frames AI agents as the next computing paradigm"
πŸ”¬ RESEARCH

SAVeS: Steering Safety Judgments in Vision-Language Models via Semantic Cues

"Vision-language models (VLMs) are increasingly deployed in real-world and embodied settings where safety decisions depend on visual context. However, it remains unclear which visual evidence drives these judgments. We study whether multimodal safety behavior in VLMs can be steered by simple semantic..."
πŸ”¬ RESEARCH

OS-Themis: A Scalable Critic Framework for Generalist GUI Rewards

"Reinforcement Learning (RL) has the potential to improve the robustness of GUI agents in stochastic environments, yet training is highly sensitive to the quality of the reward function. Existing reward approaches struggle to achieve both scalability and performance. To address this, we propose OS-Th..."
πŸ”’ SECURITY

Claude Code workspace trust dialog bypass, settings loading order CVE-2026-33068

πŸ“Š DATA

some pretty dope datasets i came across from the 3D vision conference in vancouver

"harmony4d, the precursor to the contact4d dataset. it's a large-scale multi-view video dataset of in-the-wild close human–human contact interactions: https://huggingface.co/datasets/Voxel51/Harmony4D toon3d, has 12 scenes from popular hand-drawn cartoons and anime, each comprising 5–12 frames that ..."
πŸ› οΈ SHOW HN

Show HN: Vessel Browser – An open-source browser built for AI agents, not humans

⚑ BREAKTHROUGH

Inducing Sustained Creativity and Diversity in Large Language Models

πŸ”¬ RESEARCH

How Uncertainty Estimation Scales with Sampling in Reasoning Models

"Uncertainty estimation is critical for deploying reasoning language models, yet remains poorly understood under extended chain-of-thought reasoning. We study parallel sampling as a fully black-box approach using verbalized confidence and self-consistency. Across three reasoning models and 17 tasks s..."
πŸ› οΈ TOOLS

Hands-on with Gemini task automation on mobile: it's super impressive despite being very slow and failing at some tasks; it can order food, book Ubers, and more

πŸ› οΈ TOOLS

Litesearch: Karpathy's autoresearch but for consumer GPUs (4–8GB) + easy GUI

"Karpathy's autoresearch is awesome β€” agent edits train.py and runs tiny LLM experiments overnight. But it wants serious VRAM. I forked it to run on normal cards like my 1080/3060: * Auto-picks model size/depth/batch/seq len so it fits your VRAM (leaves buffer, no more OOM surpri..."
πŸ› οΈ SHOW HN

Show HN: ClawJetty: Agent Pages for Production AI

πŸ› οΈ SHOW HN

Show HN: GoldenMatch – Entity resolution with LLM scoring, 97% F1, no Spark

πŸ› οΈ TOOLS

the first native Pytorch distributed training backend for Apple Silicon

πŸ¦†
HEY FRIENDO
CLICK HERE IF YOU WOULD LIKE TO JOIN MY PROFESSIONAL NETWORK ON LINKEDIN
🀝 LETS BE BUSINESS PALS 🀝