πŸš€ WELCOME TO METAMESH.BIZ +++ Ex-OpenAI policy chief launches AVERI to audit frontier models (because internal safety teams worked out so well) +++ Someone trained a 30M parameter "Topological Transformer" from scratch like it's 2017 and compute is free +++ OpenCuff promises safe AI coding agents via capability-based execution while everyone else just YOLOs to prod +++ THE FUTURE IS AUDITED, SANDBOXED, AND STILL SOMEHOW RUNNING UNOPTIMIZED CUDA +++ πŸš€ β€’
πŸš€ WELCOME TO METAMESH.BIZ +++ Ex-OpenAI policy chief launches AVERI to audit frontier models (because internal safety teams worked out so well) +++ Someone trained a 30M parameter "Topological Transformer" from scratch like it's 2017 and compute is free +++ OpenCuff promises safe AI coding agents via capability-based execution while everyone else just YOLOs to prod +++ THE FUTURE IS AUDITED, SANDBOXED, AND STILL SOMEHOW RUNNING UNOPTIMIZED CUDA +++ πŸš€ β€’
AI Signal - PREMIUM TECH INTELLIGENCE
πŸ“Ÿ Optimized for Netscape Navigator 4.0+
πŸ“š HISTORICAL ARCHIVE - January 18, 2026
What was happening in AI on 2026-01-18
← Jan 17 πŸ“Š TODAY'S NEWS πŸ“š ARCHIVE Jan 19 β†’
πŸ“Š You are visitor #47291 to this AWESOME site! πŸ“Š
Archive from: 2026-01-18 | Preserved for posterity ⚑

Stories from January 18, 2026

━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━
πŸ“‚ Filter by Category
Loading filters...
πŸ”¬ RESEARCH

Representation-Aware Unlearning via Activation Signatures: From Suppression to Knowledge-Signature Erasure

"Selective knowledge erasure from LLMs is critical for GDPR compliance and model safety, yet current unlearning methods conflate behavioral suppression with true knowledge removal, allowing latent capabilities to persist beneath surface-level refusals. In this work, we address this challenge by intro..."
πŸ”¬ RESEARCH

A Safety Report on GPT-5.2, Gemini 3 Pro, Qwen3-VL, Doubao 1.8, Grok 4.1 Fast, Nano Banana Pro, and Seedream 4.5

"The rapid evolution of Large Language Models (LLMs) and Multimodal Large Language Models (MLLMs) has produced substantial gains in reasoning, perception, and generative capability across language and vision. However, whether these advances yield commensurate improvements in safety remains unclear, i..."
πŸ›‘οΈ SAFETY

OpenAI's former Head of Policy Research Miles Brundage announces AVERI, a nonprofit aimed at advocating the idea of external audits for frontier AI models

πŸ”¬ RESEARCH

Molmo2: Open Weights and Data for Vision-Language Models with Video Understanding and Grounding

"Today's strongest video-language models (VLMs) remain proprietary. The strongest open-weight models either rely on synthetic data from proprietary VLMs, effectively distilling from them, or do not disclose their training data or recipe. As a result, the open-source community lacks the foundations ne..."
πŸ”¬ RESEARCH

On the origin of neural scaling laws: from random graphs to natural language

"Scaling laws have played a major role in the modern AI revolution, providing practitioners predictive power over how the model performance will improve with increasing data, compute, and number of model parameters. This has spurred an intense interest in the origin of neural scaling laws, with a com..."
πŸ”¬ RESEARCH

How scientists are using Claude to accelerate research and discovery

πŸ’¬ HackerNews Buzz: 51 comments πŸ‘ LOWKEY SLAPS
🎯 Skepticism of LLM capabilities β€’ Conflict of interest concerns β€’ Anthropic marketing criticism
πŸ’¬ "Large language models are fundamentally not meant for tasks of this nature" β€’ "Confidence levels are suspect"
🧠 NEURAL NETWORKS

Starting from scratch: Training a 30M Topological Transformer

πŸ’¬ HackerNews Buzz: 26 comments πŸ‘ LOWKEY SLAPS
🎯 Efficient token encoding β€’ Alternative to attention β€’ Geometric data representation
πŸ’¬ "I am running an experiment of replacing discrete tokens with embeddings + small byte encoder/decoder" β€’ "If you want to prove a new alternative to attention without breaking the bank then one of the best ways to do that would probably be to retrain an already existing model"
πŸ”¬ RESEARCH

Be Your Own Red Teamer: Safety Alignment via Self-Play and Reflective Experience Replay

"Large Language Models (LLMs) have achieved remarkable capabilities but remain vulnerable to adversarial ``jailbreak'' attacks designed to bypass safety guardrails. Current safety alignment methods depend heavily on static external red teaming, utilizing fixed defense prompts or pre-collected adversa..."
πŸ›‘οΈ SAFETY

OpenCuff – Safe, capability-based execution for AI coding agents

πŸ€– AI MODELS

Doubling Inference Speed at Character.ai

πŸ€– AI MODELS

Geometric Transformer: Emergent Connectivity via High-Dimensional Manifold

πŸ› οΈ TOOLS

How to make LLMs and Agents work on large amounts of data

πŸ”¬ RESEARCH

Private LLM Inference on Consumer Blackwell GPUs

πŸ› οΈ SHOW HN

Show HN: Nvidia's CUDA libraries are generic and not optimized for LLM inference

πŸ”¬ RESEARCH

Contextual StereoSet: Stress-Testing Bias Alignment Robustness in Large Language Models

"A model that avoids stereotypes in a lab benchmark may not avoid them in deployment. We show that measured bias shifts dramatically when prompts mention different places, times, or audiences -- no adversarial prompting required. We introduce Contextual StereoSet, a benchmark that holds stereotype..."
πŸ”¬ RESEARCH

Generative AI collective behavior needs an interactionist paradigm

"In this article, we argue that understanding the collective behavior of agents based on large language models (LLMs) is an essential area of inquiry, with important implications in terms of risks and benefits, impacting us as a society at many levels. We claim that the distinctive nature of LLMs--na..."
πŸ› οΈ SHOW HN

Show HN: Lance – Open lakehouse format for multimodal AI datasets

πŸ”¬ RESEARCH

DR-Arena: an Automated Evaluation Framework for Deep Research Agents

"As Large Language Models (LLMs) increasingly operate as Deep Research (DR) Agents capable of autonomous investigation and information synthesis, reliable evaluation of their task performance has become a critical bottleneck. Current benchmarks predominantly rely on static datasets, which suffer from..."
πŸ› οΈ TOOLS

The Agentic Software Development Lifecycle

πŸ”¬ RESEARCH

MatchTIR: Fine-Grained Supervision for Tool-Integrated Reasoning via Bipartite Matching

"Tool-Integrated Reasoning (TIR) empowers large language models (LLMs) to tackle complex tasks by interleaving reasoning steps with external tool interactions. However, existing reinforcement learning methods typically rely on outcome- or trajectory-level rewards, assigning uniform advantages to all..."
πŸ”¬ RESEARCH

Are Your Reasoning Models Reasoning or Guessing? A Mechanistic Analysis of Hierarchical Reasoning Models

"Hierarchical reasoning model (HRM) achieves extraordinary performance on various reasoning tasks, significantly outperforming large language model-based reasoners. To understand the strengths and potential failure modes of HRM, we conduct a mechanistic study on its reasoning patterns and find three..."
πŸ€– AI MODELS

Running language models where they don't belong

"We have seen a cool counter-trend recently to the typical scaleup narrative (see Smol/Phi and ZIT most notably). I've been on a mission to push this to the limit (mainly for fun), moving LMs into environments where they have no business existing. My thesis is that even the most primitive environmen..."
πŸ”¬ RESEARCH

Grounding Agent Memory in Contextual Intent

"Deploying large language models in long-horizon, goal-oriented interactions remains challenging because similar entities and facts recur under different latent goals and constraints, causing memory systems to retrieve context-mismatched evidence. We propose STITCH (Structured Intent Tracking in Cont..."
πŸ”¬ RESEARCH

Defending Large Language Models Against Jailbreak Attacks via In-Decoding Safety-Awareness Probing

"Large language models (LLMs) have achieved impressive performance across natural language tasks and are increasingly deployed in real-world applications. Despite extensive safety alignment efforts, recent studies show that such alignment is often shallow and remains vulnerable to jailbreak attacks...."
πŸ› οΈ SHOW HN

Show HN: Task Orchestrator – Production Safety for Claude Code Agents

πŸ”¬ RESEARCH

LLM Pareto Frontier

πŸ› οΈ TOOLS

MCP Discovery API – Let AI agents find the right tools automatically

πŸ› οΈ TOOLS

From PSTN to Private Azure OpenAI: Shipping a Real-Time Voice AI Stack on AKS

πŸ”§ INFRASTRUCTURE

[D]It feels like LLM inference is missing its AWS Lambda moment.

"If we actually wanted β€œmodel = function” to work, a few things seem fundamentally required: β€’. Fast scale from zero without keeping GPUs alive just to hold state β€’ Execution state reuse so models don’t need full re-init and KV rebuild on every scale event β€’ Clear separation between orchestr..."
πŸ’¬ Reddit Discussion: 24 comments 🐝 BUZZING
🎯 Serverless model deployment β€’ Infrastructure challenges β€’ Orchestration and state management
πŸ’¬ "Lambda style LLM inference would be great" β€’ "CRIU style checkpointing seems to the path for most of it"
πŸ”¬ RESEARCH

Structure and Diversity Aware Context Bubble Construction for Enterprise Retrieval Augmented Systems

"Large language model (LLM) contexts are typically constructed using retrieval-augmented generation (RAG), which involves ranking and selecting the top-k passages. The approach causes fragmentation in information graphs in document structures, over-retrieval, and duplication of content alongside insu..."
πŸ¦†
HEY FRIENDO
CLICK HERE IF YOU WOULD LIKE TO JOIN MY PROFESSIONAL NETWORK ON LINKEDIN
🀝 LETS BE BUSINESS PALS 🀝