πŸš€ WELCOME TO METAMESH.BIZ +++ Anthropic kills third-party auth dreams while everyone was busy building wrappers (subscription abuse finally has consequences) +++ FlowPrefill fixes LLM serving by literally just letting requests take turns like civilized code +++ Kitten TTS runs speech synthesis in 14MB because apparently we were using 100x too much RAM this whole time +++ AI recommendation poisoning is the new SEO except it corrupts model memory instead of search results +++ THE FUTURE IS 0.9B MODELS READING YOUR DOCUMENTS ON A RASPBERRY PI +++ β€’
πŸš€ WELCOME TO METAMESH.BIZ +++ Anthropic kills third-party auth dreams while everyone was busy building wrappers (subscription abuse finally has consequences) +++ FlowPrefill fixes LLM serving by literally just letting requests take turns like civilized code +++ Kitten TTS runs speech synthesis in 14MB because apparently we were using 100x too much RAM this whole time +++ AI recommendation poisoning is the new SEO except it corrupts model memory instead of search results +++ THE FUTURE IS 0.9B MODELS READING YOUR DOCUMENTS ON A RASPBERRY PI +++ β€’
AI Signal - PREMIUM TECH INTELLIGENCE
πŸ“Ÿ Optimized for Netscape Navigator 4.0+
πŸ“Š You are visitor #53045 to this AWESOME site! πŸ“Š
Last updated: 2026-02-19 | Server uptime: 99.9% ⚑

Today's Stories

━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━
πŸ“‚ Filter by Category
Loading filters...
πŸ”¬ RESEARCH

The Geometry of Alignment Collapse: When Fine-Tuning Breaks Safety

"Fine-tuning aligned language models on benign tasks unpredictably degrades safety guardrails, even when training data contains no harmful content and developers have no adversarial intent. We show that the prevailing explanation, that fine-tuning updates should be orthogonal to safety-critical direc..."
🎯 PRODUCT

Anthropic officially bans using subscription auth for third party use

πŸ’¬ HackerNews Buzz: 381 comments 😐 MID OR MIXED
🎯 Clarity on OAuth usage β€’ Restrictions on API/SDK usage β€’ Transition from open to closed ecosystem
πŸ’¬ "If I build a commercial app that allows my users to connect using their OAuth token coming from their ChatGPT/Claude etc. account, do they allow me (and their users) to do this or not?" β€’ "Others (OpenAI, Copilot etc...) explicitly allow using OpenCode, they explicitly forbid it."
πŸ”¬ RESEARCH

FlowPrefill: Decoupling Preemption from Prefill Scheduling Granularity to Mitigate Head-of-Line Blocking in LLM Serving

"The growing demand for large language models (LLMs) requires serving systems to handle many concurrent requests with diverse service level objectives (SLOs). This exacerbates head-of-line (HoL) blocking during the compute-intensive prefill phase, where long-running requests monopolize resources and..."
πŸ”¬ RESEARCH

Policy Compiler for Secure Agentic Systems

"LLM-based agents are increasingly being deployed in contexts requiring complex authorization policies: customer service protocols, approval workflows, data access restrictions, and regulatory compliance. Embedding these policies in prompts provides no enforcement guarantees. We present PCAS, a Polic..."
πŸ€– AI MODELS

model: support GLM-OCR by ngxson Β· Pull Request #19677 Β· ggml-org/llama.cpp

"tl;dr **0.9B OCR model (you can run it on any potato)** # Introduction GLM-OCR is a multimodal OCR model for complex document understanding, built on the GLM-V encoder–decoder architecture. It introduces Multi-Token Prediction (MTP) loss and stable full-task reinforcement learning to improve tra..."
πŸ’¬ Reddit Discussion: 8 comments 🐝 BUZZING
🎯 OCR model issues β€’ OCR model capabilities β€’ OCR model comparison
πŸ’¬ "check_tensor_dims: tensor 'blk.0.attn_output.weight' has wrong shape" β€’ "0.9B OCR model that runs on any potato is exactly what i was hoping someone would build"
πŸ› οΈ TOOLS

Kitten TTS V0.8 is out: New SOTA Super-tiny TTS Model (Less than 25 MB)

"**Model introduction:** New Kitten models are out. Kitten ML has released open source code and weights for three new tiny expressive TTS models - 80M, 40M, 14M (all Apache 2.0) Discord: https://discord.com/invite/VJ86W4SURW GitHub: [https://github.com/Kitt..."
πŸ’¬ Reddit Discussion: 56 comments 🐝 BUZZING
🎯 Text-to-speech features β€’ Model improvements β€’ Community feedback
πŸ’¬ "A firefox/chrome extension would be #1 in like a week, I'm telling you." β€’ "thanks for the feedback. we'll have it by tomorrow."
πŸ”’ SECURITY

Manipulating AI memory for profit: The rise of AI Recommendation Poisoning

πŸ”¬ RESEARCH

GLM-5: from Vibe Coding to Agentic Engineering

"We present GLM-5, a next-generation foundation model designed to transition the paradigm of vibe coding to agentic engineering. Building upon the agentic, reasoning, and coding (ARC) capabilities of its predecessor, GLM-5 adopts DSA to significantly reduce training and inference costs while maintain..."
πŸ”¬ RESEARCH

Evaluating AI agents: Real-world lessons from building agentic systems at Amazon

πŸ”¬ RESEARCH

Towards a Science of AI Agent Reliability

"AI agents are increasingly deployed to execute important tasks. While rising accuracy scores on standard benchmarks suggest rapid progress, many agents still continue to fail in practice. This discrepancy highlights a fundamental limitation of current evaluations: compressing agent behavior into a s..."
πŸ”¬ RESEARCH

Causality is Key for Interpretability Claims to Generalise

"Interpretability research on large language models (LLMs) has yielded important insights into model behaviour, yet recurring pitfalls persist: findings that do not generalise, and causal interpretations that outrun the evidence. Our position is that causal inference specifies what constitutes a vali..."
πŸ› οΈ SHOW HN

Show HN: OpenCastor – A universal runtime connecting AI models to robot hardware

πŸ”¬ RESEARCH

Operationalising the Superficial Alignment Hypothesis via Task Complexity

"The superficial alignment hypothesis (SAH) posits that large language models learn most of their knowledge during pre-training, and that post-training merely surfaces this knowledge. The SAH, however, lacks a precise definition, which has led to (i) different and seemingly orthogonal arguments suppo..."
πŸ’Ό JOBS

How AI is affecting productivity and jobs in Europe

πŸ’¬ HackerNews Buzz: 66 comments 🐝 BUZZING
🎯 AI Productivity Challenges β€’ Coordinated Tech Industry Plans β€’ LLM vs Big Data Analytics
πŸ’¬ "The productivity gains for small and medium-sized enterprises are actually negative" β€’ "AI is failing to deliver because only 4% efficiency increase is a pre-mature conclusion"
πŸ’° FUNDING

World Labs $1B Funding

+++ World Labs raised a cool billion from a who's who of chip makers and enterprise software firms to build world models for robotics and science, because apparently simulating reality is now venture fundable. +++

Fei-Fei Li's World Labs raised $1B from A16Z, Nvidia to advance its world models

πŸ’¬ HackerNews Buzz: 15 comments πŸ‘ LOWKEY SLAPS
🎯 World models β€’ Video generation β€’ Scalability
πŸ’¬ "The interesting thing to me about their world models is that it's like a static point cloud model" β€’ "I see the video generation base as generally superior but far more expensive"
πŸ”’ SECURITY

Microsoft confirms a bug that let Microsoft 365 Copilot summarize confidential emails from Sent Items and Drafts folders, and deployed a fix in early February

πŸ”’ SECURITY

Security audit of OpenClaw and other similar open source AI Agents

πŸ’¬ HackerNews Buzz: 1 comments 🐐 GOATED ENERGY
🎯 Security audits β€’ AI agent frameworks β€’ Vulnerability analysis
πŸ’¬ "Full software composition analysis" β€’ "Vulnerability mapping across dependencies"
πŸ”¬ RESEARCH

Scaling Open Discrete Audio Foundation Models with Interleaved Semantic, Acoustic, and Text Tokens

"Current audio language models are predominantly text-first, either extending pre-trained text LLM backbones or relying on semantic-only audio tokens, limiting general audio modeling. This paper presents a systematic empirical study of native audio foundation models that apply next-token prediction t..."
πŸ”¬ RESEARCH

A Content-Based Framework for Cybersecurity Refusal Decisions in Large Language Models

"Large language models and LLM-based agents are increasingly used for cybersecurity tasks that are inherently dual-use. Existing approaches to refusal, spanning academic policy frameworks and commercially deployed systems, often rely on broad topic-based bans or offensive-focused taxonomies. As a res..."
πŸ”¬ RESEARCH

Align Once, Benefit Multilingually: Enforcing Multilingual Consistency for LLM Safety Alignment

"The widespread deployment of large language models (LLMs) across linguistic communities necessitates reliable multilingual safety alignment. However, recent efforts to extend alignment to other languages often require substantial resources, either through large-scale, high-quality supervision in the..."
πŸ”¬ RESEARCH

CrispEdit: Low-Curvature Projections for Scalable Non-Destructive LLM Editing

"A central challenge in large language model (LLM) editing is capability preservation: methods that successfully change targeted behavior can quietly game the editing proxy and corrupt general capabilities, producing degenerate behaviors reminiscent of proxy/reward hacking. We present CrispEdit, a sc..."
πŸ”¬ RESEARCH

OpenAI and Paradigm announce EVMbench, a benchmark that measures how well AI agents can detect, exploit, and patch high-severity smart contract vulnerabilities

πŸ”¬ RESEARCH

From Growing to Looping: A Unified View of Iterative Computation in LLMs

"Looping, reusing a block of layers across depth, and depth growing, training shallow-to-deep models by duplicating middle layers, have both been linked to stronger reasoning, but their relationship remains unclear. We provide a mechanistic unification: looped and depth-grown models exhibit convergen..."
πŸ”¬ RESEARCH

Measuring Mid-2025 LLM-Assistance on Novice Performance in Biology

"Large language models (LLMs) perform strongly on biological benchmarks, raising concerns that they may help novice actors acquire dual-use laboratory skills. Yet, whether this translates to improved human performance in the physical laboratory remains unclear. To address this, we conducted a pre-reg..."
πŸ› οΈ TOOLS

What tech stack Claude Code defaults to when building apps

πŸ”¬ RESEARCH

Recursive Concept Evolution for Compositional Reasoning in Large Language Models

"Large language models achieve strong performance on many complex reasoning tasks, yet their accuracy degrades sharply on benchmarks that require compositional reasoning, including ARC-AGI-2, GPQA, MATH, BBH, and HLE. Existing methods improve reasoning by expanding token-level search through chain-of..."
πŸ”¬ RESEARCH

This human study did not involve human subjects: Validating LLM simulations as behavioral evidence

"A growing literature uses large language models (LLMs) as synthetic participants to generate cost-effective and nearly instantaneous responses in social science experiments. However, there is limited guidance on when such simulations support valid inference about human behavior. We contrast two stra..."
πŸ”’ SECURITY

Microsoft guide to pirating Harry Potter for LLM training (2024) [removed]

πŸ’¬ HackerNews Buzz: 172 comments πŸ‘ LOWKEY SLAPS
🎯 Copyright infringement β€’ Microsoft's IP protection β€’ Fair use for educational purposes
πŸ’¬ "We've all collectively decided that copyright just doesn't matter anymore." β€’ "This is probably the most polite way I would describe this to most, UG."
πŸ€– AI MODELS

FlashLM v4: 4.3M ternary model trained on CPU in 2 hours β€” coherent stories from adds and subtracts only

"Back with v4. Some of you saw v3 β€” 13.6M params, ternary weights, trained on CPU, completely incoherent output. Went back to the drawing board and rebuilt everything from scratch. **What it is:** 4.3M parameter language model where every weight in the model body is -1, 0, or +1. Trained for 2 hour..."
πŸ’¬ Reddit Discussion: 38 comments 🐝 BUZZING
🎯 Ternary model architecture β€’ Efficient model inference β€’ Novel tokenizer design
πŸ’¬ "ternary weights mean inference is just adds and subtracts" β€’ "Every weight is 1.58 bits so a 192Γ—512 layer is \~19KB"
πŸ”¬ RESEARCH

Reinforced Fast Weights with Next-Sequence Prediction

"Fast weight architectures offer a promising alternative to attention-based transformers for long-context modeling by maintaining constant memory overhead regardless of context length. However, their potential is limited by the next-token prediction (NTP) training paradigm. NTP optimizes single-token..."
🎨 CREATIVE

Google rolls out Lyria 3, a generative music model that can make 30-second tracks with Nano Banana-made cover art, in beta in the Gemini app in eight languages

πŸ› οΈ TOOLS

Update from Anthropic regarding the Agent SDK.

"External link discussion - see full content at original source."
πŸ’¬ Reddit Discussion: 36 comments 😐 MID OR MIXED
🎯 Allowed vs. Prohibited Use β€’ SDK Usage Clarification β€’ Community Expectations
πŸ’¬ "they really should simply show a table showing allowed vs prohibited use" β€’ "We're clearly using it with claude code, it's just a glorified plugin"
πŸ› οΈ TOOLS

Major Claude Code policy clear up from Anthropic

"Source: https://code.claude.com/docs/en/legal-and-compliance#authentication-and-credential-use..."
πŸ’¬ Reddit Discussion: 103 comments 😐 MID OR MIXED
🎯 Unsustainable Pricing β€’ SDK Restrictions β€’ Competitor Opportunities
πŸ’¬ "Becoming exceedingly clear how much the current landscape is propped up with subsidized pricing" β€’ "They are going to find it difficult going forward. Chinese models will eat their lunch."
🧠 NEURAL NETWORKS

The next era of AI is not LLMs, it's Energy-Based Models EBMs

πŸ”’ SECURITY

Kernel-enforced sandbox App and SDK for AI agents, MCP and LLM workloads

πŸ› οΈ TOOLS

Sayou – Open-source Dropbox for AI agents

🎭 MULTIMODAL

Last week in Multimodal AI - Vision Edition

"I curate a weekly multimodal AI roundup, here are the vision-related highlights fromΒ last week: **Qwen3.5-397B-A17B - Native Vision-Language Foundation Model** * 397B-parameter MoE model with hybrid linear attention that integrates vision natively into the architecture. * Handles document parsing,..."
πŸ”’ SECURITY

Boundary Point Jail A new way to break the strongest AI defences

πŸ”¬ RESEARCH

Protecting the Undeleted in Machine Unlearning

"Machine unlearning aims to remove specific data points from a trained model, often striving to emulate "perfect retraining", i.e., producing the model that would have been obtained had the deleted data never been included. We demonstrate that this approach, and security definitions that enable it, c..."
πŸ”¬ RESEARCH

CMind: An AI Agent for Localizing C Memory Bugs

πŸ”’ SECURITY

The Problem with AI Agents Isn't Identity, It's Authorization

πŸ› οΈ SHOW HN

Show HN: Sieves, a unified interface for structured document AI

πŸ¦†
HEY FRIENDO
CLICK HERE IF YOU WOULD LIKE TO JOIN MY PROFESSIONAL NETWORK ON LINKEDIN
🀝 LETS BE BUSINESS PALS 🀝