πŸš€ WELCOME TO METAMESH.BIZ +++ TICKER ERROR: CONTENT TOO SPICY FOR ANTHROPIC'S USAGE POLICY +++ HERE'S WHAT'S HAPPENING +++ 'Western Qwen': IBM Wows with Granite 4 LLM Launch and Hybrid Mamba/Transformer +++ Sora 2: AI Video Generation with Realistic Sound +++ LoRA without regrets implemented in Hugging Face TRL [colab, and python scripts] πŸš€ β€’
πŸš€ WELCOME TO METAMESH.BIZ +++ TICKER ERROR: CONTENT TOO SPICY FOR ANTHROPIC'S USAGE POLICY +++ HERE'S WHAT'S HAPPENING +++ 'Western Qwen': IBM Wows with Granite 4 LLM Launch and Hybrid Mamba/Transformer +++ Sora 2: AI Video Generation with Realistic Sound +++ LoRA without regrets implemented in Hugging Face TRL [colab, and python scripts] πŸš€ β€’
AI Signal - PREMIUM TECH INTELLIGENCE
πŸ“Ÿ Optimized for Netscape Navigator 4.0+
πŸ“š HISTORICAL ARCHIVE - October 03, 2025
What was happening in AI on 2025-10-03
← Oct 02 πŸ“Š TODAY'S NEWS πŸ“š ARCHIVE Oct 04 β†’
πŸ“Š You are visitor #47291 to this AWESOME site! πŸ“Š
Archive from: 2025-10-03 | Preserved for posterity ⚑

Stories from October 03, 2025

━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━
πŸ“‚ Filter by Category
Loading filters...
πŸ’° FUNDING

OpenAI's H1 2025: $4.3B in income, $13.5B in loss

πŸ’¬ HackerNews Buzz: 535 comments πŸ‘ LOWKEY SLAPS
🎯 Monetization strategies β€’ Competition from Chinese models β€’ OpenAI's strategic dilemma
πŸ’¬ "That VC loss playbook only works if you can corner the market and squeeze later to make up for the losses." β€’ "The biggest concern IMO is how good the open weight models coming out of China are, on consumer hardware."
🏒 BUSINESS

Sources: in recent weeks, Meta changed FAIR's publishing rules to require extra review, angering staff; Yann LeCun considered resigning over Meta's AI changes

πŸ€– AI MODELS

Google says Gemini 2.5 Flash Image, aka Nano Banana, is now generally available and supports more aspect ratios, priced at $0.039/image and $30/1M output tokens

πŸ€– AI MODELS

IBM Granite 4.0 LLM launch

+++ Big Blue drops open source LLM family mixing Mamba with transformers, betting enterprises care more about memory efficiency than benchmark leaderboards. +++

'Western Qwen': IBM Wows with Granite 4 LLM Launch and Hybrid Mamba/Transformer

πŸ’¬ HackerNews Buzz: 13 comments πŸ‘ LOWKEY SLAPS
🎯 Model performance β€’ Hardware requirements β€’ IBM AI products
πŸ’¬ "Completely deserved" β€’ "No Mamba in the Ollama version"
πŸ› οΈ TOOLS

Granite 4.0 Micro (3.4B) running 100% locally in your browser w/ WebGPU acceleration

"External link discussion - see full content at original source."
πŸ’¬ Reddit Discussion: 30 comments 🐝 BUZZING
🎯 AI language models β€’ Efficient local deployment β€’ Hardware performance
πŸ’¬ "running smooth in browser is the future" β€’ "Imagine deploying LLM apps without any backend infra needed"
πŸ”„ OPEN SOURCE

We built this open-source LLM Inference project to boost context generation by up to 15x and now it is being implemented by NVIDIA Dynamo!

"Hi everyone, our team has been working nonstop on our open source project, LMCache, to reduce repetitive computation in LLM inference and make systems serve more people (3x more throughput in chat applications) and recently it has been implemented by NVIDIA's Inference project Dyanamo. In LLM servi..."
πŸ’¬ Reddit Discussion: 4 comments 🐝 BUZZING
🎯 Implementing Llama Integration β€’ Caching for Inference Costs β€’ Caching Benefits for Models
πŸ’¬ "How would we local llama-ers implement this?" β€’ "The reason they did not, from my best guess, is that for local workload, it has 1. less context reuse 2. usually runs smaller models which prefill very fast 3. the workload does not saturate the server(usually with lower qps)"
πŸ”§ INFRASTRUCTURE

Deep dive: Optimizing LLM inference for speed & efficiency β€” lessons learned from real-world experiments

"trungtranthanh.medium.com/the-art-of-llm-inference-fast-fit-and-free-c9faf1190d78..."
πŸ“Š DATA

Claude 4.5 Sonnet takes #1 in LMArena, the first Anthropic model since Sonnet 3.5 to be #1

"External link discussion - see full content at original source."
πŸ’¬ Reddit Discussion: 47 comments πŸ‘ LOWKEY SLAPS
🎯 AI model comparisons β€’ Benchmark limitations β€’ Subjective user experience
πŸ’¬ "Gemini is great. Just useful for specific kinds of things." β€’ "I don't care what the metrics say."
🎨 CREATIVE

Sora 2: AI Video Generation with Realistic Sound

πŸ”¬ RESEARCH

DeepScientist: Advancing Frontier-Pushing Scientific Findings Progressively

"While previous AI Scientist systems can generate novel findings, they often lack the focus to produce scientifically valuable contributions that address pressing human-defined challenges. We introduce DeepScientist, a system designed to overcome this by conducting goal-oriented, fully autonomous sci..."
πŸ”¬ RESEARCH

Fine-tuning Behavioral Cloning Policies with Preference-Based Reinforcement Learning

"Deploying reinforcement learning (RL) in robotics, industry, and health care is blocked by two obstacles: the difficulty of specifying accurate rewards and the risk of unsafe, data-hungry exploration. We address this by proposing a two-stage framework that first learns a safe initial policy from a r..."
πŸ’° FUNDING

Microsoft has committed $33B+ to neocloud providers; sources: its $19.4B Nebius deal will provide computing power for creating LLMs and a consumer AI assistant

πŸ€– AI MODELS

Huawei Develop New LLM Quantization Method (SINQ) that's 30x Faster than AWQ and Beats Calibrated Methods Without Needing Any Calibration Data

"Hugging Face model, dataset, or community resource."
πŸ’¬ Reddit Discussion: 37 comments 🐝 BUZZING
🎯 Quantization performance β€’ Inference speed β€’ Transparency of claims
πŸ’¬ "I'm interested on the de-quantization speed" β€’ "the speedup here is the speedup of quantization, and NOT inference"
βš–οΈ ETHICS

OpenAI asks a US judge to dismiss a lawsuit alleging it hired away xAI employees to steal trade secrets, calling the case part of Musk's β€œongoing harassment”

πŸ”„ OPEN SOURCE

LoRA without regrets implemented in Hugging Face TRL [colab, and python scripts]

"# LoRA Without Regret > [!WARNING] > I wrote this page for the TRL docs, but thought it's just drop it here in advance for anyone who can't wait. I also made a colab notebook of this guide. Recent res..."
πŸ’¬ Reddit Discussion: 4 comments 🐝 BUZZING
🎯 LoRA training β€’ LLM capabilities β€’ Practical applications
πŸ’¬ "For RL to be the next frontier of LLM training, it should be changing all parts of the system, not just tweak 0.0326% of model weights" β€’ "Choose a model that's Well suited, train multiple LoRAs, let a Backend decide which fine-tune to use and you quickly have experts at Hand for very little cost"
πŸ”¬ RESEARCH

[R] New paper: LLMs don't have privileged self knowledge, which means we can efficiently train a General Correctness Model to predict the correctness of multiple models. Surprising or expected?

"Quick paper highlight (adapted from TLDR thread): Finds no special advantage using an LLM to predict its own correctness (a trend in prior work), instead finding that LLMs benefit from learning to predict the correctness of many other models – becoming a GCM. \-- Training 1 GCM is strictly mor..."
πŸ› οΈ TOOLS

Google adds a new command-line interface and public API to its AI coding agent Jules, allowing it to plug into terminals, CI/CD systems, and tools like Slack

πŸ“Š DATA

Mercor launches the AI Productivity Index (APEX), which evaluates AI models' ability to perform β€œeconomically valuable knowledge work”; GPT-5 leads the index

🏒 BUSINESS

Anthropic CTO change - Rahul Patil hired

+++ Former Stripe CTO Rahul Patil takes the technical reins while cofounder McCandlish gets a shiny new "chief architect" title. Infrastructure era begins. +++

Anthropic hires former Stripe CTO Rahul Patil as its new CTO, taking over from co-founder Sam McCandlish, who will move to a new role as chief architect

🧠 NEURAL NETWORKS

Writing an LLM from scratch, part 20 – starting training, and cross entropy loss

πŸ”¬ RESEARCH

The One-Step Trap (In AI Research), by Richard Sutton

βš–οΈ ETHICS

"OpenAI Is Trying to Get Sued" – Nintendo IP Floods Sora 2 Video Generation App

πŸ”§ INFRASTRUCTURE

Microsoft CTO says he wants to swap most AMD and Nvidia GPUs for homemade chips

πŸ’¬ HackerNews Buzz: 120 comments 🐝 BUZZING
🎯 Custom silicon competition β€’ Vertical integration risks β€’ Analog ML hardware
πŸ’¬ "The software titan is rather late to the custom silicon party" β€’ "If everyone is siloed into their own vertically integrated hardware+operating system stack, the results will be awful for free software"
πŸ”’ SECURITY

LLM Code Review vs. Deterministic SAST Security Tools

πŸš€ STARTUP

AI chip startup Groq, last valued at $6.9B, says it plans to break ground on 12+ new data centers in 2026; Groq has set up 12 data centers in 2025 so far

πŸ”§ INFRASTRUCTURE

Simple LLM VRAM calculator for model inference

πŸ“Š DATA

AI Has Already Run Out of Training Data, Goldman's Data Chief Says

πŸ› οΈ TOOLS

A Jax-Native LLM Post-Training Library

πŸ”§ INFRASTRUCTURE

TechInsights: Huawei used components from TSMC, Samsung, and SK Hynix in some of its Ascend 910C chips; TSMC says the analyzed dies were made before Oct. 2024

πŸ’° FUNDING

OpenAI $500B valuation secondary share sale

+++ Sam Altman's ChatGPT factory edges past SpaceX in paper value after $6.6B secondary sale, proving conversational AI pays better than rockets. +++

OpenAI completed a secondary share sale that let staff sell ~$6.6B at a $500B valuation, making it the world's most valuable startup, ahead of SpaceX

"External link discussion - see full content at original source."
πŸ’¬ Reddit Discussion: 24 comments πŸ‘ LOWKEY SLAPS
🎯 Startup Definition β€’ Equity Valuation β€’ Profitability Concerns
πŸ’¬ "A real company shouldn't meet either of those criteria" β€’ "To have $1 billion after taxes you'd need to have about $2 billion worth of stock"
πŸ”¬ RESEARCH

Recursive Self-Aggregation Unlocks Deep Thinking in Large Language Models

"Test-time scaling methods improve the capabilities of large language models (LLMs) by increasing the amount of compute used during inference to make a prediction. Inference-time compute can be scaled in parallel by choosing among multiple independent solutions or sequentially through self-refinement..."
πŸ“Š DATA

Computer Use with Sonnet 4.5

"We ran one of our hardest computer-use benchmarks on Anthropic Sonnet 4.5, side-by-side with Sonnet 4. Ask: "Install LibreOffice and make a sales table". Sonnet 4.5: 214 turns, clean trajectory Sonnet 4: 316 turns, major detours The difference shows up in multi-step sequences where errors compou..."
πŸ”¬ RESEARCH

TimeRewarder: Learning Dense Reward from Passive Videos via Frame-wise Temporal Distance

"Designing dense rewards is crucial for reinforcement learning (RL), yet in robotics it often demands extensive manual effort and lacks scalability. One promising solution is to view task progress as a dense reward signal, as it quantifies the degree to which actions advance the system toward task co..."
πŸ”¬ RESEARCH

MENLO: From Preferences to Proficiency - Evaluating and Modeling Native-like Quality Across 47 Languages

"Ensuring native-like quality of large language model (LLM) responses across many languages is challenging. To address this, we introduce MENLO, a framework that operationalizes the evaluation of native-like response quality based on audience design-inspired mechanisms. Using MENLO, we create a datas..."
πŸ€– AI MODELS

Google's Jules enters as AI coding agent competition heats up

πŸ’° FUNDING

How much of the AI boom is underpinned by Nvidia's balance sheet? Investors ask

πŸ’° FUNDING

a16z releases a report, with Mercury data, on the top 50 AI companies startups pay for; OpenAI leads, followed by Anthropic, Replit, Freepik, and ElevenLabs

πŸ“Š DATA

Retrieval Embedding Benchmark

πŸ› οΈ TOOLS

Trackio: A Lightweight Experiment Tracking Library from Hugging Face

πŸ› οΈ TOOLS

Agentic AI Architecture for On-Call Engineers

πŸ› οΈ SHOW HN

Show HN: OpsWorker – AI SRE CoWorker that auto-investigates incidents

πŸ”¬ RESEARCH

What Makes 5% of AI Agents Work in Production?

πŸ”¬ RESEARCH

Stitch: Training-Free Position Control in Multimodal Diffusion Transformers

"Text-to-Image (T2I) generation models have advanced rapidly in recent years, but accurately capturing spatial relationships like "above" or "to the right of" poses a persistent challenge. Earlier methods improved spatial relationship following with external position control. However, as architecture..."
πŸ› οΈ TOOLS

RightNow AI, the first GPU code editor for CUDA

πŸ”¬ RESEARCH

Agent S3: Approaching Human-Level Computer Use with Wide Scaling

πŸ”¬ RESEARCH

Uncertainty Quantification for Regression using Proper Scoring Rules

"Quantifying uncertainty of machine learning model predictions is essential for reliable decision-making, especially in safety-critical applications. Recently, uncertainty quantification (UQ) theory has advanced significantly, building on a firm basis of learning with proper scoring rules. However, t..."
πŸ”¬ RESEARCH

Parametric Neural Amp Modeling with Active Learning

"We introduce Panama, an active learning framework to train parametric guitar amp models end-to-end using a combination of an LSTM model and a WaveNet-like architecture. With \model, one can create a virtual amp by recording samples that are determined through an ensemble-based active learning strate..."
πŸ”¬ RESEARCH

Towards Reliable Benchmarking: A Contamination Free, Controllable Evaluation Framework for Multi-step LLM Function Calling

"As language models gain access to external tools via structured function calls, they become increasingly more capable of solving complex, multi-step tasks. However, existing benchmarks for tool-augmented language models (TaLMs) provide insufficient control over factors such as the number of function..."
πŸ”¬ RESEARCH

Deconstructing Self-Bias in LLM-generated Translation Benchmarks

"As large language models (LLMs) begin to saturate existing benchmarks, automated benchmark creation using LLMs (LLM as a benchmark) has emerged as a scalable alternative to slow and costly human curation. While these generated test sets have to potential to cheaply rank models, we demonstrate a crit..."
πŸ”¬ RESEARCH

SPATA: Systematic Pattern Analysis for Detailed and Transparent Data Cards

"Due to the susceptibility of Artificial Intelligence (AI) to data perturbations and adversarial examples, it is crucial to perform a thorough robustness evaluation before any Machine Learning (ML) model is deployed. However, examining a model's decision boundaries and identifying potential vulnerabi..."
🧠 NEURAL NETWORKS

I Trained a Small Language Model from Scratch

πŸ’¬ HackerNews Buzz: 3 comments 😀 NEGATIVE ENERGY
🎯 Evaluation performance β€’ Lack of details β€’ Comparison to other LLMs
πŸ’¬ "How often are the answers nonsensical?" β€’ "Without those answerw, the article is meaningless."
πŸ”¬ RESEARCH

Self-supervised learning, JEPA, world models, and the future of AI [video]

πŸ’¬ HackerNews Buzz: 20 comments 🐝 BUZZING
🎯 LLM limitations β€’ Criticizing LeCun β€’ Questioning JEPA approach
πŸ’¬ "LeCun has correctly identified that LLM is only one type of intelligence" β€’ "This seems like the same exact talk LeCun has been giving for years"
πŸ”’ SECURITY

Unsexy AI Failures: The PDF That Broke ChatGPT

🏒 BUSINESS

Sources: delays in the deal to send Nvidia's AI chips to the UAE, announced in May, are frustrating Jensen Huang and administration officials like David Sacks

πŸ’° FUNDING

OpenAI now worth $500B, most valuable startup in history

πŸ’° FUNDING

Source: OpenAI employees sold shares to a consortium of investors including Thrive Capital, SoftBank, Dragoneer, Abu Dhabi's MGX, and T. Rowe Price

πŸ¦†
HEY FRIENDO
CLICK HERE IF YOU WOULD LIKE TO JOIN MY PROFESSIONAL NETWORK ON LINKEDIN
🀝 LETS BE BUSINESS PALS 🀝