This Week in Turing Post: | | Our news digest is always free. Upgrade to receive our deep dives in full, directly into your inbox. | |
|
| A day late, but for good reason β today we are reporting straight from NVIDIA GTC in DC |
|
| | It happened to be my second Jensen Huang keynote this year. The first one was during CES in January. And oh my, has the narrative been upgraded. It was peppered with praise for the current administration and the words βBuild in America,β βMade in America,β and finally βLetβs make America great again.β If President Donald Trump had been in the room β and apparently, according to Mr. Huang, he was supposed to be β he wouldβve been beaming like a shiny dime. It was an impressive exercise in tightening the knots with the current administration. If you can admire a skill, that was for sure a beautiful policy performance. NVIDIA spoke like a national champion, not a Silicon Valley company. The tone that used to orbit around innovation and scale now anchors itself in sovereignty, state power, and the energy race. The market reacted: +9.54%. | The business opportunities that friendship with Trump opens up for NVIDIA are massive. The reason Trump was not in DC today is because he was visiting Japan and Korea β and he really wanted Jensen Huang to join him. You might ask why. To negotiate a deal with China. Not long ago, Huang bitterly said that NVIDIAβs share in China fell from around 20 percent to zero β a brutal consequence of export controls that cut China off. | Why Trump is reversing his own actions? | Jensen Huang found the right approach. Being asked to bring manufacturing back in February, today (only 9 months later) he announced that Blackwell chips are now manufactured across Arizona, Indiana, Texas, and California β the first time NVIDIAβs flagship silicon is fully produced on US soil. He also highlighted the partnership with the Department of Energy (DoE) to build seven national AI supercomputers placed that hardware inside the countryβs scientific core. Even though Nokia is a Finnish company, Jensen introduced the deal with them on 6G as tying American chips back into the wireless stack. And the Foxconn factory in Houston β built as a digital twin in Omniverse β framed reindustrialization as a fusion of robotics and AI. | But of course just that wouldnβt work. Huang laid out the logic of selling chips and software to China in plain terms: | Platforms win by scale. More users make the system better. China has a huge user base that improves US AI platforms through volume and usage frequency. App ecosystems matter. Like phones need apps, AI hardware needs many models. Letting Chinese developers build and run models on US chips grows the CUDA/model ecosystem and strengthens the architecture globally. Set the global standard. If American tech is available in China, itβs more likely to become the worldwide default stack β benefiting both US industry and national security. Developers are the leverage. A large share of the worldβs AI researchers are in China; winning their loyalty keeps the American stack ahead.
| So is Trump mending the broken ties? Thatβs what the Secretary of DoE said during the Q&A: βPresident Trump is a master negotiator. Itβs his vision thatβs allowing America to unleash energy, lift government restrictions, and enable Jensen and me to have this dialogue today. Thatβs whatβs making this dream of AI leadership in science and defense possible.β He continued with diplomatic optimism: βChina is a scientific powerhouse. We have our differences, but also common ground. Iβm confident an agreement will come that benefits both nations β and the world.β | Huang understands this game better than most. Through the todayβs show it became clear: heβs positioning himself not as the CEO of a chipmaker, not even as a leader of accelerated computing and AI company but as an architect of national infrastructure. Every phrase about βAI factoriesβ and βsovereign computeβ is a signal to policymakers that NVIDIA is now part of Americaβs industrial DNA. | There is one more thing: energy and energy race. I asked Jensen Huang: βWe hear a lot about scaling and infrastructure, but not much about the energy behind it. How do solar, wind and other renewables feed into NVIDIAβs vision for AI factories, and how should government support that β given carbon concerns β and given that China may be ahead?β | That was his answer: | The U.S. must expand energy supply to win the AI race. Pro-energy, pro-growth policy is essential for any industrial buildout. Use every form of energy. AI is important enough that all sources should be deployed. We need all the energy to win. But the most important is the win on efficiency. Grace Blackwell delivers ~10Γ the performance of Hopper at ~50% more power, so system-level energy efficiency improves sharply. Strategy: pair more energy with better efficiency rather than relying on brute-force energy alone.
| He didnβt mention that, but his alignment with Trump is rooted exactly there β in energy. His administration is pushing nuclear power as the backbone for Americaβs βAI energy economy.β Heβs signing deals (like the new $80 billion Westinghouse-Japan partnership that he signed just a few hours ago) and easing approvals for new reactors. Nvidia doesnβt build nuclear plants, but its future depends on the kind of dense, continuous power only nuclear can provide. Trump wants to prove America can build again and is betting on nuclear as the backbone of the new AI economy. Huang needs that backbone to keep his AI factories running. For NVIDIA itβs, again, about efficiency, this time thermodynamic. | What a right place and right time to be NVIDIA. | | The NVIDIA GTC brought a range of very interesting and industry-shifting announcements β read about them below in the News from the Usual Suspects segment. |
|
| | π€ Recommended: NODES 2025 β Where Graphs Meet AI Innovation β streaming live Nov 6 | | Must see: Andrew Ng and Emil Eifrem open the day with a fireside chat on the future of AI. After that, choose from 140+ community sessions, packed with real-world use cases in GraphRAG, context engineering, and knowledge graphs. | Discover how developers everywhere are building the next wave of graph-powered AI agents. Secure your free spot today! | | | Attention Span on YouTube: It was interesting that the entire opening of Huangβs keynote focused on history β and was generated with AI. We didnβt know that while creating this video, but it turned out to be similar in spirit, also centered on remembering the past and the minds that shaped it. Honoring the Halloween week with the Festival of Memory β watch it in full and comment on it, please. Truly appreciated. |  | The Festival of Memory: When the Computers Were Women feat. Sora 2 |
|
| | We are also reading/watching: | | | | Curated Collections β AI Coding Repos | | | News from The Usual Suspects Β© | NVIDIA NVIDIA ARC and Partnership with Nokia for 6G A new product line, NVIDIA Arc Aerial RAN Computer, was unveiled β a fully software-defined, programmable computer that merges wireless communication and AI processing on the same platform. In a landmark move, Nokia will integrate ARC into its future base stations. Why it matters: βAI on RANβ β AI on the Radio Access Network β is a new kind of edge cloud, where industrial robots, drones, and cars can think locally and act instantly. Very impressive. The Rubin Platform and the Future of AI Supercomputing Cableless, liquid-cooled, and absurdly powerful β Rubin, NVIDIAβs third-generation, rack-scale supercomputer, has 100 petaflops in a single rack. Thatβs a 100x the power of the DGX-1 delivered to OpenAI just nine years ago. Why it matters: Rubin shows how NVIDIA plans to outgrow Mooreβs Law through βextreme co-design.β Huang loves this concept of βco-design.β The company now does scale up (the chips), scale out (the racks), and scale across (AI factories). This is achieved by building racks into a single giant GPU, connecting those racks with AI-optimized networking, and then linking entire data centers together. He calls it βthe virtual cycle of AIβ. With half a trillion dollars in cumulative orders already visible through 2026, Huangβs message was unmistakable β the AI industrial age has arrived. Omniverse DSX β Digital Twins for AI Factories Huang also introduced NVIDIA Omniverse DSX, a full-stack blueprint for designing and operating gigascale AI factories. Everything from cooling systems to compute clusters can be co-designed virtually before a single wall is built. Why it matters: Building an AI factory is no small feat β itβs energy, physics, and capital at planetary scale. With DSX, the entire process can be simulated and optimized beforehand. Itβs the digital twin of industrial civilization, reducing years of planning to months. NVIDIA Drive Hyperion and Partnership with Uber Finally, NVIDIA Drive Hyperion β the standardized architecture for βrobotaxi readyβ vehicles. Already adopted by Mercedes-Benz and Lucid, it now joins forces with Uber. Hyperion cars will plug directly into Uberβs global network. Why it matters: This is the missing bridge between self-driving prototypes and commercial deployment. By unifying sensors, compute, and infrastructure, NVIDIA is turning autonomous driving from a moonshot into a market. Incredible range of announcements.
OpenAI: Unchained Ambition OpenAI is now officially off the nonprofit leash. A new restructuring deal with Microsoft removes prior fundraising constraints, positioning the ChatGPT maker for a likely IPO. Microsoft retains a 27% stake, but forfeits exclusivity on compute. Altmanβs vision includes $1.4 trillion in data centers and gigawatt-scale ambitions. The nonprofit still holds the reins β at least on paper. Microsoftβs return? Nearly 10x on its $13.8B bet. Google: Entanglement Meets Advantage
|  | Google @Google |  |
| |
This week at Google, we announced a major breakthrough in quantum computing. Here are three things to know. | |  | | | 10:27 PM β’ Oct 24, 2025 | | | | | | 824 Likes 143 Retweets | 153 Replies |
|
| |  | TuringPost @TheTuringPost |  |
| |
Googleβs TPUs are finally having their breakout moment - a decade after the debut. Anthropic just signed a deal for up to 1 million TPUs - over a gigawatt of compute - making it one of the biggest AI infrastructure agreements to date. βͺοΈ Basically, TPUs accelerate neural | |  | | | 10:09 PM β’ Oct 26, 2025 | | | | | | 1.12K Likes 159 Retweets | 33 Replies |
|
| | Interesting Dataset: | | Are They Lovers or Friends? Evaluating LLMsβ Social Reasoning in English and Korean Dialogues β assess social reasoning across languages, showing weak cross-lingual and relational inference capabilities in current models βread the paper | | Models to pay attention to | Every attention matters and Every step evolves Researchers from Ling Team, Ant Group developed Ring-mini-linear-2.0 (16B parameters, 957M activations) and Ring-flash-linear-2.0 (104B parameters, 6.1B activations), using a hybrid architecture combining linear and softmax attention. These LLMs cut inference costs to 1/10 compared to 32B dense models and over 50% vs. earlier Ring models. Training efficiency improved 50% via the custom FP8 library linghe. Optimal attention ratios were identified, enabling SOTA performance in complex reasoning benchmarks βread the paper They also released Ring-1T, a 1-trillion parameter Mixture-of-Experts model activating 50B parameters per token. To tackle training-inference misalignment and RL inefficiencies, they introduced IcePop (token-level discrepancy masking), C3PO++ (dynamic rollout partitioning), and ASystem (a high-performance RL framework). Ring-1T achieves 93.4 on AIME-2025, 86.72 on HMMT-2025, 2088 on CodeForces, 55.94 on ARC-AGI-1, and silver medal level on IMO-2025, setting new open-source reasoning benchmarks βread the paper DeepAnalyze: Agentic large language models for autonomous data science Researchers from Tsinghua University introduced DeepAnalyze-8B, an 8B-parameter agentic LLM that autonomously generates analyst-grade research reports from raw data. It uses a curriculum-based training approach simulating human learning, and a data-grounded trajectory synthesis framework for high-quality training data. DeepAnalyze handles tasks from data Q&A to open-ended analysis. Despite its size, it outperforms workflow-based agents built on proprietary LLMs. All model components, including code and data, are open-sourced βread the paper Aion-1: Omnimodal foundation model for astronomical sciences Researchers from Flatiron Institute, Berkeley and other institutions introduced AION-1, a multimodal foundation model for astronomy integrating imaging, spectroscopic, and scalar data using modality-specific tokenization and transformer-based masked modeling. Pretrained on over 200 million observations from surveys like SDSS, DESI, and Gaia, AION-1 (300Mβ3.1B parameters) excels in tasks such as galaxy property estimation, morphology classification, and spectral super-resolution. It uses a single frozen encoder and is fully open-sourced, offering a scalable blueprint for scientific multimodal modeling βread the paper
| | The freshest research papers, categorized for your convenience | We organize research papers by goal-oriented or functional categories to make it easier to explore related developments and compare approaches. As always, papers we particularly recommend are marked with π | Reinforcement Learning, Distillation & Reasoning Dynamics | BAPO: Stabilizing Off-Policy Reinforcement Learning for LLMs via Balanced Policy Optimization with Adaptive Clipping β stabilize off-policy RL by adaptively balancing positive and negative gradients, preserving entropy and exploration for data-efficient, large-scale reasoning improvements βread the paper LoongRL: Reinforcement Learning for Advanced Reasoning over Long Contexts β train models to reason across long contexts by generating complex multi-hop tasks that teach retrieval, planning, and rechecking at scale βread the paper π QueST: Incentivizing LLMs to Generate Difficult Problems (by University of Zurich, Microsoft)) β generate high-difficulty coding problems through difficulty-aware sampling and rejection tuning to advance reasoning and distillation data quality βread the paper π Unified Reinforcement and Imitation Learning for Vision-Language Models (by NVIDIA, Kaist) β combine RL with adversarial imitation to help small VLMs mimic and improve upon teacher models across modalities βread the paper Search Self-play: Pushing the Frontier of Agent Capability without Supervision β achieve self-supervised RL for search-based agents by co-evolving task proposers and solvers in retrieval-augmented self-play βread the paper π On-policy Distillation (Thinking Machines) β train smaller models by sampling their own outputs and receiving token-level feedback from a teacher, achieving RL-level performance with 10β30Γ less compute βread the paper π Deep Self-Evolving Reasoning β model reasoning as a stochastic process where self-improvement probability compounds across iterations, enabling small models to surpass their teachers through long-horizon refinement βread the paper
| Efficiency, Compression & System Design | π Efficient Long-context Language Model Training by Core Attention Disaggregation β decouple attention into dedicated compute servers to balance workloads and eliminate stragglers, achieving up to 1.35Γ throughput on 512 GPUs βread the paper π Glyph: Scaling Context Windows via Visual-Text Compression (by Zhipu AI) β render text as images for visual-language processing, compressing long contexts 3β4Γ while maintaining reasoning accuracy and enabling million-token tasks βread the paper π LightMem: Lightweight and Efficient Memory-Augmented Generation β organize LLM memory into sensory, short-term, and long-term layers to preserve past context efficiently with over 10Γ runtime savings βread the paper AdaSPEC: Selective Knowledge Distillation for Efficient Speculative Decoders β accelerate speculative decoding by filtering difficult tokens during distillation, improving acceptance rates and generation efficiency βread the paper ARC-Encoder: Learning Compressed Text Representations for Large Language Models β train portable encoders that replace token embeddings with continuous compressed representations to reduce context length and inference cost βread the paper Sparser Block-Sparse Attention via Token Permutation β reorder tokens to maximize block sparsity and achieve up to 2.75Γ faster prefilling while maintaining accuracy close to full attention βread the paper
| Agents, Memory & Communication | DeepAgent: A General Reasoning Agent with Scalable Toolsets β unify reasoning, tool discovery, and action execution with RL-based tool credit assignment and dynamic memory folding for scalable, autonomous agents βread the paper Thought Communication in Multiagent Collaboration β propose βthought communication,β enabling agents to exchange latent states directly instead of language, revealing theoretical guarantees for shared and private thoughts βread the paper
| Alignment, Safety & Interpretability | π Extracting Alignment Data in Open Models (by Google DeepMind) β show that alignment datasets can be semantically extracted from post-trained models using embedding similarity, revealing risks and implications for distillation practices βread the paper Soft Instruction De-escalation Defense β mitigate prompt injections with iterative sanitization and rewriting loops for safer tool-augmented LLM agents βread the paper
| Theory, Representation & Foundational Insights | Language Models are Injective and Hence Invertible β prove that transformers are injective mappings, allowing exact reconstruction of inputs from activations, enabling theoretical and practical interpretability βread the paper RECALL: REpresentation-aligned Catastrophic-forgetting ALLeviation β merge models via hierarchical, representation-based fusion to prevent forgetting and align domain knowledge without old data βread the paper From Masks to Worlds: A Hitchhikerβs Guide to World Models β trace the evolution from masked models to interactive, memory-augmented architectures, charting the path toward persistent generative world models βread the paper
| Evaluation, Metrics & Cognitive Frameworks | Redefining Retrieval Evaluation in the Era of LLMs β propose utility- and distraction-aware metrics (UDCG) that correlate better with generation quality than classical IR measures βread the paper A Definition of AGI β define AGI quantitatively through ten cognitive domains based on psychometric theory, establishing measurable cognitive profiles and AGI scoring for modern models βread the paper Unfortunately this paper evoke some critique:
|  | Michael Saxon @m2saxon |  |
| |
The viral new "Definition of AGI" paper has fake citations which do not exist. And it specifically TELLS you to read them! Proof: different articles present at the specified journal/volume/page number, and their titles exist nowhere on any searchable repository. | |     | | | 12:50 AM β’ Oct 18, 2025 | | | | | | 1.66K Likes 216 Retweets | 103 Replies |
|
| Thatβs all for today. Thank you for reading! Please send this newsletter to colleagues if it can help them enhance their understanding of AI and stay ahead of the curve. | How did you like it? | | |
|