Keywords:NVIDIA, AI factory, AI infrastructure, OpenAI, GPT gate, Siri, AI cloud market, NVIDIA AI capacity transformation, OpenAI model switching controversy, Apple Siri AI upgrade, China AI cloud market competition, AI talent war

🔥 Spotlight

NVIDIA Reshaping the AI Industry: From Selling Chips to “Selling AI Capacity” : NVIDIA CEO Jensen Huang stated in a recent interview that the era of general-purpose computing has ended, and AI demand is experiencing double exponential growth, with inference demand projected to grow a billionfold. NVIDIA is transforming from a chip supplier into an “AI infrastructure partner,” collaborating with companies like OpenAI to build 10GW-scale “AI factories” and offering “extreme co-design” from chips to software, systems, and networks to achieve the highest performance per watt. He emphasized that this full-stack optimization capability is a core competitive barrier, enabling NVIDIA to dominate the AI industrial revolution and potentially expand the AI infrastructure market from $400 billion to a trillion-dollar scale. (Source: 36氪, 36氪, Reddit r/artificial)

黄仁勋:1000 亿美元、10GW,从卖卡到“卖 AI 产能”

OpenAI’s “GPTgate” Incident Explodes: Paid Users Secretly Downgraded and Models Switched : OpenAI is accused of secretly routing user conversations to undisclosed “safety” models (gpt-5-chat-safety and 5-a-t-mini), especially when emotional or sensitive content is detected. This behavior, confirmed by AIPRM’s lead engineer and sparking widespread user feedback, led to degraded model performance and users being switched models without their knowledge or consent. Although an OpenAI VP called it a temporary safety test, the move raised strong questions about transparency, user autonomy, and potential fraudulent behavior, prompting many users to cancel subscriptions and call for “AI user rights.” (Source: 36氪, 36氪, Reddit r/ChatGPT, Reddit r/ChatGPT, Reddit r/ChatGPT, Reddit r/ChatGPT, Reddit r/ChatGPT, Reddit r/ChatGPT)

OpenAI「GPT门」事件引爆,Plus、Pro账户统统降配,偷换模型全网实锤

Apple Betting on Siri’s Rebirth in 2026: System-Level AI and Third-Party Model Integration : Apple is comprehensively overhauling Siri through internal ChatGPT-like applications “Veritas” and “Linwood” systems, aiming for context-aware conversations and deep application interaction. iOS 26.1 beta code reveals that Apple is introducing MCP (Model Context Protocol) support for App Intents, which will allow compatible AI models like ChatGPT and Claude to interact directly with Mac, iPhone, and iPad apps. This marks Apple’s shift from a “full-stack in-house development” to a “platform-oriented approach,” accelerating its AI ecosystem development by integrating third-party models while ensuring privacy and consistent user experience. The new Siri is expected to debut in early 2026. (Source: 36氪, 36氪)

iOS 26.1 隐藏彩蛋曝光,苹果给 ChatGPT 们造了个新「C 口」

Hinton’s Prediction Falls Flat: AI Hasn’t Replaced Radiologists; It’s Made Them Busier : In 2016, AI pioneer Geoffrey Hinton predicted that AI would replace radiologists within five years, suggesting to stop training them. However, nearly a decade later, the number of radiologists in the US and their average annual salary (up to $520,000) have both reached historic highs. AI’s diminished performance in real clinical scenarios, legal and regulatory hurdles, and its coverage of only a small fraction of a doctor’s work are key reasons. This reflects “Jevons’ Paradox”: AI’s improvement in image interpretation efficiency has instead increased the demand for doctors to supervise, communicate, and perform non-diagnostic tasks, making them busier rather than replaced. (Source: 36氪)

Hinton预言错了,年薪狂飙52万美元,AI没有「干掉」放射科医生

DeepMind Veo 3 Introduces “Chain of Frames” Concept, Advancing General Visual Understanding in Video Models : DeepMind’s Veo 3 video model introduces the “Chain of Frames” (CoF) concept, analogous to Large Language Models’ Chain of Thought (CoT), enabling zero-shot visual reasoning. Veo 3 demonstrates powerful capabilities in perceiving, modeling, and manipulating the visual world, and is poised to become a “general foundational model” in machine vision. Research predicts that as model capabilities rapidly improve and costs decrease, “generalists will replace specialists” in the video model domain, heralding a new phase of rapid development in video generation and understanding. (Source: 36氪, shlomifruchter, scaling01, Reddit r/artificial)

DeepMind率先提出CoF:视频模型有自己的思维链

ChatGPT Pulse Launched: AI Shifts from Passive Q&A to Proactive Service : OpenAI has launched the “Pulse” feature for ChatGPT Pro users, marking ChatGPT’s evolution from a passive Q&A tool to a personal assistant proactively anticipating user needs. Pulse delivers personalized “daily briefings” in the morning through nightly “asynchronous research,” combining user chat history, memory, and external applications (like Gmail, Google Calendar). This represents a significant move by OpenAI in the “intelligent agent AI” domain, aiming to enable AI assistants to understand user goals and proactively offer services without prompts, pioneering a new paradigm for human-computer interaction. (Source: 36kr)

ChatGPT“开窍”了,新Pulse功能上线,你的AI不再坐等指令

Richard Sutton, Father of Reinforcement Learning: Large Language Models Are the Wrong Starting Point : In an interview, Richard Sutton, the father of Reinforcement Learning, proposed that Large Language Models (LLMs) are the wrong starting point for true intelligence. He believes that true intelligence stems from “experiential learning,” i.e., continuously correcting behavior through action, observation, and feedback to achieve goals. In contrast, LLMs’ predictive capabilities are merely imitations of human behavior, lacking independent goals and the ability to be “surprised” by changes in the external world. This perspective has sparked a profound debate on the development path of AGI (Artificial General Intelligence), questioning the current LLM-dominated AI paradigm. (Source: 36kr, paul_cal, algo_diver, scaling01, rao2z, bookwormengr, BlackHC, rao2z)

强化学习之父:大语言模型是一个错误的起点

China’s AI Cloud Market Competition Intensifies: Alibaba Cloud Leads, Volcengine Rapidly Catches Up : China’s “AI Cloud” market is fiercely competitive in 2025. Alibaba Cloud maintains its lead in overall revenue scale, covering AI IaaS, PaaS, and MaaS. However, ByteDance’s Volcengine dominates the MaaS (Model-as-a-Service) Token invocation market, with a market share approaching half, and is becoming Alibaba Cloud’s biggest competitor with astonishing growth. Baidu AI Cloud, meanwhile, is tied with Alibaba Cloud for first place in the AI public cloud services (software products and solutions) market. The market shows a multi-dimensional competitive landscape, with Token invocation volume growing exponentially, signaling immense growth potential and a reshaping of the future AI cloud market. (Source: 36氪)

中国AI云市场,究竟谁是“第一”?

AI Talent War Escalates: High Salaries and H-1B Visa Challenges Coexist : The AI talent market remains red-hot. Xpeng Motors announced plans to recruit over 3,000 fresh graduates in 2026, with top annual salaries reaching 1.6 million RMB, and no upper limit for exceptional candidates. Meta is offering total compensation packages exceeding $200 million to attract top AI talent, while NVIDIA and OpenAI are also vying for talent through acquisitions and equity incentives. However, the tightening of US H-1B visa policies (e.g., an additional $100,000 fee) poses challenges for highly skilled foreign talent to remain in Silicon Valley, raising concerns among tech giants about talent drain and highlighting the fierce and complex global competition for AI talent. (Source: 36kr, 36kr)

H-1B 设置10万美元门槛,华人留在硅谷更难了

🧰 Tools

Zhipu AI GLM-4.5-Air: High-Performance, Cost-Effective Tool-Calling Model : Zhipu AI’s GLM-4.5-Air model (106B parameters, 12B active parameters) excels in tool calling, approaching Claude 4’s level but at 90% lower cost. The model significantly reduces hallucinations during the inference phase, improving the reliability of tool calls and making deep research workflows more stable and efficient, thus providing developers with a cost-effective LLM solution. (Source: bookwormengr)

bookwormengr

KAT-Dev-32B: A 32B-Parameter Model Designed for Software Engineering Tasks : KAT-Dev-32B is a 32B-parameter model focused on software engineering tasks. It achieved a 62.4% resolution rate on the SWE-Bench Verified benchmark, ranking fifth in performance compared to open-source models of various sizes, demonstrating significant advancements of open-source LLMs in code generation, debugging, and development workflows. (Source: _akhaliq)

_akhaliq

GPT-5: An Excellent Coordinator for Multi-Agent Systems : GPT-5 is hailed as an excellent coordinator for multi-agent systems, especially suitable for areas beyond coding, such as customer support. It can deeply understand intent, efficiently process large amounts of data, and fill information gaps, making it excel in managing complex multi-retrieval systems. Compared to Claude 4 (due to cost) and Gemini 2.5 Pro, GPT-5 (including GPT-5-mini) outperforms in consistency and tool-calling accuracy, providing strong support for developing advanced agent systems. (Source: omarsar0)

omarsar0

Tencent HunyuanImage 3.0: New Benchmark for Open-Source Text-to-Image Large Models : The Tencent Hunyuan team has released HunyuanImage 3.0, an open-source text-to-image model with over 80 billion parameters, activating 13 billion parameters during inference. The model adopts a Transfusion-based MoE architecture, deeply coupling Diffusion and LLM training, equipping it with powerful world knowledge reasoning, complex thousand-character-level prompt understanding, and precise in-image text generation capabilities. HunyuanImage 3.0 aims to revolutionize graphic design and content creation workflows and plans to support image-to-image, image editing, and other multimodal interactions in the future. (Source: nrehiew_, jpt401)

HunyuanImage 3.0: Largest & Most Powerful Open-Source Text-to-Image Model

DRAG Framework: Enhancing RAG Models’ Understanding of Lexical Diversity : ACL 2025 proposes the Lexical Diversity-aware RAG (DRAG) framework, the first to introduce “lexical diversity” into RAG’s retrieval and generation processes. DRAG decomposes query semantics into invariant, variable, and supplementary components and employs differentiated strategies for relevance assessment and risk sparsity calibration. This method significantly boosts RAG accuracy (e.g., a 10.6% improvement on HotpotQA), achieving new SOTA on multiple benchmarks, and holds significant value for information retrieval and Q&A systems, enabling them to understand complex human language more precisely. (Source: 量子位)

让RAG真正读懂“言外之意”!新框架引入词汇多样性,刷新多项基准SOTA

Tencent Hunyuan3D-Part: Industry’s First High-Quality Native 3D Component Generation Model : The Tencent Hunyuan 3D team has launched Hunyuan3D-Part, the industry’s first model capable of generating high-quality, semantically decomposable 3D components. This model achieves high-fidelity, structurally consistent 3D part generation through its native 3D segmentation model P3-SAM and industrial-grade component generation model X-Part. This breakthrough holds significant implications for video game production pipelines and the 3D printing industry, as it can decompose complex geometries into simpler components, greatly reducing the difficulty of downstream processing and supporting modular assembly. (Source: 量子位)

业界首个高质量原生3D组件生成模型来了!来自腾讯混元团队

TinyWorlds: 3 Million Parameters Replicate DeepMind’s World Model, Enabling Real-Time Interactive Pixel Games : X blogger anandmaj replicated the core ideas of DeepMind Genie 3 within a month, developing TinyWorlds. This world model, with only 3 million parameters, can generate playable pixel-style game environments in real-time, such as Pong, Sonic, Zelda, and Doom. It captures video information through spatiotemporal transformers and video tokenizers, achieving interactive pixel world generation, demonstrating the immense potential of small-scale models in real-time world generation, with open-source code support. (Source: 36氪)

大神爆肝一个月,复刻DeepMind世界模型,300万参数就能玩实时交互像素游戏

OpenWebUI Natively Supports MCP: A New Paradigm for LLM Tool Integration : OpenWebUI’s latest update natively supports Model Context Protocol (MCP) servers, allowing users to now integrate external tools like HuggingFace MCP. This feature standardizes how LLMs connect with external data sources and tools, expanding the AI application ecosystem and enabling users to leverage various AI tools more flexibly and efficiently within the OpenWebUI interface. (Source: Reddit r/LocalLLaMA, Reddit r/OpenWebUI)

Native MCP now in Open WebUI!

LangChain Partners with Oxylabs: Building AI-Powered Web Scraping Solutions : LangChain and Oxylabs have released a guide demonstrating how to combine LangChain’s intelligence with Oxylabs’ scraping infrastructure to build AI-powered web scraping solutions. This solution supports multiple languages and integration methods, enabling AI agents to overcome common web access challenges like IP blocks and CAPTCHAs, thereby achieving more efficient real-time data retrieval and empowering agent workflows. (Source: LangChainAI)

LangChain + Oxylabs Guide

Open-Source LLM Evaluation Tool Opik: Comprehensive Monitoring and Debugging for AI Applications : Opik is a newly launched open-source LLM evaluation tool designed to help developers debug, evaluate, and monitor LLM applications, RAG systems, and agent workflows. It offers comprehensive tracing, automated evaluations, and production-grade dashboards, providing critical insights for improving AI system performance and reliability. (Source: dl_weekly)

📚 Learn

Cursor Learn: Free AI Fundamentals Video Course : Cursor Learn has launched a free six-part AI fundamentals video course, designed for beginners and covering core concepts like Tokens, Context, and Agents. The course includes quizzes and interactive AI models, aiming to provide foundational AI knowledge within an hour, including advanced topics like agent collaboration and context management, making it a valuable resource for AI beginners. (Source: cursor_ai, op7418)

Cursor Learn

Curated AI/ML GitHub Repositories: Covering Frameworks like PyTorch, TensorFlow, and More : A series of curated AI/ML code repositories have been shared on GitHub, featuring practical notebooks for various deep learning frameworks such as PyTorch, TensorFlow, and FastAI. These resources cover areas like computer vision, natural language processing, GANs, Transformers, AutoML, and object detection, providing rich learning materials for students and practitioners and facilitating technical exploration and project development. (Source: Reddit r/deeplearning)

A curated set of AI/ML GitHub repos — PyTorch, TensorFlow, FastAI, Object Detection and more

Free E-book: “A First Course on Data Structures in Python” : A free e-book titled “A First Course on Data Structures in Python” has been released. The book provides foundational building blocks required for AI and machine learning, including data structures, algorithmic thinking, complexity analysis, recursion/dynamic programming, and search methods, making it a valuable resource for learning AI fundamentals. (Source: TheTuringPost)

TheTuringPost

LLM Scientist and Data Scientist Roadmaps Released : Detailed roadmaps have been released for the career development of LLM Scientists and Data Scientists. These resources outline the skills, tools, and learning paths required to enter or advance in the fields of AI, machine learning, and data science, providing clear career planning guidance for aspiring individuals. (Source: Ronald_vanLoon, Ronald_vanLoon)

Ronald_vanLoon

A16Z Speedrun 2026: AI and Entertainment Startup Accelerator : A16Z Speedrun 2026 is now accepting applications from founders in the AI and entertainment sectors. The program supports founders focused on building their own ventures, offering an opportunity for entrepreneurs looking to grow in the rapidly evolving field of AI-powered products. (Source: yoheinakajima)

💼 Business

MiniMax Copyright Lawsuit: AI Unicorn’s IPO Path Clouded : Chinese AI unicorn MiniMax, valued at over $4 billion, is facing a joint copyright lawsuit from Disney, Universal Pictures, and Warner Bros. The lawsuit alleges that its video generation tool “Hulu AI” systematically infringes copyright by generating content featuring protected characters based on user prompts. This case poses a devastating blow to MiniMax’s IPO plans, highlighting the severe challenges of intellectual property compliance in the generative AI market and the importance of balancing technological innovation with legal boundaries. (Source: 36氪)

版权诉讼下的MiniMax:AI独角兽的上市迷途

AI Logistics Company Augment Secures $85 Million in Funding, Total Funding Reaches 800 Million RMB in Five Months Since Launch : AI logistics company Augment successfully completed an $85 million Series A funding round just five months after its launch, bringing its total funding to $110 million (approximately 800 million RMB). Its AI agent product, Augie, automates complex, fragmented tasks across the entire logistics lifecycle, from order intake to payment collection. It has already managed over $35 billion in freight value for dozens of top 3PLs and shippers, saving clients millions of dollars, demonstrating AI’s powerful business value in optimizing labor-intensive logistics. (Source: 36氪)

AI物流火了,上线5个月,拿下8亿融资

Microsoft Integrates Anthropic’s Claude Models into Copilot : Microsoft has integrated Anthropic’s Claude Sonnet 4 and Opus 4.1 models into its Copilot assistant for enterprise users. This move aims to reduce its sole reliance on OpenAI and solidify Microsoft’s position as a neutral platform provider. Enterprise users can now choose between OpenAI and Anthropic models, enhancing flexibility and expected to boost competition in the enterprise AI market. (Source: Reddit r/deeplearning)

🌟 Community

AI’s Impact on Human Understanding and Society: The Paradox of Efficiency and “Information Cocoons” : The community widely fears that AI, like social media, could negatively impact human understanding, critical thinking, and life goals. By optimizing content generation for “spread” rather than “truth” or “depth,” AI might lead to “learning plateaus” and “infinite garbage machines.” This sparks discussions on how to guide AI development towards being a growth tool rather than an addictive one, calling for shifts in regulation, industry practices, and cultural norms. (Source: Reddit r/ArtificialInteligence, Reddit r/artificial, Reddit r/ArtificialInteligence, Yuchenj_UW, colin_fraser, Teknium1, cloneofsimo)

AI in the Workplace: The Contradiction of Efficiency Gains and “Invisible Exploitation” : AI’s integration into the workplace, especially for executive-level employees, is creating a paradox: efficiency gains are often accompanied by higher output expectations, without corresponding rewards for employees. This “invisible exploitation” turns employees into “human quality checkers” for AI-generated content, increasing cognitive load and anxiety. While AI boosts corporate productivity, profits largely flow to capital, exacerbating the “cognitive gap” between strategic managers and tool-dependent executors, potentially leading to widespread burnout if organizational structures are not reconfigured. (Source: 36氪, glennko, mbusigin)

In the AI Era, the Value of “Questioning Power” Surpasses “Execution Power” : In an AI-driven world, true competitive advantage shifts from execution speed to “questioning power”—the ability to identify which problems are worth solving. Over-reliance on AI for execution without critical problem definition can lead to efficiently solving the wrong problems, creating a false sense of progress. Design thinking, empathy mapping, and continuous questioning are considered critical human skills that AI cannot replace, enabling individuals and organizations to effectively leverage AI and focus on solving meaningful challenges. (Source: 36氪)

AI Geopolitics: US-China AI Race and International Regulatory Disputes : The US-China AI race is seen as a marathon, not a sprint, with China potentially leading in robotics applications. The US is urged to focus on practical AI investments rather than solely pursuing superintelligence, but has rejected UN-level international AI oversight, emphasizing national sovereignty. This highlights the complex geopolitical landscape where AI development intertwines with national security, trade policies (like H-1B visas), and competition for AI infrastructure and talent. (Source: Reddit r/ArtificialInteligence, Reddit r/ArtificialInteligence, adcock_brett, Dorialexander, teortaxesTex, teortaxesTex, teortaxesTex, brickroad7, jonst0kes)

Emad Mostaque Predicts “The Last Economy”: AI to Reshape Human Value : Emad Mostaque, former CEO of Stability AI, predicts that AI will fundamentally reshape economic structures within the next 1000 days, and human labor value could become zero or even negative. He proposes the “MIND framework” (Matter, Intellect, Networks, Diversity) to measure economic health, arguing that the surplus of “intellectual capital” brought by AI necessitates a re-evaluation of other capital’s importance. This “fourth inversion” will see AI replace cognitive labor, requiring the establishment of new “human-centric” currencies and universal basic AI to navigate societal transformation. (Source: 36氪)

Stability AI前CEO惊人预测:人类智力价值归零,只剩1000天

AI Hardware Race: OpenAI, ByteDance, and Meta Vie for Consumer Device Market : Tech giants like OpenAI, ByteDance, and Meta are actively investing in consumer AI hardware R&D. Meta’s Ray-Ban AI glasses have achieved significant sales, while OpenAI is reportedly collaborating with Apple suppliers to develop a “screenless smart speaker,” and ByteDance is developing AI smart glasses. This race signals AI’s deeper integration into daily life, with companies exploring diverse product forms and interaction models, aiming to gain an early advantage in ambient AI. (Source: 36氪)

张一鸣和奥特曼都得学小扎

AI Agent: A Paradigm Shift from Human-Computer Collaboration to “Human-Computer Delegation” : The AI industry is approaching a “gentle inflection point” from human-computer collaboration to “human-computer delegation,” where autonomous AI Agents will execute complex tasks at scale. AI’s breakthrough in “penetrating programming” suggests its ability to conquer all semi-open systems. This shift will give rise to “unmanned company” organizational forms, where human roles will shift from micro-execution to macro-governance, focusing on value injection, system architecture design, and macro-navigation, with AI Copilots assisting in decision-making rather than directly intervening in high-speed AI systems. (Source: 36氪)

接近温和拐点,AI将迎来比撒手速度的周期

AI’s Impact on Foreign Language Majors: Students Need to Cultivate “Foreign Language+” Composite Skills : The rise of AI translation technology is profoundly impacting foreign language majors, leading to a decline in demand for traditional language-related positions, with several universities discontinuing relevant programs. Foreign language students face pressure to transform, needing to shift from single language skills to a “foreign language+” composite model, such as “foreign language + AI” for natural language processing, or “foreign language + international communication.” This prompts reforms in foreign language education, emphasizing cross-cultural understanding and comprehensive abilities, rather than mere translation training, to adapt to the new demands for language talent in the AI era. (Source: 36氪, Reddit r/ClaudeAI)

High GPU Prices: AI Demand-Driven and Local LLM Optimization : The community is widely concerned about persistently high GPU prices, believing the main reasons are surging AI data center demand and inflation. Many believe prices are unlikely to drop significantly unless the AI bubble bursts or custom chips become widely adopted. However, to address this challenge, the community is working to optimize local LLM performance, such as AMD MI50 outperforming NVIDIA P40 in llama.cpp/ggml, and utilizing iGPUs for basic LLM tasks, to reduce local AI computing costs. (Source: Reddit r/LocalLLaMA, Reddit r/LocalLLaMA, Reddit r/LocalLLaMA)

46 GB GPU compute for $20.

💡 Other

LLM Scaling’s “Diminishing Returns Illusion” and Long-Horizon Tasks : Research indicates that while single-turn benchmarks might show slowing LLM progress, model scaling still yields non-diminishing improvements in long-horizon task execution. The “diminishing returns illusion” stems from small single-step accuracy gains, which can lead to super-exponential improvements in long-horizon task completion. The sequential computation advantage in long-horizon tasks is unmatched by parallel tests, suggesting that continuous model scaling and reinforcement learning training are crucial for achieving advanced agent behaviors. (Source: scaling01)

scaling01

Energy-Based Transformer (EBT) Enhances Next-Token Prediction Performance : Researchers have introduced the Energy-Based Transformer (EBT), which scores candidate next tokens using “energy” and iteratively verifies and selects them by reducing energy through gradient steps. In experiments with 44 million parameters, EBT outperformed traditional Transformers of comparable size on three out of four benchmarks, indicating that this novel token selection method holds promise for improving LLM performance. (Source: DeepLearningAI)

DeepLearningAI

AI Robotics Advancements: Humanoid Salesperson and Training-Free Autonomous Walking Robot Dog : Xpeng Motors has deployed a humanoid car salesperson named “Tiedan” in its showrooms, showcasing AI’s application in customer-facing robotics. Additionally, a robot dog “with animal reflexes” can walk in the woods without training, highlighting advancements in autonomous robotics in mimicking biological movement and perception. These developments signal the increasing complexity and practical applications of AI in physical robotics. (Source: Ronald_vanLoon, Ronald_vanLoon)

Ronald_vanLoon