Keywords:World Robot Conference, humanoid robots, embodied intelligence, GPT-5, Google DeepMind, AI glasses, LangChain, Reality Proxy AI glasses, Genie 3 world simulator, LEANN vector indexing, Qwen Code free API, GPT-5 priority processing service

🔥 FOCUS

The “Gala” of Embodied AI: 200 Robots Compete : The World Robot Conference (WRC 2025) was grandly held in Beijing, attracting over 220 exhibiting companies and showcasing over 1500 exhibits, including over 100 debut new products from 50 humanoid robot companies. The conference showcased the latest advancements of humanoid robots in various fields such as home services (e.g., making beds, folding clothes), commercial services (e.g., cashiering, coffee making, bartending), industrial applications (e.g., precision assembly, sorting, handling), and medical care (e.g., rehabilitation training, massage). Furthermore, components in the robot industry chain (e.g., planetary roller screws, dexterous hands, tactile sensors) also demonstrated significant innovations, signaling that embodied intelligence is accelerating its integration into the physical world, which is expected to drive the deep fusion of AI with real-world scenarios. (Source: 36氪)
AI Glasses Enable “Remote Object Retrieval”: Reality Proxy : A team of Zhejiang University alumni has developed an AI glasses technology called “Reality Proxy,” which enables users to “retrieve” real-world objects remotely and interact intuitively with them through “digital proxies.” This technology can capture scene structures and generate operable digital proxies, supporting diverse interactive functions such as browsing previews, multi-object selection, filtering by attributes, semantic grouping, and spatial scaling grouping. This innovation merges the physical and digital worlds, significantly enhancing the interaction efficiency and precision of XR devices in complex scenarios like book retrieval, architectural navigation, and drone control, and is considered a key step towards a “Jarvis”-like AI assistant. (Source: 量子位)

AI眼镜“隔空取物”,戴上即可随心选中现实世界任意物体

🎯 DEVELOPMENTS

OpenAI GPT-5 Release and Subsequent Adjustments : OpenAI officially released GPT-5, emphasizing its “routing system” that dynamically allocates model resources based on task complexity and user intent, enabling seamless multimodal collaboration, and significantly reducing factual error rates and hallucinations. However, after the release, users reported a “degradation in performance,” which Sam Altman attributed to an automatic switcher malfunction. He promised a fix and stated that GPT-4o would be restored as an option for Plus users. OpenAI also plans to increase GPT-5’s “temperature” and personalization options to address user preferences for the model’s “personality.” (Source: 36氪, The Verge, The Verge, sama, openai, nickaturley, sama, openai, dotey, dotey, Reddit r/ChatGPT, Reddit r/ChatGPT, Reddit r/artificial, Reddit r/ChatGPT)

Just posted by Sam regarding 4o

Google DeepMind Latest Developments Summary : Google DeepMind recently released a series of AI achievements, including the state-of-the-art world simulator Genie 3, Gemini 2.5 Pro Deep Think opened to Ultra subscribers, providing Gemini Pro for free to university students and investing $1 billion to support US education, the release of the global geospatial model AlphaEarth, and the Aeneas model for deciphering ancient texts. Additionally, Gemini achieved gold medal level in the IMO (International Mathematical Olympiad), launched Storybook, a storybook app with art and audio, added a Kaggle Game Arena LLM benchmark, the asynchronous coding Agent Jules exited Beta, AI search mode launched in the UK, a NotebookLM video overview was released, and Gemma model downloads exceeded 200 million. (Source: demishassabis, Google, Ar_Douillard, _rockt, quocleix)
GLM-4.5 Series Models to be Open-Sourced Soon : Zhipu AI (GLM) announced that its GLM-4.5 series new models will be open-sourced soon, revealing that the model defeated 99% of real players in a map search competition within 16 hours. This move heralds new advancements in the visual model domain, potentially impacting geolocation and image recognition applications. The community has shown high interest in the new model’s specific capabilities and open-source details. (Source: Reddit r/LocalLLaMA)

GLM-4.5 series new models will be open source soon

Cohere Command A Vision Released : The Cohere team has launched Command A Vision, a state-of-the-art generative model designed to provide exceptional multimodal visual task performance for enterprises while maintaining strong text processing capabilities. The release of this model will further boost efficiency and effectiveness for enterprises in applications combining images and text. (Source: dl_weekly)
Meta V-JEPA 2 Released : Meta AI has released V-JEPA 2, a groundbreaking world model focused on visual understanding and prediction. This model is expected to bring significant advancements in robotics and AI, as it helps AI systems better understand and predict visual environments, leading to more complex autonomous behaviors. (Source: Ronald_vanLoon)
OpenAI GPT-5 Introduces Priority Processing Service : OpenAI has introduced “Priority Processing” for GPT-5, allowing developers to achieve faster first-token generation speeds by setting "service_tier": "priority". This feature is crucial for applications sensitive to millisecond-level latency but comes at an additional cost, reflecting OpenAI’s exploration in optimizing model service experience and commercialization. (Source: jeffintime, OpenAIDevs, swyx, juberti)

Priority Processing debuts with GPT-5. under-hyped imo

🧰 TOOLS

Qwen Code Offers Free Call Quota : Alibaba’s Tongyi Qianwen announced that Qwen Code offers 2000 free calls daily, with international users getting 1000 calls via OpenRouter. This initiative significantly lowers the barrier for developers to use code generation tools, potentially promoting innovative applications based on Qwen Code and the popularization of “vibe coding,” making it a strong contender in the field of AI-assisted programming. (Source: huybery, jeremyphoward, op7418, Reddit r/LocalLLaMA)

💡 You get 2,000 free Qwen Code runs every day!

Genie 3 Explores the World of Paintings : Google DeepMind’s Genie 3 demonstrates an amazing ability, allowing users to “step into” and explore their favorite paintings, transforming them into interactive 3D worlds. This feature brings a new dimension to art appreciation, education, and virtual experiences, for instance, enabling users to stroll through Edward Hopper’s “Nighthawks” or Jacques-Louis David’s “The Death of Socrates” for an immersive art experience. (Source: cloneofsimo, jparkerholder, BorisMPower, francoisfleuret, shlomifruchter, _rockt, Vtrivedy10, rbhar90, fchollet, bookwormengr)

Gpt - 5 Thinking, one shot vibe coding: Space sim, meditation app, duo lingo clone, Windows 95

LangChain Launches GPT-5 Playground : LangChain has integrated OpenAI’s latest GPT-5 models (including gpt-5, gpt-5-mini, gpt-5-nano) into its LangSmith Playground, featuring a built-in cost tracking function. This provides developers with a convenient platform to test and build GPT-5-based applications while monitoring API usage costs, helping to optimize development workflows and resource management. (Source: LangChainAI, hwchase17)

Test Driving GPT-5

Claude Code Facilitates Mobile Hotfixes : A developer successfully handled an urgent hotfix in a production environment using Claude Code via a mobile browser at a Taco Bell drive-thru. This demonstrates the powerful practicality of AI coding tools in mobile scenarios, allowing developers to break free from desk constraints and perform code debugging and problem-solving anytime, anywhere, increasing work flexibility. (Source: Reddit r/ClaudeAI)

Shipped a hotfix from a Taco Bell drive-thru using Claude Code

Clode Studio Remote Access Feature : Clode Studio has released an update, adding a built-in Relay Server and multi-tunnel support, allowing users to remotely access desktop IDEs and control Claude Code Chat from any device. This feature offers multiple tunnel options (Clode, Cloudflare, Custom), supports mobile and tablet touch control, and ensures secure authentication, aiming to enhance remote development experience and flexibility. (Source: Reddit r/ClaudeAI)
LEANN: Extremely Lightweight Vector Index : LEANN is an innovative, extremely lightweight vector index that enables fast, accurate, and 100% private RAG (Retrieval Augmented Generation) on MacBooks without an internet connection, with index files 97% smaller than traditional methods. It allows users to perform semantic searches on their local devices, processing personal data like emails and chat logs, providing a personal Jarvis-like experience. (Source: matei_zaharia)

1/N 🚀 Launching LEANN — the tiniest vector index on Earth!

Qwen-Image LoRA Trainer Launched : The WaveSpeedAI platform has launched the Qwen-Image LoRA Trainer, the world’s first platform to offer an online Qwen-Image LoRA trainer. Users can now train their own custom styles in minutes, greatly simplifying the AI art creation process and enhancing the personalization capabilities of image generation models. (Source: Alibaba_Qwen)

🚀 Just Launched: Qwen-Image LoRA Trainer @Alibaba_Qwen

Jules Launches Interactive Plan : Google’s asynchronous coding Agent Jules has released its Interactive Plan feature, allowing Jules to read codebases, ask clarifying questions, and collaborate with users to refine development plans. This collaborative approach increases the likelihood of users clarifying their goals, ensuring human-AI collaboration remains consistent in code generation and solution building, thereby improving code quality and reliability. (Source: julesagent)

We had a fun week this week. One more ship.

Grok 4 PDF Processing Capabilities Upgraded : xAI announced that Grok 4’s PDF processing capabilities have been significantly enhanced. It can now seamlessly process ultra-large PDF files spanning hundreds of pages and better understand PDF content through more acute recognition capabilities. This upgrade has been launched in Grok’s web and mobile applications, greatly enhancing user efficiency in processing and analyzing complex documents. (Source: xai, Yuhu_ai_, Yuhu_ai_, Yuhu_ai_)

Let the games begin

📚 LEARNING

HuggingFace Launches AI Courses : HuggingFace has released 9 free elite-level AI courses, covering core topics such as LLMs, Agents, and AI systems. These courses aim to help developers and researchers master cutting-edge AI technologies, lower learning barriers, and promote the development of the open-source AI community. (Source: huggingface)

If you’re serious about mastering LLMs, agents, and AI systems start here.

Attention Basin: Research on LLM Contextual Position Sensitivity : A study revealed the significant sensitivity of Large Language Models (LLMs) to the contextual position of input information, termed the “attention basin” phenomenon: models tend to allocate higher attention to information at the beginning and end of a sequence while neglecting the middle parts. The study proposes the Attention-Driven Reranking (AttnRank) framework, which significantly improved the performance of 10 different LLMs on multi-hop question answering and few-shot learning tasks by calibrating model attention preferences and reordering retrieved documents or few-shot examples. (Source: HuggingFace Daily Papers)

LisanBenchV2 - Grok-4 smashes expectations

MLLMSeg: Lightweight Mask Decoder Enhances Referring Expression Segmentation : MLLMSeg is a novel framework designed to address the challenges of pixel-level dense prediction in referring expression segmentation (RES) tasks for Multimodal Large Language Models (MLLMs). This framework fully leverages the inherent visual detail features within MLLM visual encoders and proposes detail-enhanced and semantically consistent feature fusion modules. Combined with a lightweight mask decoder, it achieves a better balance between performance and cost, surpassing existing SAM-based and SAM-free methods. (Source: HuggingFace Daily Papers)

Tencent AI Lab introduces R-Zero!

Learning to Reason for Factuality : A study proposes a novel reward function aimed at addressing the high hallucination rates of reasoning-based Large Language Models (R-LLMs) in long-form factual tasks. This reward function simultaneously considers factual accuracy, level of response detail, and answer relevance. Through online reinforcement learning training, the model’s average hallucination rate was reduced by 23.1 percentage points across six factual benchmarks, the answer detail level increased by 23%, without compromising overall response utility. (Source: HuggingFace Daily Papers)

Learning to Reason for Factuality

LangChain to Host Hacking Hours : LangChain will host “LangChain Hacking Hours,” providing a focused co-working environment where developers can make tangible progress on their LangChain or LangGraph projects, receive direct technical guidance from the team, and network with other builders in the community. (Source: LangChainAI)

🛠️ Building AI Agents solo can feel isolating when you hit technical roadblocks or need fresh perspectives on your approach.

DSPy: Faithfulness in RAG Pipelines : Discussions on social media highlighted the advantages of the DSPy framework in maintaining faithfulness within RAG (Retrieval Augmented Generation) pipelines. With DSPy, developers can engineer systems to proactively output “I don’t know” when the context does not contain the necessary information, thereby avoiding model hallucinations. It also simplifies the complexity of prompt engineering by separating business objectives, models, processes, and training data. (Source: lateinteraction, lateinteraction, lateinteraction)

I love this @DSPyOSS pattern for keeping a RAG pipeline faithful.

AI Evals Course Insights : Hamel Husain shared 14 highlights from his AI Evals course, particularly prominent ideas regarding retrieval (RAG). The course emphasized the importance of evaluation in AI system development and how to effectively leverage retrieval techniques to enhance model performance, especially when dealing with complex data and multi-source information. (Source: HamelHusain)

14 standout ideas from Lesson 5, Lesson 6 and Chapter 7 from Hamel and Shreya's AI Evals course.

Anthropic Pledges to Advance AI Education : Anthropic has joined the “Pledge to America’s Youth” initiative, collaborating with over 100 organizations to advance AI education. They will work with educators, students, and communities nationwide to cultivate essential AI and cybersecurity skills for the next generation, addressing the challenges of future technological developments. (Source: AnthropicAI)

We joined the Pledge to America's Youth along with 100+ organizations committed to advancing AI education.

The Nature of Chain-of-Thought (CoT) Reasoning : There’s been a heated discussion regarding whether CoT reasoning is a “mirage.” One study, analyzing from a data distribution perspective, questions CoT’s true understanding capabilities, suggesting it may overfit benchmark tasks and be prone to hallucinations. Meanwhile, other views contend that CoT can still provide valuable information in complex cognitive tasks, and its “traces of thought” remain credible under specific conditions. (Source: togelius, METR_Evals, rao2z, METR_Evals, METR_Evals)

📢 Excited to share our new arXiv preprint: "Is Chain-of-Thought Reasoning of LLMs a Mirage? A Data Distribution Lens"

How LLMs Predict the Next Word : A video shared on social media visually demonstrates how Large Language Models (LLMs) generate text by predicting the next word. This helps users understand the fundamental working principle of LLMs, which involves selecting the most probable next word based on probability distributions to construct coherent and meaningful sequences. (Source: Reddit r/deeplearning)
Necessity of Independent Projections for Q, K, V in Transformer Models : The community discussed the reasons for independent projections of Query (Q), Key (K), and Value (V) in Transformer models. The discussion pointed out that directly binding Q and V to input embeddings would compromise the model’s expressive power and flexibility, as independent projections allow the model to query, match, and extract information in different semantic spaces, thereby capturing more complex dependencies and multi-head attention mechanisms. (Source: Reddit r/deeplearning)
Adaptive Classifiers: New Architecture for Few-Shot Learning : A study proposes the “Adaptive Classifiers” architecture, enabling text classifiers to learn from a few samples (5-10 per class), continuously adapt to new data without catastrophic forgetting, and dynamically add new categories without retraining. This solution combines prototype learning and elastic weight consolidation, achieving 90-100% accuracy in enterprise-level tasks with fast inference speed, addressing ML deployment challenges in data-scarce and rapidly changing scenarios. (Source: Reddit r/MachineLearning)

Adaptive Classifiers: Few-Shot Learning with Continuous Adaptation and Dynamic Class Addition

Dynamic Fine-Tuning (DFT) Enhances SFT : A study proposes “Dynamic Fine-Tuning (DFT),” which redefines SFT (Supervised Fine-Tuning) as reinforcement learning and introduces a single-line code modification to stabilize token updates, thereby improving SFT’s performance. DFT surpassed RL methods like PPO, DPO, and GRPO in some cases, offering a more efficient and stable new approach for model fine-tuning. (Source: TheTuringPost)

Dynamic Fine-Tuning (DFT) - a single-line code change generalizing SFT.

💼 BUSINESS

OpenAI GPT-5 Pricing Strategy Triggers Price War Speculation : OpenAI released GPT-5 with API pricing ($1.25/1M input, $10/1M output) significantly lower than its competitor Anthropic Claude Opus 4.1 ($15/1M input, $75/1M output). This move is seen as a “game-changer” that could trigger a price war in the LLM market. The industry is watching to see if this is a short-term market share disruption or the beginning of long-term AI cost reduction, and how it will impact AI tool development, business models, and AI accessibility. (Source: Reddit r/ArtificialInteligence)

OpenAI just priced GPT-5 so low it might trigger an AI price war ,who wins here?

GPU Resource Centralization and the AI Industry Landscape : Comments indicate that the high concentration of GPU resources has led to “GPU-rich labs” dominating the general AI field, making it difficult for open models to compete. The article suggests that 2025 will be the year of Agents and application layers, and enterprises should focus on building acceptable solutions on the smallest LLMs rather than spending vast sums training large models, reflecting a strategic shift in the AI industry from model training to application deployment. (Source: Reddit r/artificial)
AI Company Equity Trading Irregularities : Social media has revealed phenomena of “underlying predators” and “scammers” in AI lab equity trading. These multi-layered SPV (Special Purpose Vehicle) brokers, who have no direct affiliation with the companies themselves, engage in fraudulent activities, warning investors and the public to be wary of the growing irrational exuberance and potential risks in the AI sector. (Source: saranormous)

🌟 COMMUNITY

GPT-5 Release Triggers Strong User Reactions and Controversy : Following OpenAI’s release of GPT-5, widespread discussion erupted within the community. Some users expressed disappointment with GPT-5’s performance (especially in programming and creative writing), finding it inferior to GPT-4o or Claude Code, and even perceiving a “regression.” They also voiced dissatisfaction with OpenAI’s “automatic switcher” feature, model transparency, and adjustments to Plus user access limits. Many users expressed nostalgia and fondness for GPT-4o’s “personality” and “emotions,” considering it not just a tool but a “friend” or “companion,” even launching a petition to demand OpenAI restore the 4o option. Sam Altman responded, stating that the company underestimated user preferences for 4o’s “personality” and promised to restore 4o as an option for Plus users, while also improving GPT-5’s “temperature” and personalization features. He also explained that initial suboptimal model performance was due to technical glitches during the launch. (Source: maithra_raghu, teortaxesTex, teortaxesTex, teortaxesTex, SebastienBubeck, SebastienBubeck, shaneguML, OfirPress, cloneofsimo, TheZachMueller, scaling01, Smol_AI, natolambert, teortaxesTex, Vtrivedy10, tokenbender, ClementDelangue, TheZachMueller, TomLikesRobots, METR_Evals, Ronald_vanLoon, teortaxesTex, teortaxesTex, scaling01, scaling01, scaling01, scaling01, scaling01, scaling01, scaling01, scaling01, scaling01, scaling01, Teknium1, Teknium1, Teknium1, Teknium1

Leave a Reply

Your email address will not be published. Required fields are marked *