In this talk, we present our experience in scaling hybrid linear attention architectures to trillion-scale, through two models from the Ling Team: Ling-2.5-1T and Ring-2.5-1T. These models integrate linear attention with selected softmax attention layers to support efficient long-context training while preserving strong reasoning and representation capability. We share key algorithm–system co-design insights that make trillion-scale hybrid attention practical, including stability techniques for large-scale linear attention training and efficient distributed training for ultra-long sequences.
We are excited to present Amazon Nova, a groundbreaking portfolio of AI offerings that deliver frontier intelligence and industry-leading price performance. Nova is built on advanced AI technologies originally developed for Amazon's internal applications, such as Alexa+, Amazon Ads, and AWS Marketplace, and is now available to AWS customers. Amazon Nova includes Nova models, fast and cost-effective foundation models for text and multimodal needs; Nova Forge, a new service to build your own frontier models; and Nova Act, a new service to build agents that automate browser-based UI workflows powered by a custom Nova 2 Lite model. These models and services have built-in controls for the safe and responsible use of AI, delivering robust protections, content filters, and policy-aligned behaviors to meet compliance requirements. During our demo, a product engineer and researcher will showcase the product and the science behind it. We will demonstrate how Nova models can be customized to deliver responses that reflect industry expertise, powering interactive chat interfaces, Retrieval-Augmented Generation (RAG) systems, agentic applications, video analysis, and UI workflow automation solutions. We will also highlight the multimodal capabilities of Nova, which accept text, image, or video inputs to generate text output, and the creative content generation models that accept text and image inputs to generate image or video output. Amazon Nova has been adopted by tens of thousands of customers across industries, delivering measurable impact with cost savings and gains in productivity, automation, and quality with real-world deployments. We believe that Nova will be a valuable addition to the ICLR conference and look forward to sharing our insights and experiences with you.
Building Physical AI at Scale: Data, Infrastructure, and Evaluation for the Real World
Physical AI — robots, autonomous vehicles, and embodied agents — is approaching a genuine inflection point. Foundation models for real-world interaction are becoming viable, hardware costs are dropping, and developer interest is surging. Yet most teams building in this space are still stitching together their development stack from incompatible pieces, and it is slowing them down. The core bottlenecks are well understood but rarely addressed together. Real-world robotic behavior cannot be learned from synthetic data alone — collecting, annotating, and validating diverse physical-world data at scale is a full operational discipline. Training multimodal vision-language-action models demands infrastructure purpose-built for the task. And evaluating whether a model actually works in the physical world requires benchmarking approaches that go far beyond standard leaderboards. This social will bring together researchers and practitioners to examine all three problems in parallel. Short talks from speakers with hands-on experience in physical AI development will cover the state of real-world data pipelines, what purpose-built infrastructure for physical AI actually looks like, and how the community is approaching evaluation for embodied systems. An open discussion will follow, focused on where the biggest unsolved problems lie and how the research community can contribute.
AI compensation and negotiation
As AI reshapes industries, compensation is changing faster than researchers can track. New labs, startups, and top companies are competing for talent with vastly different pay structures, currencies, and cultural norms. Yet most researchers are never formally taught how to understand their worth or navigate these systems. The result is an uneven landscape where brilliant minds often make life-changing decisions without the information they need.
The session begins with a concise, data-driven talk on current AI compensation and negotiation trends, grounded in real stories and case studies. From there, a fireside chat and open Q&A invite candid, experience-driven insights from researchers who have navigated these conversations firsthand.
Key Takeaways for attendees: * How to evaluate compensation (salary, equity, bonuses) across industry roles such as Research/Applied/Data Scientists, Research/Machine Learning/Software Engineers, and more * How to compare global opportunities and account for regional differences in pay structures * How to identify leverage points and negotiate effectively at different career stages and levels * How to respond to pushback and recognize red vs. green flags in job offers * How to negotiate the deadline, and avoid having an offer rescinded * How to advocate for yourself without counteroffers and amidst having fears
Digital Real Estate Social
Real estate is one of the most financially significant and emotionally complex decisions people make — yet digital experiences have historically optimized for information retrieval (i.e., home search) rather than intelligent and personalized guidance. In this talk, I’ll share how we are evolving from high-IQ systems that answer questions to agentic systems that demonstrate AQ (Agentic Quotient) and EQ (Emotional Intelligence). By combining deep, panoptic understanding of homes and users with advanced reasoning, planning, and tool use, we are building AI copilots that don’t just respond passively — they guide, anticipate, and act proactively. I’ll discuss the key technical challenges, and how we operationalize agentic AI in production, architect for trust in high-stakes decisions, and design experiences that transform the home journey from fragmented search into confident, personalized progression.
Digital Olfaction Social
This discussion booth explores the emerging intersection of machine learning and smell technology, also known as digital olfaction. The space is designed for researchers, students, builders, and curious attendees interested in how AI can be used to detect, classify, generate, and interpret scent-related data. Topics may include electronic noses, olfactory sensing, multimodal AI, applications in healthcare, food quality, environmental monitoring, fragrance, and human-computer interaction. The goal is to create an open and interdisciplinary conversation around both the technical challenges and the creative opportunities in this fast-growing field. Whether you work directly on ML models for chemical signal analysis or are simply interested in the future of smell interfaces and sensory intelligence, this booth offers a place to exchange ideas, share projects, discuss collaborations, and connect with others working at the frontier of AI and olfaction.
Our results show that hybrid linear attention enables efficient scaling, strong reasoning, and robust long-context capability, providing a practical path toward next-generation trillion-scale foundation models.
openJiuwen is an open-source Agent framework and platform aimed at helping users build precise, simple, efficient, and engineering production-ready AI Agents. This presentation will cover two topics related to openJiuwen. First, we will introduce the key features and design philosophy of openJiuwen, including agent self-evolution, context compression and offloading, and the agent controller. Second, we will introduce an agent application developed based on openJiuwen: JiuwenClaw, an AI assistant that supports self-evolving capabilities and intelligent task management.
Digital Real Estate Social
Real estate is one of the most financially significant and emotionally complex decisions people make — yet digital experiences have historically optimized for information retrieval (i.e., home search) rather than intelligent and personalized guidance. In this talk, I’ll share how we are evolving from high-IQ systems that answer questions to agentic systems that demonstrate AQ (Agentic Quotient) and EQ (Emotional Intelligence). By combining deep, panoptic understanding of homes and users with advanced reasoning, planning, and tool use, we are building AI copilots that don’t just respond passively — they guide, anticipate, and act proactively. I’ll discuss the key technical challenges, and how we operationalize agentic AI in production, architect for trust in high-stakes decisions, and design experiences that transform the home journey from fragmented search into confident, personalized progression.
Video AI for Real-World Systems Social
Join us for an evening hosted by a team building video AI systems for real-world deployment, bringing together researchers, engineers, and builders working at the frontier of video AI. This is a relaxed, conversation-first gathering connecting people across robotics, self-driving, and embodied AI.
We’ll focus on a new shift in how video is used in real-world AI: moving beyond static labeling toward systems that can search, reason over, and validate events directly from raw footage. Discussions will explore approaches that combine vision-language models, motion understanding, and multi-stage validation to turn large-scale video into something you can actually act on.
No formal talks, just high-signal conversations with people actively building and deploying these systems.
The goal is to go beyond surface-level ideas and get into what it actually takes to make Physical AI work reliably in the real world. If you are working on autonomy, VLMs, or production AI systems, this is a chance to exchange ideas and shape what comes next.
| ICLR uses cookies for essential functions only. We do not sell your personal information. Our Privacy Policy » |