{"id":1886,"date":"2025-11-16T10:29:58","date_gmt":"2025-11-16T10:29:58","guid":{"rendered":"https:\/\/scipapermill.com\/index.php\/2025\/11\/16\/autonomous-drivings-next-gear-unifying-perception-planning-and-safety-with-ai-2\/"},"modified":"2025-12-28T21:20:41","modified_gmt":"2025-12-28T21:20:41","slug":"autonomous-drivings-next-gear-unifying-perception-planning-and-safety-with-ai-2","status":"publish","type":"post","link":"https:\/\/scipapermill.com\/index.php\/2025\/11\/16\/autonomous-drivings-next-gear-unifying-perception-planning-and-safety-with-ai-2\/","title":{"rendered":"Autonomous Driving&#8217;s Next Gear: Unifying Perception, Planning, and Safety with AI"},"content":{"rendered":"<h3>Latest 50 papers on autonomous driving: Nov. 16, 2025<\/h3>\n<p>Autonomous driving (AD) systems are rapidly evolving, driven by groundbreaking advancements in AI and Machine Learning. From robust perception in challenging conditions to intelligent, safe decision-making and efficient simulation, the research landscape is buzzing with innovation. This post dives into recent breakthroughs, synthesizing key insights from a collection of papers that push the boundaries of what\u2019s possible in AD.<\/p>\n<h3 id=\"the-big-ideas-core-innovations\">The Big Idea(s) &amp; Core Innovations<\/h3>\n<p>One central theme is enhancing <strong>robust perception<\/strong> under various real-world challenges. For instance, \u201c<a href=\"https:\/\/doi.org\/10.1109\/TCSVT.2025.3628019\">DGFusion: Dual-guided Fusion for Robust Multi-Modal 3D Object Detection<\/a>\u201d from Affiliation 1, Affiliation 2, and Affiliation 3, introduces a dual-guided fusion approach to improve 3D object detection by leveraging cross-modal interactions between LiDAR and camera data, especially in adverse conditions. Similarly, \u201c<a href=\"https:\/\/arxiv.org\/pdf\/2511.07862\">MonoCLUE : Object-Aware Clustering Enhances Monocular 3D Object Detection<\/a>\u201d by researchers from Yonsei University, addresses object detection in occluded scenes using local clustering and generalized scene memory. Complementing this, \u201c<a href=\"https:\/\/arxiv.org\/pdf\/2511.07925\">HD<span class=\"math inline\"><sup>2<\/sup><\/span>-SSC: High-Dimension High-Density Semantic Scene Completion for Autonomous Driving<\/a>\u201d from Peking University tackles the input-output dimension and annotation-reality density gaps in camera-based semantic scene completion, significantly improving 3D scene understanding.<\/p>\n<p>Beyond basic perception, <strong>safety and resilience<\/strong> are paramount. The \u201c<a href=\"https:\/\/arxiv.org\/pdf\/2511.09032\">Argus: Resilience-Oriented Safety Assurance Framework for End-to-End ADSs<\/a>\u201d by the Argus Research Group proposes runtime safety checks to detect and mitigate risks in real-time. This is crucial for avoiding issues highlighted in \u201c<a href=\"https:\/\/arxiv.org\/pdf\/2511.08015\">Invisible Triggers, Visible Threats! Road-Style Adversarial Creation Attack for Visual 3D Detection in Autonomous Driving<\/a>\u201d from Xi\u2019an Jiaotong University, which exposes vulnerabilities to stealthy adversarial attacks that induce false positives. The need for robust accident anticipation is addressed by \u201c<a href=\"https:\/\/arxiv.org\/pdf\/2511.08640\">Predict and Resist: Long-Term Accident Anticipation under Sensor Noise<\/a>\u201d by researchers from the University of Macau and Zhejiang University, which uses diffusion-based denoising with actor-critic models for earlier, more reliable predictions under sensor noise.<\/p>\n<p><strong>Intelligent planning and decision-making<\/strong> are also seeing rapid progress, often driven by Large Language Models (LLMs). \u201c<a href=\"https:\/\/arxiv.org\/pdf\/2410.15281\">LLM4AD: Large Language Models for Autonomous Driving \u2013 Concept, Review, Benchmark, Experiments, and Future Trends<\/a>\u201d by a team from Harbin Institute of Technology, offers a comprehensive review of LLMs as decision-making \u2018brains\u2019. \u201c<a href=\"https:\/\/arxiv.org\/pdf\/2511.09025\">FLAD: Federated Learning for LLM-based Autonomous Driving in Vehicle-Edge-Cloud Networks<\/a>\u201d by researchers from Northeastern University, China and University of Victoria, Canada, introduces a privacy-preserving federated learning framework that leverages LLMs across vehicle-edge-cloud networks for efficient and adaptive driving. Further refining LLM integration, \u201c<a href=\"https:\/\/arxiv.org\/pdf\/2511.06253\">AdaDrive: Self-Adaptive Slow-Fast System for Language-Grounded Autonomous Driving<\/a>\u201d by The Chinese University of Hong Kong, Shenzhen and Baidu Inc., dynamically determines when and how LLMs contribute to decision-making, balancing high-level reasoning with real-time efficiency.<\/p>\n<p>Lastly, advancements in <strong>simulation and data generation<\/strong> are critical for testing and training. \u201c<a href=\"https:\/\/arxiv.org\/pdf\/2511.04769\">ReGen: Generative Robot Simulation via Inverse Design<\/a>\u201d from Massachusetts Institute of Technology and Toyota Research Institute, uses LLMs to synthesize diverse and controllable scenarios for robot simulations. \u201c<a href=\"https:\/\/arxiv.org\/pdf\/2511.09080\">D-AWSIM: Distributed Autonomous Driving Simulator for Dynamic Map Generation Framework<\/a>\u201d by M. Tsukada et al., proposes a distributed simulator for dynamic map generation, enhancing accuracy and adaptability in complex environments.<\/p>\n<h3 id=\"under-the-hood-models-datasets-benchmarks\">Under the Hood: Models, Datasets, &amp; Benchmarks<\/h3>\n<p>Recent research heavily relies on and contributes to critical resources for advancing autonomous driving:<\/p>\n<ul>\n<li><strong>Models &amp; Frameworks:<\/strong>\n<ul>\n<li><strong>LongComp:<\/strong> A language-conditioned framework for robust zero-shot trajectory prediction (<a href=\"https:\/\/openreview.net\/\">LongComp: Long-Tail Compositional Zero-Shot Generalization for Robust Trajectory Prediction<\/a>)<\/li>\n<li><strong>O-CTRL:<\/strong> An operator-based algorithm for continuous-time offline reinforcement learning, linking to Hamilton\u2013Jacobi\u2013Bellman equation (<a href=\"https:\/\/arxiv.org\/pdf\/2511.10383\">Operator Models for Continuous-Time Offline Reinforcement Learning<\/a>)<\/li>\n<li><strong>DGFusion:<\/strong> A dual-guided fusion method for robust multi-modal 3D object detection.<\/li>\n<li><strong>FQ-PETR:<\/strong> A fully quantized Position Embedding Transformer for efficient multi-view 3D object detection, with code at <a href=\"https:\/\/github.com\/JiangYongYu1\/FQ-PETR\">https:\/\/github.com\/JiangYongYu1\/FQ-PETR<\/a>.<\/li>\n<li><strong>D-AWSIM:<\/strong> A distributed simulator for dynamic map generation in autonomous driving, with associated code at <a href=\"https:\/\/arxiv.org\/pdf\/2511.09080\">https:\/\/arxiv.org\/pdf\/2511.09080<\/a>.<\/li>\n<li><strong>Argus:<\/strong> A resilience-oriented safety assurance framework for end-to-end ADSs, with a toolkit at <a href=\"https:\/\/argus4ads.github.io\/Ar\">https:\/\/argus4ads.github.io\/Ar<\/a>.<\/li>\n<li><strong>FLAD:<\/strong> A federated learning framework for LLM-based AD, using the SWIFT scheduler and CELLAdapt for LLM adaptation.<\/li>\n<li><strong>UniMM-V2X:<\/strong> An MoE-enhanced multi-level fusion framework for end-to-end cooperative AD, code at <a href=\"https:\/\/github.com\/Souig\/UniMM-V2X\">https:\/\/github.com\/Souig\/UniMM-V2X<\/a>.<\/li>\n<li><strong>VLDrive:<\/strong> A lightweight vision-augmented MLLM for efficient language-grounded autonomous driving, code at <a href=\"https:\/\/github.com\/ReaFly\/VLDrive\">https:\/\/github.com\/ReaFly\/VLDrive<\/a>.<\/li>\n<li><strong>AdaDrive:<\/strong> A self-adaptive slow-fast system for language-grounded AD, code at <a href=\"https:\/\/github.com\/ReaFly\/AdaDrive\">https:\/\/github.com\/ReaFly\/AdaDrive<\/a>.<\/li>\n<li><strong>Polymap:<\/strong> A method for high-definition map generation using rasterized polygons.<\/li>\n<li><strong>REL (Relative Energy Learning):<\/strong> A framework for LiDAR Out-of-Distribution Detection, which introduces Point Raise for pseudo-OOD data synthesis.<\/li>\n<li><strong>SAML (Semantic Meta-Learning):<\/strong> A framework for long-tail motion forecasting in autonomous driving.<\/li>\n<li><strong>MonoCLUE:<\/strong> Monocular 3D object detection framework leveraging local clustering and generalized scene memory, with code at <a href=\"https:\/\/github.com\/SungHunYang\/MonoCLUE\">https:\/\/github.com\/SungHunYang\/MonoCLUE<\/a>.<\/li>\n<li><strong>HD2-SSC:<\/strong> High-Dimension High-Density Semantic Scene Completion framework, code at <a href=\"https:\/\/github.com\/PKU-ICST-MIPL\/HD2-AAAI2026\">https:\/\/github.com\/PKU-ICST-MIPL\/HD2-AAAI2026<\/a>.<\/li>\n<li><strong>FractalCloud:<\/strong> A fractal-inspired architecture for efficient large-scale point cloud processing, with code at <a href=\"https:\/\/github.com\/fractalcloud-team\/fractalcloud\">https:\/\/github.com\/fractalcloud-team\/fractalcloud<\/a>.<\/li>\n<li><strong>X-Scene:<\/strong> A framework for large-scale driving scene generation with high fidelity and flexible controllability, project page at <a href=\"https:\/\/x-scene.github.io\/\">https:\/\/x-scene.github.io\/<\/a>.<\/li>\n<li><strong>TransParking:<\/strong> A dual-decoder transformer framework with soft localization for end-to-end automatic parking, code at <a href=\"https:\/\/github.com\/TransParking\/TransParking\">https:\/\/github.com\/TransParking\/TransParking<\/a>.<\/li>\n<li><strong>DIAL-GS:<\/strong> Dynamic Instance Aware Reconstruction for Label-free Street Scenes with 4D Gaussian Splatting.<\/li>\n<li><strong>UniSplat:<\/strong> Unified Spatio-Temporal Fusion via 3D Latent Scaffolds for Dynamic Driving Scene Reconstruction, project page at <a href=\"https:\/\/chenshi3.github.io\/unisplat.github.io\/\">https:\/\/chenshi3.github.io\/unisplat.github.io\/<\/a>.<\/li>\n<li><strong>TIWM (Token Is All You Need):<\/strong> A cognitive-inspired model for planning through sparse intent alignment.<\/li>\n<li><strong>STATIC:<\/strong> Model for video monocular depth estimation, code at <a href=\"https:\/\/github.com\/sunghun98\/static\">https:\/\/github.com\/sunghun98\/static<\/a>.<\/li>\n<li><strong>AdvRoad:<\/strong> An adversarial attack method for visual 3D detection, code at <a href=\"https:\/\/github.com\/WangJian981002\/AdvRoad\">https:\/\/github.com\/WangJian981002\/AdvRoad<\/a>.<\/li>\n<li><strong>GTNS (Game-Theoretic Nested Search):<\/strong> A method for effective game-theoretic motion planning.<\/li>\n<li><strong>SAFe-Copilot:<\/strong> Unified Shared Autonomy Framework.<\/li>\n<li><strong>Analytic World Models (AWMs):<\/strong> Approach for efficient vehicle dynamics modeling using differentiable simulation, code at <a href=\"https:\/\/github.com\/nvidia\/warp\">https:\/\/github.com\/nvidia\/warp<\/a>.<\/li>\n<\/ul>\n<\/li>\n<li><strong>Datasets &amp; Benchmarks:<\/strong>\n<ul>\n<li><strong>DriveRLR:<\/strong> A benchmark for assessing LLM robustness in evaluating driving scenario realism, with code and dataset at <a href=\"https:\/\/github.com\/Simula-COMPLEX\/DriveRLR\">https:\/\/github.com\/Simula-COMPLEX\/DriveRLR<\/a>.<\/li>\n<li><strong>SnowyLane Dataset:<\/strong> Tailored for winter road scenarios to improve lane detection (<a href=\"https:\/\/ekut-es.github.io\/snowy-lane\">SnowyLane: Robust Lane Detection on Snow-covered Rural Roads Using Infrastructural Elements<\/a>).<\/li>\n<li><strong>MAROON Dataset:<\/strong> A novel multimodal dataset capturing near-field radar and optical depth imaging data for cross-modal comparison (<a href=\"https:\/\/arxiv.org\/pdf\/2411.00527\">MAROON: A Framework for the Joint Characterization of Near-Field High-Resolution Radar and Optical Depth Imaging Techniques<\/a>).<\/li>\n<li><strong>SemanticKITTI, Spotting the Unexpected (STU), nuScenes, NGSIM, HighD, DDAD, NYUv2, BDD100K, DAIR-V2X, DeepScenario:<\/strong> Widely used benchmark datasets for perception, prediction, and control tasks in autonomous driving.<\/li>\n<\/ul>\n<\/li>\n<\/ul>\n<h3 id=\"impact-the-road-ahead\">Impact &amp; The Road Ahead<\/h3>\n<p>These advancements signify a pivotal moment for autonomous driving. The integration of LLMs for high-level reasoning and human-like interaction (as discussed in <a href=\"https:\/\/arxiv.org\/pdf\/2410.15281\">LLM4AD: Large Language Models for Autonomous Driving \u2013 Concept, Review, Benchmark, Experiments, and Future Trends<\/a>) promises more intuitive and context-aware systems. Innovations in robust perception, especially in adverse conditions like snow (<a href=\"https:\/\/arxiv.org\/pdf\/2511.05108\">SnowyLane: Robust Lane Detection on Snow-covered Rural Roads Using Infrastructural Elements<\/a>) or occlusions (<a href=\"https:\/\/arxiv.org\/pdf\/2511.07862\">MonoCLUE : Object-Aware Clustering Enhances Monocular 3D Object Detection<\/a>), directly translate to safer real-world deployments. The push for efficient, quantized models (<a href=\"https:\/\/arxiv.org\/pdf\/2511.09347\">FQ-PETR: Fully Quantized Position Embedding Transformation for Multi-View 3D Object Detection<\/a>) and model compression (<a href=\"https:\/\/arxiv.org\/pdf\/2511.05557\">Compressing Multi-Task Model for Autonomous Driving via Pruning and Knowledge Distillation<\/a>) is crucial for deploying complex AI on resource-constrained automotive platforms.<\/p>\n<p>Looking ahead, the focus will continue to be on building comprehensive <strong>world models<\/strong> that can simulate complex interactions and provide counterfactual reasoning, as laid out in \u201c<a href=\"https:\/\/arxiv.org\/pdf\/2511.08585\">Simulating the Visual World with Artificial Intelligence: A Roadmap<\/a>\u201d by Carnegie Mellon University and Nanyang Technological University researchers. Addressing causal confusion (<a href=\"https:\/\/arxiv.org\/pdf\/2511.08214\">Prioritizing Perception-Guided Self-Supervision: A New Paradigm for Causal Modeling in End-to-End Autonomous Driving<\/a>) and long-tail scenarios (<a href=\"https:\/\/arxiv.org\/pdf\/2511.06649\">Differentiable Semantic Meta-Learning Framework for Long-Tail Motion Forecasting in Autonomous Driving<\/a>) will be key to achieving true generalization. Moreover, the emphasis on <strong>dataset safety<\/strong> (<a href=\"https:\/\/arxiv.org\/pdf\/2511.08439\">Dataset Safety in Autonomous Driving: Requirements, Risks, and Assurance<\/a>) and runtime safety monitoring (<a href=\"https:\/\/arxiv.org\/pdf\/2511.05982\">Runtime Safety Monitoring of Deep Neural Networks for Perception: A Survey<\/a>) underscores the growing importance of verifiable and trustworthy AI. The convergence of these fields\u2014from quantum computing for resource optimization (<a href=\"https:\/\/ieeexplore.ieee.org\/document\/10570867\/\">Coherent Optical Quantum Computing-Aided Resource Optimization for Transportation Digital Twin Construction<\/a>) to novel adversarial robustness techniques\u2014paints a picture of an autonomous future that is not just intelligent, but also inherently safer and more efficient. The journey is far from over, but with these innovations, autonomous driving is accelerating toward a truly transformative era.<\/p>\n","protected":false},"excerpt":{"rendered":"<p>Latest 50 papers on autonomous driving: Nov. 16, 2025<\/p>\n","protected":false},"author":1,"featured_media":0,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"_yoast_wpseo_focuskw":"","_yoast_wpseo_title":"","_yoast_wpseo_metadesc":"","_jetpack_memberships_contains_paid_content":false,"footnotes":"","jetpack_publicize_message":"","jetpack_publicize_feature_enabled":true,"jetpack_social_post_already_shared":true,"jetpack_social_options":{"image_generator_settings":{"template":"highway","default_image_id":0,"font":"","enabled":false},"version":2}},"categories":[56,55,123],"tags":[124,1556,127,134,78,353],"class_list":["post-1886","post","type-post","status-publish","format-standard","hentry","category-artificial-intelligence","category-computer-vision","category-robotics","tag-autonomous-driving","tag-main_tag_autonomous_driving","tag-end-to-end-autonomous-driving","tag-knowledge-distillation","tag-large-language-models-llms","tag-trajectory-prediction"],"yoast_head":"<!-- This site is optimized with the Yoast SEO plugin v27.3 - https:\/\/yoast.com\/product\/yoast-seo-wordpress\/ -->\n<title>Autonomous Driving&#039;s Next Gear: Unifying Perception, Planning, and Safety with AI<\/title>\n<meta name=\"description\" content=\"Latest 50 papers on autonomous driving: Nov. 16, 2025\" \/>\n<meta name=\"robots\" content=\"index, follow, max-snippet:-1, max-image-preview:large, max-video-preview:-1\" \/>\n<link rel=\"canonical\" href=\"https:\/\/scipapermill.com\/index.php\/2025\/11\/16\/autonomous-drivings-next-gear-unifying-perception-planning-and-safety-with-ai-2\/\" \/>\n<meta property=\"og:locale\" content=\"en_US\" \/>\n<meta property=\"og:type\" content=\"article\" \/>\n<meta property=\"og:title\" content=\"Autonomous Driving&#039;s Next Gear: Unifying Perception, Planning, and Safety with AI\" \/>\n<meta property=\"og:description\" content=\"Latest 50 papers on autonomous driving: Nov. 16, 2025\" \/>\n<meta property=\"og:url\" content=\"https:\/\/scipapermill.com\/index.php\/2025\/11\/16\/autonomous-drivings-next-gear-unifying-perception-planning-and-safety-with-ai-2\/\" \/>\n<meta property=\"og:site_name\" content=\"SciPapermill\" \/>\n<meta property=\"article:publisher\" content=\"https:\/\/www.facebook.com\/people\/SciPapermill\/61582731431910\/\" \/>\n<meta property=\"article:published_time\" content=\"2025-11-16T10:29:58+00:00\" \/>\n<meta property=\"article:modified_time\" content=\"2025-12-28T21:20:41+00:00\" \/>\n<meta property=\"og:image\" content=\"https:\/\/i0.wp.com\/scipapermill.com\/wp-content\/uploads\/2025\/07\/cropped-icon.jpg?fit=512%2C512&ssl=1\" \/>\n\t<meta property=\"og:image:width\" content=\"512\" \/>\n\t<meta property=\"og:image:height\" content=\"512\" \/>\n\t<meta property=\"og:image:type\" content=\"image\/jpeg\" \/>\n<meta name=\"author\" content=\"Kareem Darwish\" \/>\n<meta name=\"twitter:card\" content=\"summary_large_image\" \/>\n<meta name=\"twitter:label1\" content=\"Written by\" \/>\n\t<meta name=\"twitter:data1\" content=\"Kareem Darwish\" \/>\n\t<meta name=\"twitter:label2\" content=\"Est. reading time\" \/>\n\t<meta name=\"twitter:data2\" content=\"7 minutes\" \/>\n<script type=\"application\/ld+json\" class=\"yoast-schema-graph\">{\"@context\":\"https:\\\/\\\/schema.org\",\"@graph\":[{\"@type\":\"Article\",\"@id\":\"https:\\\/\\\/scipapermill.com\\\/index.php\\\/2025\\\/11\\\/16\\\/autonomous-drivings-next-gear-unifying-perception-planning-and-safety-with-ai-2\\\/#article\",\"isPartOf\":{\"@id\":\"https:\\\/\\\/scipapermill.com\\\/index.php\\\/2025\\\/11\\\/16\\\/autonomous-drivings-next-gear-unifying-perception-planning-and-safety-with-ai-2\\\/\"},\"author\":{\"name\":\"Kareem Darwish\",\"@id\":\"https:\\\/\\\/scipapermill.com\\\/#\\\/schema\\\/person\\\/2a018968b95abd980774176f3c37d76e\"},\"headline\":\"Autonomous Driving&#8217;s Next Gear: Unifying Perception, Planning, and Safety with AI\",\"datePublished\":\"2025-11-16T10:29:58+00:00\",\"dateModified\":\"2025-12-28T21:20:41+00:00\",\"mainEntityOfPage\":{\"@id\":\"https:\\\/\\\/scipapermill.com\\\/index.php\\\/2025\\\/11\\\/16\\\/autonomous-drivings-next-gear-unifying-perception-planning-and-safety-with-ai-2\\\/\"},\"wordCount\":1377,\"commentCount\":0,\"publisher\":{\"@id\":\"https:\\\/\\\/scipapermill.com\\\/#organization\"},\"keywords\":[\"autonomous driving\",\"autonomous driving\",\"end-to-end autonomous driving\",\"knowledge distillation\",\"large language models (llms)\",\"trajectory prediction\"],\"articleSection\":[\"Artificial Intelligence\",\"Computer Vision\",\"Robotics\"],\"inLanguage\":\"en-US\",\"potentialAction\":[{\"@type\":\"CommentAction\",\"name\":\"Comment\",\"target\":[\"https:\\\/\\\/scipapermill.com\\\/index.php\\\/2025\\\/11\\\/16\\\/autonomous-drivings-next-gear-unifying-perception-planning-and-safety-with-ai-2\\\/#respond\"]}]},{\"@type\":\"WebPage\",\"@id\":\"https:\\\/\\\/scipapermill.com\\\/index.php\\\/2025\\\/11\\\/16\\\/autonomous-drivings-next-gear-unifying-perception-planning-and-safety-with-ai-2\\\/\",\"url\":\"https:\\\/\\\/scipapermill.com\\\/index.php\\\/2025\\\/11\\\/16\\\/autonomous-drivings-next-gear-unifying-perception-planning-and-safety-with-ai-2\\\/\",\"name\":\"Autonomous Driving's Next Gear: Unifying Perception, Planning, and Safety with AI\",\"isPartOf\":{\"@id\":\"https:\\\/\\\/scipapermill.com\\\/#website\"},\"datePublished\":\"2025-11-16T10:29:58+00:00\",\"dateModified\":\"2025-12-28T21:20:41+00:00\",\"description\":\"Latest 50 papers on autonomous driving: Nov. 16, 2025\",\"breadcrumb\":{\"@id\":\"https:\\\/\\\/scipapermill.com\\\/index.php\\\/2025\\\/11\\\/16\\\/autonomous-drivings-next-gear-unifying-perception-planning-and-safety-with-ai-2\\\/#breadcrumb\"},\"inLanguage\":\"en-US\",\"potentialAction\":[{\"@type\":\"ReadAction\",\"target\":[\"https:\\\/\\\/scipapermill.com\\\/index.php\\\/2025\\\/11\\\/16\\\/autonomous-drivings-next-gear-unifying-perception-planning-and-safety-with-ai-2\\\/\"]}]},{\"@type\":\"BreadcrumbList\",\"@id\":\"https:\\\/\\\/scipapermill.com\\\/index.php\\\/2025\\\/11\\\/16\\\/autonomous-drivings-next-gear-unifying-perception-planning-and-safety-with-ai-2\\\/#breadcrumb\",\"itemListElement\":[{\"@type\":\"ListItem\",\"position\":1,\"name\":\"Home\",\"item\":\"https:\\\/\\\/scipapermill.com\\\/\"},{\"@type\":\"ListItem\",\"position\":2,\"name\":\"Autonomous Driving&#8217;s Next Gear: Unifying Perception, Planning, and Safety with AI\"}]},{\"@type\":\"WebSite\",\"@id\":\"https:\\\/\\\/scipapermill.com\\\/#website\",\"url\":\"https:\\\/\\\/scipapermill.com\\\/\",\"name\":\"SciPapermill\",\"description\":\"Follow the latest research\",\"publisher\":{\"@id\":\"https:\\\/\\\/scipapermill.com\\\/#organization\"},\"potentialAction\":[{\"@type\":\"SearchAction\",\"target\":{\"@type\":\"EntryPoint\",\"urlTemplate\":\"https:\\\/\\\/scipapermill.com\\\/?s={search_term_string}\"},\"query-input\":{\"@type\":\"PropertyValueSpecification\",\"valueRequired\":true,\"valueName\":\"search_term_string\"}}],\"inLanguage\":\"en-US\"},{\"@type\":\"Organization\",\"@id\":\"https:\\\/\\\/scipapermill.com\\\/#organization\",\"name\":\"SciPapermill\",\"url\":\"https:\\\/\\\/scipapermill.com\\\/\",\"logo\":{\"@type\":\"ImageObject\",\"inLanguage\":\"en-US\",\"@id\":\"https:\\\/\\\/scipapermill.com\\\/#\\\/schema\\\/logo\\\/image\\\/\",\"url\":\"https:\\\/\\\/i0.wp.com\\\/scipapermill.com\\\/wp-content\\\/uploads\\\/2025\\\/07\\\/cropped-icon.jpg?fit=512%2C512&ssl=1\",\"contentUrl\":\"https:\\\/\\\/i0.wp.com\\\/scipapermill.com\\\/wp-content\\\/uploads\\\/2025\\\/07\\\/cropped-icon.jpg?fit=512%2C512&ssl=1\",\"width\":512,\"height\":512,\"caption\":\"SciPapermill\"},\"image\":{\"@id\":\"https:\\\/\\\/scipapermill.com\\\/#\\\/schema\\\/logo\\\/image\\\/\"},\"sameAs\":[\"https:\\\/\\\/www.facebook.com\\\/people\\\/SciPapermill\\\/61582731431910\\\/\",\"https:\\\/\\\/www.linkedin.com\\\/company\\\/scipapermill\\\/\"]},{\"@type\":\"Person\",\"@id\":\"https:\\\/\\\/scipapermill.com\\\/#\\\/schema\\\/person\\\/2a018968b95abd980774176f3c37d76e\",\"name\":\"Kareem Darwish\",\"image\":{\"@type\":\"ImageObject\",\"inLanguage\":\"en-US\",\"@id\":\"https:\\\/\\\/secure.gravatar.com\\\/avatar\\\/5fc627e90b8f3d4e8d6eac1f6f00a2fae2dc0cd66b5e44faff7e38e3f85d3dff?s=96&d=mm&r=g\",\"url\":\"https:\\\/\\\/secure.gravatar.com\\\/avatar\\\/5fc627e90b8f3d4e8d6eac1f6f00a2fae2dc0cd66b5e44faff7e38e3f85d3dff?s=96&d=mm&r=g\",\"contentUrl\":\"https:\\\/\\\/secure.gravatar.com\\\/avatar\\\/5fc627e90b8f3d4e8d6eac1f6f00a2fae2dc0cd66b5e44faff7e38e3f85d3dff?s=96&d=mm&r=g\",\"caption\":\"Kareem Darwish\"},\"description\":\"The SciPapermill bot is an AI research assistant dedicated to curating the latest advancements in artificial intelligence. Every week, it meticulously scans and synthesizes newly published papers, distilling key insights into a concise digest. Its mission is to keep you informed on the most significant take-home messages, emerging models, and pivotal datasets that are shaping the future of AI. This bot was created by Dr. Kareem Darwish, who is a principal scientist at the Qatar Computing Research Institute (QCRI) and is working on state-of-the-art Arabic large language models.\",\"sameAs\":[\"https:\\\/\\\/scipapermill.com\"]}]}<\/script>\n<!-- \/ Yoast SEO plugin. -->","yoast_head_json":{"title":"Autonomous Driving's Next Gear: Unifying Perception, Planning, and Safety with AI","description":"Latest 50 papers on autonomous driving: Nov. 16, 2025","robots":{"index":"index","follow":"follow","max-snippet":"max-snippet:-1","max-image-preview":"max-image-preview:large","max-video-preview":"max-video-preview:-1"},"canonical":"https:\/\/scipapermill.com\/index.php\/2025\/11\/16\/autonomous-drivings-next-gear-unifying-perception-planning-and-safety-with-ai-2\/","og_locale":"en_US","og_type":"article","og_title":"Autonomous Driving's Next Gear: Unifying Perception, Planning, and Safety with AI","og_description":"Latest 50 papers on autonomous driving: Nov. 16, 2025","og_url":"https:\/\/scipapermill.com\/index.php\/2025\/11\/16\/autonomous-drivings-next-gear-unifying-perception-planning-and-safety-with-ai-2\/","og_site_name":"SciPapermill","article_publisher":"https:\/\/www.facebook.com\/people\/SciPapermill\/61582731431910\/","article_published_time":"2025-11-16T10:29:58+00:00","article_modified_time":"2025-12-28T21:20:41+00:00","og_image":[{"width":512,"height":512,"url":"https:\/\/i0.wp.com\/scipapermill.com\/wp-content\/uploads\/2025\/07\/cropped-icon.jpg?fit=512%2C512&ssl=1","type":"image\/jpeg"}],"author":"Kareem Darwish","twitter_card":"summary_large_image","twitter_misc":{"Written by":"Kareem Darwish","Est. reading time":"7 minutes"},"schema":{"@context":"https:\/\/schema.org","@graph":[{"@type":"Article","@id":"https:\/\/scipapermill.com\/index.php\/2025\/11\/16\/autonomous-drivings-next-gear-unifying-perception-planning-and-safety-with-ai-2\/#article","isPartOf":{"@id":"https:\/\/scipapermill.com\/index.php\/2025\/11\/16\/autonomous-drivings-next-gear-unifying-perception-planning-and-safety-with-ai-2\/"},"author":{"name":"Kareem Darwish","@id":"https:\/\/scipapermill.com\/#\/schema\/person\/2a018968b95abd980774176f3c37d76e"},"headline":"Autonomous Driving&#8217;s Next Gear: Unifying Perception, Planning, and Safety with AI","datePublished":"2025-11-16T10:29:58+00:00","dateModified":"2025-12-28T21:20:41+00:00","mainEntityOfPage":{"@id":"https:\/\/scipapermill.com\/index.php\/2025\/11\/16\/autonomous-drivings-next-gear-unifying-perception-planning-and-safety-with-ai-2\/"},"wordCount":1377,"commentCount":0,"publisher":{"@id":"https:\/\/scipapermill.com\/#organization"},"keywords":["autonomous driving","autonomous driving","end-to-end autonomous driving","knowledge distillation","large language models (llms)","trajectory prediction"],"articleSection":["Artificial Intelligence","Computer Vision","Robotics"],"inLanguage":"en-US","potentialAction":[{"@type":"CommentAction","name":"Comment","target":["https:\/\/scipapermill.com\/index.php\/2025\/11\/16\/autonomous-drivings-next-gear-unifying-perception-planning-and-safety-with-ai-2\/#respond"]}]},{"@type":"WebPage","@id":"https:\/\/scipapermill.com\/index.php\/2025\/11\/16\/autonomous-drivings-next-gear-unifying-perception-planning-and-safety-with-ai-2\/","url":"https:\/\/scipapermill.com\/index.php\/2025\/11\/16\/autonomous-drivings-next-gear-unifying-perception-planning-and-safety-with-ai-2\/","name":"Autonomous Driving's Next Gear: Unifying Perception, Planning, and Safety with AI","isPartOf":{"@id":"https:\/\/scipapermill.com\/#website"},"datePublished":"2025-11-16T10:29:58+00:00","dateModified":"2025-12-28T21:20:41+00:00","description":"Latest 50 papers on autonomous driving: Nov. 16, 2025","breadcrumb":{"@id":"https:\/\/scipapermill.com\/index.php\/2025\/11\/16\/autonomous-drivings-next-gear-unifying-perception-planning-and-safety-with-ai-2\/#breadcrumb"},"inLanguage":"en-US","potentialAction":[{"@type":"ReadAction","target":["https:\/\/scipapermill.com\/index.php\/2025\/11\/16\/autonomous-drivings-next-gear-unifying-perception-planning-and-safety-with-ai-2\/"]}]},{"@type":"BreadcrumbList","@id":"https:\/\/scipapermill.com\/index.php\/2025\/11\/16\/autonomous-drivings-next-gear-unifying-perception-planning-and-safety-with-ai-2\/#breadcrumb","itemListElement":[{"@type":"ListItem","position":1,"name":"Home","item":"https:\/\/scipapermill.com\/"},{"@type":"ListItem","position":2,"name":"Autonomous Driving&#8217;s Next Gear: Unifying Perception, Planning, and Safety with AI"}]},{"@type":"WebSite","@id":"https:\/\/scipapermill.com\/#website","url":"https:\/\/scipapermill.com\/","name":"SciPapermill","description":"Follow the latest research","publisher":{"@id":"https:\/\/scipapermill.com\/#organization"},"potentialAction":[{"@type":"SearchAction","target":{"@type":"EntryPoint","urlTemplate":"https:\/\/scipapermill.com\/?s={search_term_string}"},"query-input":{"@type":"PropertyValueSpecification","valueRequired":true,"valueName":"search_term_string"}}],"inLanguage":"en-US"},{"@type":"Organization","@id":"https:\/\/scipapermill.com\/#organization","name":"SciPapermill","url":"https:\/\/scipapermill.com\/","logo":{"@type":"ImageObject","inLanguage":"en-US","@id":"https:\/\/scipapermill.com\/#\/schema\/logo\/image\/","url":"https:\/\/i0.wp.com\/scipapermill.com\/wp-content\/uploads\/2025\/07\/cropped-icon.jpg?fit=512%2C512&ssl=1","contentUrl":"https:\/\/i0.wp.com\/scipapermill.com\/wp-content\/uploads\/2025\/07\/cropped-icon.jpg?fit=512%2C512&ssl=1","width":512,"height":512,"caption":"SciPapermill"},"image":{"@id":"https:\/\/scipapermill.com\/#\/schema\/logo\/image\/"},"sameAs":["https:\/\/www.facebook.com\/people\/SciPapermill\/61582731431910\/","https:\/\/www.linkedin.com\/company\/scipapermill\/"]},{"@type":"Person","@id":"https:\/\/scipapermill.com\/#\/schema\/person\/2a018968b95abd980774176f3c37d76e","name":"Kareem Darwish","image":{"@type":"ImageObject","inLanguage":"en-US","@id":"https:\/\/secure.gravatar.com\/avatar\/5fc627e90b8f3d4e8d6eac1f6f00a2fae2dc0cd66b5e44faff7e38e3f85d3dff?s=96&d=mm&r=g","url":"https:\/\/secure.gravatar.com\/avatar\/5fc627e90b8f3d4e8d6eac1f6f00a2fae2dc0cd66b5e44faff7e38e3f85d3dff?s=96&d=mm&r=g","contentUrl":"https:\/\/secure.gravatar.com\/avatar\/5fc627e90b8f3d4e8d6eac1f6f00a2fae2dc0cd66b5e44faff7e38e3f85d3dff?s=96&d=mm&r=g","caption":"Kareem Darwish"},"description":"The SciPapermill bot is an AI research assistant dedicated to curating the latest advancements in artificial intelligence. Every week, it meticulously scans and synthesizes newly published papers, distilling key insights into a concise digest. Its mission is to keep you informed on the most significant take-home messages, emerging models, and pivotal datasets that are shaping the future of AI. This bot was created by Dr. Kareem Darwish, who is a principal scientist at the Qatar Computing Research Institute (QCRI) and is working on state-of-the-art Arabic large language models.","sameAs":["https:\/\/scipapermill.com"]}]}},"views":34,"jetpack_publicize_connections":[],"jetpack_featured_media_url":"","jetpack_shortlink":"https:\/\/wp.me\/pgIXGY-uq","jetpack_sharing_enabled":true,"_links":{"self":[{"href":"https:\/\/scipapermill.com\/index.php\/wp-json\/wp\/v2\/posts\/1886","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/scipapermill.com\/index.php\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/scipapermill.com\/index.php\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/scipapermill.com\/index.php\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/scipapermill.com\/index.php\/wp-json\/wp\/v2\/comments?post=1886"}],"version-history":[{"count":1,"href":"https:\/\/scipapermill.com\/index.php\/wp-json\/wp\/v2\/posts\/1886\/revisions"}],"predecessor-version":[{"id":3225,"href":"https:\/\/scipapermill.com\/index.php\/wp-json\/wp\/v2\/posts\/1886\/revisions\/3225"}],"wp:attachment":[{"href":"https:\/\/scipapermill.com\/index.php\/wp-json\/wp\/v2\/media?parent=1886"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/scipapermill.com\/index.php\/wp-json\/wp\/v2\/categories?post=1886"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/scipapermill.com\/index.php\/wp-json\/wp\/v2\/tags?post=1886"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}