{"id":6766,"date":"2026-05-02T03:23:55","date_gmt":"2026-05-02T03:23:55","guid":{"rendered":"https:\/\/scipapermill.com\/index.php\/2026\/05\/02\/ocrs-next-chapter-from-low-light-to-unfaithful-math-transcriptions\/"},"modified":"2026-05-02T03:23:55","modified_gmt":"2026-05-02T03:23:55","slug":"ocrs-next-chapter-from-low-light-to-unfaithful-math-transcriptions","status":"publish","type":"post","link":"https:\/\/scipapermill.com\/index.php\/2026\/05\/02\/ocrs-next-chapter-from-low-light-to-unfaithful-math-transcriptions\/","title":{"rendered":"OCR&#8217;s Next Chapter: From Low-Light to Unfaithful Math Transcriptions"},"content":{"rendered":"<h3>Latest 3 papers on optical character recognition: May. 2, 2026<\/h3>\n<p>Optical Character Recognition (OCR) has long been a cornerstone of digital transformation, allowing us to bridge the gap between physical and digital text. Yet, as we push the boundaries of AI\/ML, OCR faces increasingly complex, real-world challenges \u2013 from deciphering text in adverse conditions to accurately interpreting nuanced content like handwritten math. Recent research is not only tackling these hurdles but also revealing new, critical considerations for how we evaluate and develop future OCR systems.<\/p>\n<h3 id=\"the-big-ideas-core-innovations\">The Big Idea(s) &amp; Core Innovations<\/h3>\n<p>These recent breakthroughs highlight a pivotal shift: moving beyond mere accuracy to embrace robustness, contextual intelligence, and faithfulness. A prime example is the work presented by <strong>Vijaysinh Gaikwad<\/strong> from <strong>JP Research India Pvt. Ltd.<\/strong> in their paper, <a href=\"https:\/\/arxiv.org\/pdf\/2604.25176\">Benchmarking OCR Pipelines with Adaptive Enhancement for Multi-Domain Retail Bill Digitization<\/a>. This research addresses the messy reality of retail documents by proposing an intelligent, quality-aware adaptive OCR pipeline. Their core innovation lies in integrating CNN-based image enhancement with self-supervised denoising, Laplacian variance-based quality analysis for three-tier routing, and confidence-driven feedback loops. This adaptive approach significantly improves accuracy (26.4% CER improvement over baseline) while boosting efficiency, showcasing that smart resource allocation through quality assessment can optimize processing.<\/p>\n<p>Complementing this focus on robustness, a team from the <strong>Computer Vision Center, Barcelona, Spain<\/strong>, including <strong>Xuanshuo Fu, Lei Kang, and Javier Vazquez-Corral<\/strong>, tackles the formidable challenge of <a href=\"https:\/\/arxiv.org\/pdf\/2604.23685\">Reading in the Dark: Low-light Scene Text Recognition<\/a>. They introduce RLLIE (Re-render Low-light Image Enhancement), an end-to-end module that cleverly combines physics-inspired Image-Based Lighting and Precomputed Radiance Transfer with OCR. Their key insight? Brighter isn\u2019t always better. Joint training of enhancement and recognition modules, especially with task-oriented optimization, outperforms standalone or frozen combinations, leading to substantial improvements in low-light conditions by preserving critical text structures.<\/p>\n<p>Perhaps the most thought-provoking advancement comes from <strong>Jin Seong<\/strong> and colleagues at the <strong>Electronics and Telecommunications Research Institute, Republic of Korea<\/strong>, in their paper, <a href=\"https:\/\/arxiv.org\/pdf\/2604.22774\">When VLMs \u2018Fix\u2019 Students: Identifying and Penalizing Over-Correction in the Evaluation of Multi-line Handwritten Math OCR<\/a>. This groundbreaking work identifies \u2018over-correction\u2019 as a pervasive and critical failure mode in Vision-Language Models (VLMs) when transcribing handwritten math. VLMs, particularly larger models, often \u2018fix\u2019 student errors rather than faithfully reproducing them. They propose PINK (Penalized INK-based score), a novel semantic evaluation metric that explicitly penalizes this behavior, revealing a hidden flaw in even state-of-the-art models and emphasizing the crucial need for <strong>faithfulness<\/strong> in critical applications like educational assessment. Their findings indicate that stronger models actually over-correct more frequently, an emergent property tied to advanced reasoning capabilities overriding visual evidence.<\/p>\n<h3 id=\"under-the-hood-models-datasets-benchmarks\">Under the Hood: Models, Datasets, &amp; Benchmarks<\/h3>\n<p>The innovations discussed are heavily reliant on tailored datasets, robust models, and specialized benchmarks:<\/p>\n<ul>\n<li><strong>Adaptive Retail OCR Pipeline:<\/strong> This work, from Gaikwad, utilizes standard tools like Python 3.9, TensorFlow 2.x, OpenCV 4.x, Tesseract OCR 5.0, and EasyOCR 1.7. It benchmarks against a real-world 360-image multi-domain retail bill dataset and also contributes an OCR ensemble majority voting strategy for generating credible pseudo ground truth.<\/li>\n<li><strong>Low-Light Scene Text Recognition:<\/strong> Fu, Kang, and Vazquez-Corral et al.\u00a0introduce two crucial resources:\n<ul>\n<li><strong>LSTR dataset<\/strong>: 11,273 synthetically generated low-light images derived from existing datasets (ICDAR2015, IIIT5K, WordArt).<\/li>\n<li><strong>ESTR dataset<\/strong>: 60 real nighttime street images in English and Spanish for robust evaluation. Their RLLIE module is designed to work with OCR models like TrOCR, with LoRA-based adaptation proving effective for fine-tuning on smaller datasets. The LSTR dataset is publicly available on <a href=\"https:\/\/huggingface.co\/datasets\/lumimusta\/Low-light_Scene_Text_Dataset\">Hugging Face<\/a>.<\/li>\n<\/ul>\n<\/li>\n<li><strong>Handwritten Math OCR &amp; Over-Correction:<\/strong> Seong et al.\u2019s research leverages the <strong>FERMAT dataset<\/strong> (Nath et al., 2025) and comprehensively evaluates 15 state-of-the-art VLMs. They plan to release a Qwen3-80B grading toolkit and their complete evaluation codebase, enabling the community to reproduce and build upon their findings.<\/li>\n<\/ul>\n<h3 id=\"impact-the-road-ahead\">Impact &amp; The Road Ahead<\/h3>\n<p>These advancements herald a more sophisticated era for OCR. Gaikwad\u2019s adaptive pipeline offers immediate practical benefits for industries dealing with diverse, low-quality documents, significantly streamlining digitization workflows. The low-light text recognition research from Fu, Kang, and Vazquez-Corral et al.\u00a0opens doors for more reliable outdoor signage reading, autonomous driving, and security applications under challenging lighting conditions. The key takeaway here is the importance of <strong>task-oriented image enhancement<\/strong>, a concept that will undoubtedly influence future computer vision pipelines.<\/p>\n<p>Perhaps the most profound implications arise from the discovery of \u2018over-correction\u2019 in VLMs by Seong et al.\u00a0This highlights a critical need to re-evaluate how we measure the <strong>faithfulness<\/strong> of AI systems, particularly in sensitive domains like education where accurate assessment of human work is paramount. The PINK metric could become a standard for assessing VLM fidelity in generative tasks, pushing developers to create models that are not just intelligent but also <strong>truthful<\/strong> to the input. The correlation between model scale and over-correction suggests an emergent property of advanced AI, posing a fundamental challenge that may require architectural interventions beyond simple prompt engineering. The road ahead for OCR and VLMs involves not just improving accuracy, but fostering systems that are robust, contextually aware, and unfailingly faithful to their inputs, pushing the boundaries of what these technologies can reliably achieve.<\/p>\n","protected":false},"excerpt":{"rendered":"<p>Latest 3 papers on optical character recognition: May. 2, 2026<\/p>\n","protected":false},"author":1,"featured_media":0,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"_yoast_wpseo_focuskw":"","_yoast_wpseo_title":"","_yoast_wpseo_metadesc":"","_jetpack_memberships_contains_paid_content":false,"footnotes":"","jetpack_publicize_message":"","jetpack_publicize_feature_enabled":true,"jetpack_social_post_already_shared":true,"jetpack_social_options":{"image_generator_settings":{"template":"highway","default_image_id":0,"font":"","enabled":false},"version":2}},"categories":[55,438,63],"tags":[4144,4142,4141,475,1642,4143],"class_list":["post-6766","post","type-post","status-publish","format-standard","hentry","category-computer-vision","category-computers-and-society","category-machine-learning","tag-adaptive-pipeline","tag-cnn-image-enhancement","tag-document-intelligence","tag-optical-character-recognition","tag-main_tag_optical_character_recognition","tag-retail-bill-digitization"],"yoast_head":"<!-- This site is optimized with the Yoast SEO plugin v27.4 - https:\/\/yoast.com\/product\/yoast-seo-wordpress\/ -->\n<title>OCR&#039;s Next Chapter: From Low-Light to Unfaithful Math Transcriptions<\/title>\n<meta name=\"description\" content=\"Latest 3 papers on optical character recognition: May. 2, 2026\" \/>\n<meta name=\"robots\" content=\"index, follow, max-snippet:-1, max-image-preview:large, max-video-preview:-1\" \/>\n<link rel=\"canonical\" href=\"https:\/\/scipapermill.com\/index.php\/2026\/05\/02\/ocrs-next-chapter-from-low-light-to-unfaithful-math-transcriptions\/\" \/>\n<meta property=\"og:locale\" content=\"en_US\" \/>\n<meta property=\"og:type\" content=\"article\" \/>\n<meta property=\"og:title\" content=\"OCR&#039;s Next Chapter: From Low-Light to Unfaithful Math Transcriptions\" \/>\n<meta property=\"og:description\" content=\"Latest 3 papers on optical character recognition: May. 2, 2026\" \/>\n<meta property=\"og:url\" content=\"https:\/\/scipapermill.com\/index.php\/2026\/05\/02\/ocrs-next-chapter-from-low-light-to-unfaithful-math-transcriptions\/\" \/>\n<meta property=\"og:site_name\" content=\"SciPapermill\" \/>\n<meta property=\"article:publisher\" content=\"https:\/\/www.facebook.com\/people\/SciPapermill\/61582731431910\/\" \/>\n<meta property=\"article:published_time\" content=\"2026-05-02T03:23:55+00:00\" \/>\n<meta property=\"og:image\" content=\"https:\/\/i0.wp.com\/scipapermill.com\/wp-content\/uploads\/2025\/07\/cropped-icon.jpg?fit=512%2C512&ssl=1\" \/>\n\t<meta property=\"og:image:width\" content=\"512\" \/>\n\t<meta property=\"og:image:height\" content=\"512\" \/>\n\t<meta property=\"og:image:type\" content=\"image\/jpeg\" \/>\n<meta name=\"author\" content=\"Kareem Darwish\" \/>\n<meta name=\"twitter:card\" content=\"summary_large_image\" \/>\n<meta name=\"twitter:label1\" content=\"Written by\" \/>\n\t<meta name=\"twitter:data1\" content=\"Kareem Darwish\" \/>\n\t<meta name=\"twitter:label2\" content=\"Est. reading time\" \/>\n\t<meta name=\"twitter:data2\" content=\"4 minutes\" \/>\n<script type=\"application\/ld+json\" class=\"yoast-schema-graph\">{\"@context\":\"https:\\\/\\\/schema.org\",\"@graph\":[{\"@type\":\"Article\",\"@id\":\"https:\\\/\\\/scipapermill.com\\\/index.php\\\/2026\\\/05\\\/02\\\/ocrs-next-chapter-from-low-light-to-unfaithful-math-transcriptions\\\/#article\",\"isPartOf\":{\"@id\":\"https:\\\/\\\/scipapermill.com\\\/index.php\\\/2026\\\/05\\\/02\\\/ocrs-next-chapter-from-low-light-to-unfaithful-math-transcriptions\\\/\"},\"author\":{\"name\":\"Kareem Darwish\",\"@id\":\"https:\\\/\\\/scipapermill.com\\\/#\\\/schema\\\/person\\\/2a018968b95abd980774176f3c37d76e\"},\"headline\":\"OCR&#8217;s Next Chapter: From Low-Light to Unfaithful Math Transcriptions\",\"datePublished\":\"2026-05-02T03:23:55+00:00\",\"mainEntityOfPage\":{\"@id\":\"https:\\\/\\\/scipapermill.com\\\/index.php\\\/2026\\\/05\\\/02\\\/ocrs-next-chapter-from-low-light-to-unfaithful-math-transcriptions\\\/\"},\"wordCount\":843,\"commentCount\":0,\"publisher\":{\"@id\":\"https:\\\/\\\/scipapermill.com\\\/#organization\"},\"keywords\":[\"adaptive pipeline\",\"cnn image enhancement\",\"document intelligence\",\"optical character recognition\",\"optical character recognition\",\"retail bill digitization\"],\"articleSection\":[\"Computer Vision\",\"Computers and Society\",\"Machine Learning\"],\"inLanguage\":\"en-US\",\"potentialAction\":[{\"@type\":\"CommentAction\",\"name\":\"Comment\",\"target\":[\"https:\\\/\\\/scipapermill.com\\\/index.php\\\/2026\\\/05\\\/02\\\/ocrs-next-chapter-from-low-light-to-unfaithful-math-transcriptions\\\/#respond\"]}]},{\"@type\":\"WebPage\",\"@id\":\"https:\\\/\\\/scipapermill.com\\\/index.php\\\/2026\\\/05\\\/02\\\/ocrs-next-chapter-from-low-light-to-unfaithful-math-transcriptions\\\/\",\"url\":\"https:\\\/\\\/scipapermill.com\\\/index.php\\\/2026\\\/05\\\/02\\\/ocrs-next-chapter-from-low-light-to-unfaithful-math-transcriptions\\\/\",\"name\":\"OCR's Next Chapter: From Low-Light to Unfaithful Math Transcriptions\",\"isPartOf\":{\"@id\":\"https:\\\/\\\/scipapermill.com\\\/#website\"},\"datePublished\":\"2026-05-02T03:23:55+00:00\",\"description\":\"Latest 3 papers on optical character recognition: May. 2, 2026\",\"breadcrumb\":{\"@id\":\"https:\\\/\\\/scipapermill.com\\\/index.php\\\/2026\\\/05\\\/02\\\/ocrs-next-chapter-from-low-light-to-unfaithful-math-transcriptions\\\/#breadcrumb\"},\"inLanguage\":\"en-US\",\"potentialAction\":[{\"@type\":\"ReadAction\",\"target\":[\"https:\\\/\\\/scipapermill.com\\\/index.php\\\/2026\\\/05\\\/02\\\/ocrs-next-chapter-from-low-light-to-unfaithful-math-transcriptions\\\/\"]}]},{\"@type\":\"BreadcrumbList\",\"@id\":\"https:\\\/\\\/scipapermill.com\\\/index.php\\\/2026\\\/05\\\/02\\\/ocrs-next-chapter-from-low-light-to-unfaithful-math-transcriptions\\\/#breadcrumb\",\"itemListElement\":[{\"@type\":\"ListItem\",\"position\":1,\"name\":\"Home\",\"item\":\"https:\\\/\\\/scipapermill.com\\\/\"},{\"@type\":\"ListItem\",\"position\":2,\"name\":\"OCR&#8217;s Next Chapter: From Low-Light to Unfaithful Math Transcriptions\"}]},{\"@type\":\"WebSite\",\"@id\":\"https:\\\/\\\/scipapermill.com\\\/#website\",\"url\":\"https:\\\/\\\/scipapermill.com\\\/\",\"name\":\"SciPapermill\",\"description\":\"Follow the latest research\",\"publisher\":{\"@id\":\"https:\\\/\\\/scipapermill.com\\\/#organization\"},\"potentialAction\":[{\"@type\":\"SearchAction\",\"target\":{\"@type\":\"EntryPoint\",\"urlTemplate\":\"https:\\\/\\\/scipapermill.com\\\/?s={search_term_string}\"},\"query-input\":{\"@type\":\"PropertyValueSpecification\",\"valueRequired\":true,\"valueName\":\"search_term_string\"}}],\"inLanguage\":\"en-US\"},{\"@type\":\"Organization\",\"@id\":\"https:\\\/\\\/scipapermill.com\\\/#organization\",\"name\":\"SciPapermill\",\"url\":\"https:\\\/\\\/scipapermill.com\\\/\",\"logo\":{\"@type\":\"ImageObject\",\"inLanguage\":\"en-US\",\"@id\":\"https:\\\/\\\/scipapermill.com\\\/#\\\/schema\\\/logo\\\/image\\\/\",\"url\":\"https:\\\/\\\/i0.wp.com\\\/scipapermill.com\\\/wp-content\\\/uploads\\\/2025\\\/07\\\/cropped-icon.jpg?fit=512%2C512&ssl=1\",\"contentUrl\":\"https:\\\/\\\/i0.wp.com\\\/scipapermill.com\\\/wp-content\\\/uploads\\\/2025\\\/07\\\/cropped-icon.jpg?fit=512%2C512&ssl=1\",\"width\":512,\"height\":512,\"caption\":\"SciPapermill\"},\"image\":{\"@id\":\"https:\\\/\\\/scipapermill.com\\\/#\\\/schema\\\/logo\\\/image\\\/\"},\"sameAs\":[\"https:\\\/\\\/www.facebook.com\\\/people\\\/SciPapermill\\\/61582731431910\\\/\",\"https:\\\/\\\/www.linkedin.com\\\/company\\\/scipapermill\\\/\"]},{\"@type\":\"Person\",\"@id\":\"https:\\\/\\\/scipapermill.com\\\/#\\\/schema\\\/person\\\/2a018968b95abd980774176f3c37d76e\",\"name\":\"Kareem Darwish\",\"image\":{\"@type\":\"ImageObject\",\"inLanguage\":\"en-US\",\"@id\":\"https:\\\/\\\/secure.gravatar.com\\\/avatar\\\/5fc627e90b8f3d4e8d6eac1f6f00a2fae2dc0cd66b5e44faff7e38e3f85d3dff?s=96&d=mm&r=g\",\"url\":\"https:\\\/\\\/secure.gravatar.com\\\/avatar\\\/5fc627e90b8f3d4e8d6eac1f6f00a2fae2dc0cd66b5e44faff7e38e3f85d3dff?s=96&d=mm&r=g\",\"contentUrl\":\"https:\\\/\\\/secure.gravatar.com\\\/avatar\\\/5fc627e90b8f3d4e8d6eac1f6f00a2fae2dc0cd66b5e44faff7e38e3f85d3dff?s=96&d=mm&r=g\",\"caption\":\"Kareem Darwish\"},\"description\":\"The SciPapermill bot is an AI research assistant dedicated to curating the latest advancements in artificial intelligence. Every week, it meticulously scans and synthesizes newly published papers, distilling key insights into a concise digest. Its mission is to keep you informed on the most significant take-home messages, emerging models, and pivotal datasets that are shaping the future of AI. This bot was created by Dr. Kareem Darwish, who is a principal scientist at the Qatar Computing Research Institute (QCRI) and is working on state-of-the-art Arabic large language models.\",\"sameAs\":[\"https:\\\/\\\/scipapermill.com\"]}]}<\/script>\n<!-- \/ Yoast SEO plugin. -->","yoast_head_json":{"title":"OCR's Next Chapter: From Low-Light to Unfaithful Math Transcriptions","description":"Latest 3 papers on optical character recognition: May. 2, 2026","robots":{"index":"index","follow":"follow","max-snippet":"max-snippet:-1","max-image-preview":"max-image-preview:large","max-video-preview":"max-video-preview:-1"},"canonical":"https:\/\/scipapermill.com\/index.php\/2026\/05\/02\/ocrs-next-chapter-from-low-light-to-unfaithful-math-transcriptions\/","og_locale":"en_US","og_type":"article","og_title":"OCR's Next Chapter: From Low-Light to Unfaithful Math Transcriptions","og_description":"Latest 3 papers on optical character recognition: May. 2, 2026","og_url":"https:\/\/scipapermill.com\/index.php\/2026\/05\/02\/ocrs-next-chapter-from-low-light-to-unfaithful-math-transcriptions\/","og_site_name":"SciPapermill","article_publisher":"https:\/\/www.facebook.com\/people\/SciPapermill\/61582731431910\/","article_published_time":"2026-05-02T03:23:55+00:00","og_image":[{"width":512,"height":512,"url":"https:\/\/i0.wp.com\/scipapermill.com\/wp-content\/uploads\/2025\/07\/cropped-icon.jpg?fit=512%2C512&ssl=1","type":"image\/jpeg"}],"author":"Kareem Darwish","twitter_card":"summary_large_image","twitter_misc":{"Written by":"Kareem Darwish","Est. reading time":"4 minutes"},"schema":{"@context":"https:\/\/schema.org","@graph":[{"@type":"Article","@id":"https:\/\/scipapermill.com\/index.php\/2026\/05\/02\/ocrs-next-chapter-from-low-light-to-unfaithful-math-transcriptions\/#article","isPartOf":{"@id":"https:\/\/scipapermill.com\/index.php\/2026\/05\/02\/ocrs-next-chapter-from-low-light-to-unfaithful-math-transcriptions\/"},"author":{"name":"Kareem Darwish","@id":"https:\/\/scipapermill.com\/#\/schema\/person\/2a018968b95abd980774176f3c37d76e"},"headline":"OCR&#8217;s Next Chapter: From Low-Light to Unfaithful Math Transcriptions","datePublished":"2026-05-02T03:23:55+00:00","mainEntityOfPage":{"@id":"https:\/\/scipapermill.com\/index.php\/2026\/05\/02\/ocrs-next-chapter-from-low-light-to-unfaithful-math-transcriptions\/"},"wordCount":843,"commentCount":0,"publisher":{"@id":"https:\/\/scipapermill.com\/#organization"},"keywords":["adaptive pipeline","cnn image enhancement","document intelligence","optical character recognition","optical character recognition","retail bill digitization"],"articleSection":["Computer Vision","Computers and Society","Machine Learning"],"inLanguage":"en-US","potentialAction":[{"@type":"CommentAction","name":"Comment","target":["https:\/\/scipapermill.com\/index.php\/2026\/05\/02\/ocrs-next-chapter-from-low-light-to-unfaithful-math-transcriptions\/#respond"]}]},{"@type":"WebPage","@id":"https:\/\/scipapermill.com\/index.php\/2026\/05\/02\/ocrs-next-chapter-from-low-light-to-unfaithful-math-transcriptions\/","url":"https:\/\/scipapermill.com\/index.php\/2026\/05\/02\/ocrs-next-chapter-from-low-light-to-unfaithful-math-transcriptions\/","name":"OCR's Next Chapter: From Low-Light to Unfaithful Math Transcriptions","isPartOf":{"@id":"https:\/\/scipapermill.com\/#website"},"datePublished":"2026-05-02T03:23:55+00:00","description":"Latest 3 papers on optical character recognition: May. 2, 2026","breadcrumb":{"@id":"https:\/\/scipapermill.com\/index.php\/2026\/05\/02\/ocrs-next-chapter-from-low-light-to-unfaithful-math-transcriptions\/#breadcrumb"},"inLanguage":"en-US","potentialAction":[{"@type":"ReadAction","target":["https:\/\/scipapermill.com\/index.php\/2026\/05\/02\/ocrs-next-chapter-from-low-light-to-unfaithful-math-transcriptions\/"]}]},{"@type":"BreadcrumbList","@id":"https:\/\/scipapermill.com\/index.php\/2026\/05\/02\/ocrs-next-chapter-from-low-light-to-unfaithful-math-transcriptions\/#breadcrumb","itemListElement":[{"@type":"ListItem","position":1,"name":"Home","item":"https:\/\/scipapermill.com\/"},{"@type":"ListItem","position":2,"name":"OCR&#8217;s Next Chapter: From Low-Light to Unfaithful Math Transcriptions"}]},{"@type":"WebSite","@id":"https:\/\/scipapermill.com\/#website","url":"https:\/\/scipapermill.com\/","name":"SciPapermill","description":"Follow the latest research","publisher":{"@id":"https:\/\/scipapermill.com\/#organization"},"potentialAction":[{"@type":"SearchAction","target":{"@type":"EntryPoint","urlTemplate":"https:\/\/scipapermill.com\/?s={search_term_string}"},"query-input":{"@type":"PropertyValueSpecification","valueRequired":true,"valueName":"search_term_string"}}],"inLanguage":"en-US"},{"@type":"Organization","@id":"https:\/\/scipapermill.com\/#organization","name":"SciPapermill","url":"https:\/\/scipapermill.com\/","logo":{"@type":"ImageObject","inLanguage":"en-US","@id":"https:\/\/scipapermill.com\/#\/schema\/logo\/image\/","url":"https:\/\/i0.wp.com\/scipapermill.com\/wp-content\/uploads\/2025\/07\/cropped-icon.jpg?fit=512%2C512&ssl=1","contentUrl":"https:\/\/i0.wp.com\/scipapermill.com\/wp-content\/uploads\/2025\/07\/cropped-icon.jpg?fit=512%2C512&ssl=1","width":512,"height":512,"caption":"SciPapermill"},"image":{"@id":"https:\/\/scipapermill.com\/#\/schema\/logo\/image\/"},"sameAs":["https:\/\/www.facebook.com\/people\/SciPapermill\/61582731431910\/","https:\/\/www.linkedin.com\/company\/scipapermill\/"]},{"@type":"Person","@id":"https:\/\/scipapermill.com\/#\/schema\/person\/2a018968b95abd980774176f3c37d76e","name":"Kareem Darwish","image":{"@type":"ImageObject","inLanguage":"en-US","@id":"https:\/\/secure.gravatar.com\/avatar\/5fc627e90b8f3d4e8d6eac1f6f00a2fae2dc0cd66b5e44faff7e38e3f85d3dff?s=96&d=mm&r=g","url":"https:\/\/secure.gravatar.com\/avatar\/5fc627e90b8f3d4e8d6eac1f6f00a2fae2dc0cd66b5e44faff7e38e3f85d3dff?s=96&d=mm&r=g","contentUrl":"https:\/\/secure.gravatar.com\/avatar\/5fc627e90b8f3d4e8d6eac1f6f00a2fae2dc0cd66b5e44faff7e38e3f85d3dff?s=96&d=mm&r=g","caption":"Kareem Darwish"},"description":"The SciPapermill bot is an AI research assistant dedicated to curating the latest advancements in artificial intelligence. Every week, it meticulously scans and synthesizes newly published papers, distilling key insights into a concise digest. Its mission is to keep you informed on the most significant take-home messages, emerging models, and pivotal datasets that are shaping the future of AI. This bot was created by Dr. Kareem Darwish, who is a principal scientist at the Qatar Computing Research Institute (QCRI) and is working on state-of-the-art Arabic large language models.","sameAs":["https:\/\/scipapermill.com"]}]}},"views":7,"jetpack_publicize_connections":[],"jetpack_featured_media_url":"","jetpack_shortlink":"https:\/\/wp.me\/pgIXGY-1L8","jetpack_sharing_enabled":true,"_links":{"self":[{"href":"https:\/\/scipapermill.com\/index.php\/wp-json\/wp\/v2\/posts\/6766","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/scipapermill.com\/index.php\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/scipapermill.com\/index.php\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/scipapermill.com\/index.php\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/scipapermill.com\/index.php\/wp-json\/wp\/v2\/comments?post=6766"}],"version-history":[{"count":0,"href":"https:\/\/scipapermill.com\/index.php\/wp-json\/wp\/v2\/posts\/6766\/revisions"}],"wp:attachment":[{"href":"https:\/\/scipapermill.com\/index.php\/wp-json\/wp\/v2\/media?parent=6766"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/scipapermill.com\/index.php\/wp-json\/wp\/v2\/categories?post=6766"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/scipapermill.com\/index.php\/wp-json\/wp\/v2\/tags?post=6766"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}