<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:googleplay="http://www.google.com/schemas/play-podcasts/1.0"><channel><title><![CDATA[AI PaperTrails]]></title><description><![CDATA[AI PaperTrails aims to make AI research accessible. State of research from top conferences (NeurIPS, ICLR, COLM, & others), the evolution of techniques, and future outlook for AI research.]]></description><link>https://www.aipapertrails.com</link><image><url>https://substackcdn.com/image/fetch/$s_!3UqA!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F880c3f3d-5940-423d-9e6e-a4e0af266374_1024x1024.png</url><title>AI PaperTrails</title><link>https://www.aipapertrails.com</link></image><generator>Substack</generator><lastBuildDate>Fri, 17 Apr 2026 05:55:33 GMT</lastBuildDate><atom:link href="https://www.aipapertrails.com/feed" rel="self" type="application/rss+xml"/><copyright><![CDATA[Prakash Kagitha]]></copyright><language><![CDATA[en]]></language><webMaster><![CDATA[aipapertrails@substack.com]]></webMaster><itunes:owner><itunes:email><![CDATA[aipapertrails@substack.com]]></itunes:email><itunes:name><![CDATA[Prakash Kagitha]]></itunes:name></itunes:owner><itunes:author><![CDATA[Prakash Kagitha]]></itunes:author><googleplay:owner><![CDATA[aipapertrails@substack.com]]></googleplay:owner><googleplay:email><![CDATA[aipapertrails@substack.com]]></googleplay:email><googleplay:author><![CDATA[Prakash Kagitha]]></googleplay:author><itunes:block><![CDATA[Yes]]></itunes:block><item><title><![CDATA[The frontier of LLMs @ NeurIPS '25]]></title><description><![CDATA[5 research themes and 70+ papers to understand the jagged line of what's now and what's next]]></description><link>https://www.aipapertrails.com/p/the-frontier-of-llms-neurips-25</link><guid isPermaLink="false">https://www.aipapertrails.com/p/the-frontier-of-llms-neurips-25</guid><dc:creator><![CDATA[Prakash Kagitha]]></dc:creator><pubDate>Sun, 30 Nov 2025 15:59:12 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!iweC!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F012152fd-551c-45d2-ab96-3c0aa8fd67e6_1536x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!iweC!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F012152fd-551c-45d2-ab96-3c0aa8fd67e6_1536x1024.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!iweC!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F012152fd-551c-45d2-ab96-3c0aa8fd67e6_1536x1024.png 424w, https://substackcdn.com/image/fetch/$s_!iweC!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F012152fd-551c-45d2-ab96-3c0aa8fd67e6_1536x1024.png 848w, https://substackcdn.com/image/fetch/$s_!iweC!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F012152fd-551c-45d2-ab96-3c0aa8fd67e6_1536x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!iweC!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F012152fd-551c-45d2-ab96-3c0aa8fd67e6_1536x1024.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!iweC!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F012152fd-551c-45d2-ab96-3c0aa8fd67e6_1536x1024.png" width="1456" height="971" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/012152fd-551c-45d2-ab96-3c0aa8fd67e6_1536x1024.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:971,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:2920171,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.aipapertrails.com/i/179893774?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F012152fd-551c-45d2-ab96-3c0aa8fd67e6_1536x1024.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!iweC!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F012152fd-551c-45d2-ab96-3c0aa8fd67e6_1536x1024.png 424w, https://substackcdn.com/image/fetch/$s_!iweC!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F012152fd-551c-45d2-ab96-3c0aa8fd67e6_1536x1024.png 848w, https://substackcdn.com/image/fetch/$s_!iweC!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F012152fd-551c-45d2-ab96-3c0aa8fd67e6_1536x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!iweC!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F012152fd-551c-45d2-ab96-3c0aa8fd67e6_1536x1024.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>We are increasingly talking about AGI/ASI<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a> as LLM systems achieve breakthroughs in <a href="https://deepmind.google/blog/advanced-version-of-gemini-with-deep-think-officially-achieves-gold-medal-standard-at-the-international-mathematical-olympiad/">mathematics</a>, <a href="https://x.com/antigravity/status/1994106813491630543?s=20">coding</a>, and even <a href="https://edisonscientific.com/articles/announcing-kosmos">scientific discovery</a>. However, the techniques underlying these breakthroughs, the current state of LLM capabilities, and the diverse research directions that push the frontier remain relatively inaccessible.</p><p>In this blog post, I try to surface interesting research to shed light on the frontier of LLMs. I do that with the latest peer-reviewed research published at NeurIPS &#8216;25. Around ~6000 papers are published at this conference on virtually every topic in ML. You can use <a href="https://neurips.aipapertrails.com">Paper Finder</a> to find specific NeurIPS papers by topic and poster session, along with their posters/slides.</p><p>These are the topics I explore:</p><ol><li><p>Scaling reasoning and RL environments</p></li><li><p>Memory &amp; long-context LLMs</p></li><li><p>Self-improving LLM systems</p></li><li><p>Computer use &amp; web Agents</p></li><li><p>AI-driven scientific discovery</p></li></ol><h3>1. Scaling reasoning and RL environments</h3><p>While the recipe of <a href="https://arxiv.org/abs/2501.12948">DeepSeek-R1</a> is prominent for eliciting reasoning, <a href="https://neurips.aipapertrails.com/paper/4OsgYD7em5">the most cited paper</a> at the conference observed that RL doesn&#8217;t really incentivize reasoning capacity beyond the base model. It turns out the performance gains from RL could be recovered if one samples ~1000 responses from the base model. The authors, in turn, call for scaling RL data &amp; compute, and for agents to use tools &amp; external info, among other directions that improve reasoning with RL.</p><p>It just so happens that other papers at the conference showed improvement exactly in these directions. <a href="https://neurips.aipapertrails.com/paper/GqYSunGmp7">Reasoning Gym</a> created 100+ RL reasoning environments with verifiable rewards to do RL for reasoning. <a href="https://neurips.aipapertrails.com/paper/xUBgfvyip3">This paper</a> created a cross-domain RL-for-reasoning dataset and showed the efficacy of a mixed-domain RL training approach.</p><p>There were other dimensions, RL for reasoning is improved: <a href="https://neurips.aipapertrails.com/paper/NFM8F5cV0V">Open-Reasoner-Zero</a>, with an alternative to GRPO, showed improvements compared to DeepSeek-R1-Zero. <a href="https://neurips.aipapertrails.com/paper/YPsJha5HXQ">ProRL</a> proposed a variant of GRPO with KL divergence control. <a href="https://neurips.aipapertrails.com/paper/un1TRwNgiv">Thinking vs Doing</a> improves agent reasoning by scaling environment interaction steps. <a href="https://neurips.aipapertrails.com/paper/pBFVoll8Xa">General-Reasoner</a> elicited reasoning in non-verifiable domains with an LLM-based verifier. <a href="https://neurips.aipapertrails.com/paper/NhAi1w3s8Z">This</a> and <a href="https://neurips.aipapertrails.com/paper/AiZxn84Wdo">this</a> work investigated learning token-efficient reasoning, while, amusingly, <a href="https://neurips.aipapertrails.com/paper/IBrRNLr6JA">this paper</a> trained models for reasoning with just one training example.</p><p>Similar recipe of eliciting reasoning is adapted for many more tasks, including long-horizon reasoning in generative reward models (<a href="https://neurips.aipapertrails.com/paper/UfQAFbP6xq">Think-RM</a>), <a href="https://neurips.aipapertrails.com/paper/kXieirlPjF">tool-integrated reasoning</a>, reasoning on <a href="https://neurips.aipapertrails.com/paper/TxedB8hI5O">long video</a>, omnimodal reasoning (<a href="https://neurips.aipapertrails.com/paper/7Q1ApHpX31">Omni-R1</a>), vision-language model reasoning (<a href="https://neurips.aipapertrails.com/paper/4oYxzssbVg">VL-Rethinker</a>), reasoning for real-world software engineering (<a href="https://neurips.aipapertrails.com/paper/ULblO61XZ0">SWE-RL</a>), visual perception (<a href="https://neurips.aipapertrails.com/paper/BeXcXrXetA">Perception-R1</a>), SQL (<a href="https://neurips.aipapertrails.com/paper/hgJQcuDwm1">SQL-R1</a>), and embodied reasoning (<a href="https://neurips.aipapertrails.com/paper/N2bLuwofZ0">Robot-R1</a>).</p><h3>2. Self-improving LLMs</h3><p>A lot of research showed self-improvement on several tasks. The recipe usually goes like this: <br>1. LLM generates a full or partial response <br>2. LLM self-corrects or self-verifies itself (also could be rewards or preferences) <br>3. Train the LLM on positive data or preference learning on positive and negative data.</p><p>At NeurIPS, there are many papers showing self-improvement. Self-improving <a href="https://neurips.aipapertrails.com/paper/KXMIIVUB9U">embodied foundations models</a> by learning with steps-to-go prediction objectives (self-verify). <a href="https://neurips.aipapertrails.com/paper/Awl3ZhDHuQ">Guided-ReST</a> self-improves with guided, reinforced self-training. (training on positive predicted examples). <a href="https://neurips.aipapertrails.com/paper/JsNUE84Hxi">Self-adapting LLMs</a> create data and train themselves based on new input. <a href="https://neurips.aipapertrails.com/paper/9yusqX9DpR">Self-challenging LLM agents</a> that generate Code-as-Task data and a verification function, and train themselves on correct samples. <a href="https://neurips.aipapertrails.com/paper/JddJvNSiHk">SPC</a> fine-tunes a sneaky generator and critic that enables improvement with self-play. <a href="https://neurips.aipapertrails.com/paper/aLGgz4SOyu">Sherlock</a> self-corrects its visual reasoning and improves. <a href="https://neurips.aipapertrails.com/paper/HL1j92hb6z">MM-UPT</a> defines a continual self-improvement loop with a self-rewarding mechanism based on majority voting. <a href="https://neurips.aipapertrails.com/paper/D1PeGJtVEu">ExPO</a> unlocks hard reasoning with self-explanations. <a href="https://neurips.aipapertrails.com/paper/D1PeGJtVEu">SPRO</a> uses self-play for improving image generation. <a href="https://neurips.aipapertrails.com/paper/0jQUNQsZra">SwS</a> is even self-aware of its weaknesses and synthesizes problems and trains itself. <a href="https://neurips.aipapertrails.com/paper/gfX1nqBKtu">on</a> and <a href="https://neurips.aipapertrails.com/paper/D88yhLFzDR">on</a> and <a href="https://neurips.aipapertrails.com/paper/WdL3O58gde">on</a>.</p><p><strong>If self-improvement is possible, where is the singularity?</strong> Like the most-rated paper at NeurIPS and other related work, there is evidence that RL training or self-improvement only brings LLM performance at k=n to k=1, i.e, the better performance with n responses of LLM to the  performance of one response of an LLM. Intuitively, in the base model, the better performance with finite n responses is not infinite! So there is no singularity yet. However, better self-improvement methods might leap that barrier, only if not for the problem below.</p><p><a href="https://neurips.aipapertrails.com/paper/mGEPbyJ8OT">Feedback Friction</a> paper shows that LLMs consistently show resistance to feedback, even with correct feedback. <strong>This is a clear limitation that stunts self-improvement.</strong></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.aipapertrails.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading AI PaperTrails! Subscribe for free to receive more NeurIPS &#8216;25 content and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><h3>3. Memory &amp; long-context LLMs</h3><p>The positioning for the <a href="https://x.com/polynoamial/status/1991212955250327768?s=20">recently released</a> GPT-5.1-Codex-Max is that it can work on a task for more than 24 hours continuously over millions of tokens, which is more than the context length of the model. This is possible because it &#8220;automatically compacts its session when it approaches its context window limit, giving it a fresh context window. It repeats this process until the task is completed.&#8220;</p><p>Scaling the effective context length and designing &#8220;memory&#8220; architecture is an important line of research. <a href="https://neurips.aipapertrails.com/paper/wtcv48HImz">An investigation</a> at NeurIPS shows that long-context model training improves reasoning for even tasks with short input lengths, showing the importance of long-context capabilities. Several works argue that, despite the perceived importance, the evaluation of long-context abilities is not objective and comprehensive. <a href="https://neurips.aipapertrails.com/paper/vAT2xlaWJY">MemSim</a> uses a Bayesian Relation Network (BRNet) to automatically create a dataset to evaluate LLM hallucinations and the capability of memorizing information from user messages. <a href="https://neurips.aipapertrails.com/paper/atjpGqjG73">LongBioBench</a> with artificial biography aims to comprehensively evaluate the long-context capabilities of text models, while <a href="https://neurips.aipapertrails.com/paper/EPQi0v0OxL">MMLongBench</a> does that for vision-language models.</p><p>There are many types of memory architectures: </p><p>1. <strong>Templatic compression of long context</strong>. <a href="https://neurips.aipapertrails.com/paper/n4V3MSqK77">Agentic plan caching</a> reuses structured plan templates from planning states of agent applications to enable the memory feature. <a href="https://neurips.aipapertrails.com/paper/l3Qq5MU5VX">AdmTree</a> compresses lengthy context into hierarchical memory while summarizing and storing it as leaves in a semantic binary tree.</p><p>2. <strong>Indexing the knowledge snippets that LLM could query</strong>. <a href="https://neurips.aipapertrails.com/paper/FiM0M8gcct">A-Mem</a> creates queryable interconnected knowledge networks with the Zettelkasten method. <a href="https://neurips.aipapertrails.com/paper/mmIAp3cVS0">G-memory</a> organizes memories with a three-tier graph hierarchy and performs bi-directional memory traversal to retrieve different levels. This is specifically to encode prior collaboration experiences of multi-agent systems. </p><p>3. <strong>Memory retrieval based on model activations</strong>. <a href="https://neurips.aipapertrails.com/paper/1PvMSoKvZG">PaceLLM</a>, inspired by mechanisms in the prefrontal cortex, designs Persistent Activity (PA) to retrieve previous states and Cortical Expert (CE) to reorganize previous states into semantic modules. <a href="https://neurips.aipapertrails.com/paper/YaQnKRtTdh">This work</a> uses a Vision-Language Model (VLM) itself to encode and retrieve memories for a VLM, while <a href="https://neurips.aipapertrails.com/paper/ARJpQtLXfe">Memory Decoder</a> uses a plug-and-play trained memory decoder module. Similarly, <a href="https://neurips.aipapertrails.com/paper/q5QaTQcUbS">3DLLM-Mem</a> retrieves past interactions based on current observations.</p><p>There is also research that adaptively attends to different heads in self-attention to simulate learned memory retrieval and forgetting. Coincidentally, at NeurIPS, both <a href="https://neurips.aipapertrails.com/paper/RlqYCpTu1P">MoBA</a> and <a href="https://neurips.aipapertrails.com/paper/Nf8yfPDFTl">SeerAttention</a> propose this type of gating.</p><h3>4. Computer use/GUI and web agents</h3><p>There is a lot of great research on this topic at NeurIPS: building datasets and benchmarks for computer use agents, methods to train and improve computer use and web agents, and investigations into safety risks with these agents.</p><p><strong>Datasets and Benchmarks</strong>. To evaluate a computer use agent comprehensively, the dataset should contain demonstrations spanning several operating systems, applications, and websites. <a href="https://neurips.aipapertrails.com/paper/6iRZvJiC9Q">OpenCUA</a> presents data with 3 operating systems and 200+ applications and websites, along with an  annotation infrastructure.<a href="https://neurips.aipapertrails.com/paper/FS0voKxELr"> OSWorld-G </a>creates a training dataset specific to different interface elements that enables compositional generalization to novel interfaces. <a href="https://neurips.aipapertrails.com/paper/YJxGJP8feU">macOSWorld</a> creates a multi-lingual benchmark with 5 languages. <a href="https://neurips.aipapertrails.com/paper/Un1sWxmZuI">REAL</a> creates a  dataset with real websites and programmatic checks as deterministic success criteria. <a href="https://neurips.aipapertrails.com/paper/LZnKNApvhG">TheAgentCompany</a> simulates real-world tasks of a digital worker, including communicating with other coworkers.</p><p><strong>Building computer use/GUI agents</strong>. <a href="https://neurips.aipapertrails.com/paper/1XLjrmKZ4p">GUI-G1</a> sets up DeepSeek-R1-Zero type of training, where models surpass all prior models of similar size. R1-style training everywhere! <a href="https://neurips.aipapertrails.com/paper/XVm8KOO3Ri">GUI Exploration Lab</a> improves agents with multi-turn reinforcement learning as opposed to single-turn. <a href="https://neurips.aipapertrails.com/paper/3uUmJzSSOW">UI-Genie</a> takes this a step forward and defines are self-improving loop with a reward model, UI-Genie-RM. <a href="https://neurips.aipapertrails.com/paper/YMPYLesItf">GUI-Rise</a> uses structured reasoning with GRPO, along with history summarization and specialized rewards. <a href="https://neurips.aipapertrails.com/paper/B0Gfxhr8V5">BTL-UI</a> proposes a brain-inspired framework with Blink, Think, and Link that demonstrates competitive performance with other methods.</p><p><strong>Safety</strong>. <a href="https://neurips.aipapertrails.com/paper/Ip1cCUAllL">WASP</a> investigates prompt injection attacks on autonomous UI agents and finds &#8220;even top-tier AI models can be deceived by simple, low-effort human-written injections in very realistic scenarios&#8220;. <a href="https://neurips.aipapertrails.com/paper/qaxf7q41aK">AgentDAM</a> investigates the inadvertent use of unnecessary sensitive information. <a href="https://neurips.aipapertrails.com/paper/Di30GwhQSX">OS-Harm</a> tests models for deliberate user data misuse, prompt injection attacks, and model misbehavior. <a href="https://neurips.aipapertrails.com/paper/aPBBkHDK4W">RiOSWorld</a> investigations risk associated with computer-use agents in two major categories: user-originated risks and environmental risks. <a href="https://neurips.aipapertrails.com/paper/ToNRHqX6xq">MIP against Agent</a> uses adversarially perturbed image patches to test the robustness of multimodel OS agents.</p><h3>5. AI-driven scientific discovery</h3><p><a href="https://neurips.aipapertrails.com/paper/PegEYWWXvx">A position paper</a> at NeurIPS argues &#8220;that foundation models are driving a fundamental shift in the scientific process&#8221;. From paradigm enhancement to paradigm transition. To align with that, there is a lot of work at NeurIPS &#8216;25 that evaluates autonomous scientific discovery with LLMs, first data science and machine learning research, then biology, physics, mathematics, and finance. <strong>The results right now are mixed, but with a scope for improvement.</strong></p><p>Not surprisingly, ML researchers first want to automate their tasks! <a href="https://neurips.aipapertrails.com/paper/kQWyOYUAC4">AI-Researcher</a> defines complete research pipelines, from literature review and hypothesis generation to algorithm implementation and publication-ready manuscript preparation, and proposes Scientist-Bench. The authors claim that the AI-Researcher approach achieves human-level quality on this benchmark. <a href="https://neurips.aipapertrails.com/paper/RwfrdKSgCE">Another work</a> proposes better performance methods on MLE-Bench, which comprises problems from Kaggle competitions. Similarly, but with a  multi-modal pipeline, <a href="https://neurips.aipapertrails.com/paper/qGFvTIMS3W">this work</a> proposes a system for automated model discovery for a given dataset. </p><p><a href="https://neurips.aipapertrails.com/paper/JX9DE6colf">MLR-Bench</a> sources 201 research tasks from NeurIPS, ICLR, and ICML workshops, which are evaluated by MLR-Judge. Similarly, <a href="https://neurips.aipapertrails.com/paper/t8Okk2PRWU">MLRC-Bench</a> curates a suite of 7 competition tasks that reveal significant challenges for LLM agents. It shows that even the best agent closes only 9.3% of the gap between baseline and the top human participant. <a href="https://neurips.aipapertrails.com/paper/VrCdsZBbIg">Another work</a> investigated language modeling itself by using LLMs to come up with efficient architectures for LLMs. <a href="https://neurips.aipapertrails.com/paper/w98hMEjzu8">LLM Speedrunning</a> benchmark tests LLM agents against reproducing NanoGPT improvements. It concludes with &#8220;LLMs combined with SoTA scaffolds struggle to reimplement already-known innovations in our benchmark.&#8220; <a href="https://neurips.aipapertrails.com/paper/a64D9Vl7wK">Another interesting</a> test for LLM was predicting empirical AI research outcomes. Surprisingly, a system with fine-tuned GPT-4.1 and a paper retrieval agent beats human experts by a large margin!</p><p>LLMs are also evaluated for R&amp;D and scientific discovery in other domains. <a href="https://neurips.aipapertrails.com/paper/9VxTXAUH7G">R&amp;D-Agent-Quant</a> for quantitative finance, <a href="https://neurips.aipapertrails.com/paper/Vzi96rTe4w">scPilot</a> for single-cell analysis, <a href="https://neurips.aipapertrails.com/paper/w8uII2qAmd">PhysGym</a> for interactive physics discovery, <a href="https://neurips.aipapertrails.com/paper/7k7cubl1iL">CIDD</a> for drug design, <a href="https://neurips.aipapertrails.com/paper/qXiTFAgEx4">AstroVisBench</a> for scientific computing and visualization in Astronomy. <a href="https://neurips.aipapertrails.com/paper/AIOq1vWSgK">LabUtopia</a> for scientific embodied agent testing LLM or VLA models in laboratory settings.</p><p>Beyond benchmarks with deterministic success criteria, a large part of R&amp;D is non-verifiable and should be evaluated by subject matter experts. <a href="https://neurips.aipapertrails.com/paper/am6RR85mnc">SciArena</a> built a platform similar to Chatbot Arena, where human researchers across diverse scientific domains judge the answer quality of different LLMs on scientific literature-grounded tasks.</p><h3>Outro</h3><p>I am excited about all the interesting papers at NeurIPS and can&#8217;t wait to visit the posters for the above papers in San Diego next week.</p><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>AGI: Artificial General Intelligence; ASI: Artificial Super Intelligence. The community has used many practically interchangeable terms. Another is Human-level AI or Transformative AI by Sam Bowman, talking about a <a href="https://sleepinyourhat.github.io/checklist/">checklist for AI safety</a>. There are a lot of definitions floating around, but let&#8217;s go with Sam Bowman&#8217;s definition: &#8220;AI that could form as a drop-in replacement for humans in all remote-work-friendly jobs, including AI R&amp;D.&#8221;</p><p></p></div></div>]]></content:encoded></item><item><title><![CDATA[Chain of thought is computation]]></title><description><![CDATA[Framework to understand "thinking" in LLMs, from Dale Schuurmans's keynote at Reinforcement Learning Conference.]]></description><link>https://www.aipapertrails.com/p/chain-of-thought-is-computation</link><guid isPermaLink="false">https://www.aipapertrails.com/p/chain-of-thought-is-computation</guid><dc:creator><![CDATA[Prakash Kagitha]]></dc:creator><pubDate>Fri, 07 Nov 2025 07:04:02 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!Jf3h!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe3f62bca-74ff-46cd-bbdf-cc9d569d2636_1536x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Assume there is a giant directed graph and we were to find the shortest path between node A to node B. If we try to write down the <em>first</em> edge of the <em>optimal</em> path, that would be the hardest edge to choose. We need to know what is the optimal path from second node to node B to choose the first node optimally. Classic dynamic programming. In other words, we must have done <em>most of the computation</em> even before identifying the first edge of the optimal path. This computation scales linearly with the problem size, linear-time computation.</p><p>Analogically, if an LLM tries to generate the <em>final answer</em> in one shot, it would have to somehow compress a multi-step computation into a single fixed-depth forward pass (when generating the first token). But if LLMs were to generate a variable size chain of thought before generating an answer, linear-time computation is possible. <strong>Chain of thought </strong><em><strong>is</strong></em><strong> computation</strong>.</p><p><a href="https://webdocs.cs.ualberta.ca/~dale/">Dale Schuurman</a> (U. Alberta &amp; Google DeepMind, in <a href="https://www.youtube.com/watch?v=hMEViRcF7o0">his keynote at Reinforcement Learning Conference</a>, offered a perspective to understand LLMs as compute engines. Specifically, it explains why chain-of-thought is necessary for reasoning and how it can be treated as computation. This perspective answers core questions like &#8220;Can LLMs reason?&#8221; and offers a framework for understanding pre-training and post-training in LLMs. It also highlights the inherent limits of learning a formal program from finite data and raises questions about out-of-distribution generalization. The following depicts a characterization of the framework:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Jf3h!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe3f62bca-74ff-46cd-bbdf-cc9d569d2636_1536x1024.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Jf3h!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe3f62bca-74ff-46cd-bbdf-cc9d569d2636_1536x1024.png 424w, https://substackcdn.com/image/fetch/$s_!Jf3h!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe3f62bca-74ff-46cd-bbdf-cc9d569d2636_1536x1024.png 848w, https://substackcdn.com/image/fetch/$s_!Jf3h!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe3f62bca-74ff-46cd-bbdf-cc9d569d2636_1536x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!Jf3h!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe3f62bca-74ff-46cd-bbdf-cc9d569d2636_1536x1024.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Jf3h!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe3f62bca-74ff-46cd-bbdf-cc9d569d2636_1536x1024.png" width="1456" height="971" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/e3f62bca-74ff-46cd-bbdf-cc9d569d2636_1536x1024.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:971,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:2021636,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://aipapertrails.substack.com/i/175856692?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe3f62bca-74ff-46cd-bbdf-cc9d569d2636_1536x1024.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Jf3h!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe3f62bca-74ff-46cd-bbdf-cc9d569d2636_1536x1024.png 424w, https://substackcdn.com/image/fetch/$s_!Jf3h!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe3f62bca-74ff-46cd-bbdf-cc9d569d2636_1536x1024.png 848w, https://substackcdn.com/image/fetch/$s_!Jf3h!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe3f62bca-74ff-46cd-bbdf-cc9d569d2636_1536x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!Jf3h!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe3f62bca-74ff-46cd-bbdf-cc9d569d2636_1536x1024.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.aipapertrails.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading AI PaperTrails! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><h3>Linear-time &#8800; one-shot</h3><p>We <em>cannot</em> distill linear-time reasoning into constant-time architectures. As seen above, shortest paths need dynamic programming type of computation and we can&#8217;t compress this inherently iterative computation into a <strong>single constant-depth forward pass</strong>. In a similar case of linear-time computation problem, propositional logic (shown in the below figure), training with CoT that follow back substiution improved the performance considerably above random.</p><p>As Dale says, training input-output pairs (without COT) to produce the answer &#8220;in one pass&#8221; was doomed: </p><blockquote><p>&#8220;This cannot work. It&#8217;s an impossibility&#8230; It&#8217;s not a statistical problem. It&#8217;s a computer science problem. [&#8230;] If you want to solve linear-time problems, <em>give the model a budget to express linear-time computations</em>.&#8221;</p></blockquote><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!bY-l!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F42c2e2af-8eed-4fa4-9266-2f6d5b18e86c_1136x1214.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!bY-l!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F42c2e2af-8eed-4fa4-9266-2f6d5b18e86c_1136x1214.png 424w, https://substackcdn.com/image/fetch/$s_!bY-l!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F42c2e2af-8eed-4fa4-9266-2f6d5b18e86c_1136x1214.png 848w, https://substackcdn.com/image/fetch/$s_!bY-l!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F42c2e2af-8eed-4fa4-9266-2f6d5b18e86c_1136x1214.png 1272w, https://substackcdn.com/image/fetch/$s_!bY-l!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F42c2e2af-8eed-4fa4-9266-2f6d5b18e86c_1136x1214.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!bY-l!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F42c2e2af-8eed-4fa4-9266-2f6d5b18e86c_1136x1214.png" width="1136" height="1214" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/42c2e2af-8eed-4fa4-9266-2f6d5b18e86c_1136x1214.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1214,&quot;width&quot;:1136,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:507486,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://aipapertrails.substack.com/i/175856692?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F42c2e2af-8eed-4fa4-9266-2f6d5b18e86c_1136x1214.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!bY-l!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F42c2e2af-8eed-4fa4-9266-2f6d5b18e86c_1136x1214.png 424w, https://substackcdn.com/image/fetch/$s_!bY-l!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F42c2e2af-8eed-4fa4-9266-2f6d5b18e86c_1136x1214.png 848w, https://substackcdn.com/image/fetch/$s_!bY-l!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F42c2e2af-8eed-4fa4-9266-2f6d5b18e86c_1136x1214.png 1272w, https://substackcdn.com/image/fetch/$s_!bY-l!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F42c2e2af-8eed-4fa4-9266-2f6d5b18e86c_1136x1214.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>There is an alternative, as seen in <a href="https://x.com/prakashkagitha/status/1985824668599873630">NeurIPS 2025 papers</a>, to token-by-token CoT: <strong>continuous, latent-space thinking</strong>. Instead of reasoning tokens, the model allocates computation inside its vectors. </p><p><a href="http://arxiv.org/abs/2505.12514v2">Reasoning by Superposition: A Theoretical Perspective on Chain of Continuous Thought</a> argues that continuous CoT can encode multiple search frontiers in superposition, enabling something like parallel BFS with fewer explicit steps than discrete CoT. This doesn&#8217;t make compute vanish, it changes where the compute lives (in vectors rather than many printed tokens) and shows how a small number of &#8220;thinking updates&#8221; can still represent multi-path exploration. </p><p><a href="http://arxiv.org/abs/2502.05171v2">Scaling up Test-Time Compute with Latent Reasoning</a> proposes iterating a recurrent block at test time, no special CoT data, so the model scales depth as needed (adaptive, budgeted compute) without spitting out long text traces<em>. </em></p><p>Read together, these results pressure-test token CoT by showing that compute doesn&#8217;t disappear, it <strong>moves</strong>: from explicit text to a handful of latent updates. They don&#8217;t refute &#8220;linear-time &#8800; one-shot&#8221;; they offer tighter <strong>allocation and compression</strong> of the required iterative compute.</p><h3>Pre-training and post-training</h3><p>If the internet is a giant directed graph, <strong>pre-training</strong> learns the <em>edges</em> (token-transition statistics). <strong>Post-training</strong> learns how to <em>traverse</em> the graph for a goal. Schuurmans&#8217;s metaphor is crisp: &#8220;Pre-training is learning the graph&#8230; Post-training is about turning that into a policy that finds paths.&#8221;  This elicits reasoning which executes the policy learned. Talking about the performance of reasoning models for logical reasoning tasks, <a href="https://openreview.net/forum?id=MPTlWIVSMU#discussion">this paper</a> says:</p><blockquote><p>&#8220;Unlike other LLMs (GPT-4o &amp; DeepSeek V3), R1 shows signs of having learned the underlying reasoning&#8220;</p></blockquote><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!N2Fh!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F19b630dc-3dc6-44c2-89b6-5f256dd88e66_2284x1246.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!N2Fh!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F19b630dc-3dc6-44c2-89b6-5f256dd88e66_2284x1246.png 424w, https://substackcdn.com/image/fetch/$s_!N2Fh!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F19b630dc-3dc6-44c2-89b6-5f256dd88e66_2284x1246.png 848w, https://substackcdn.com/image/fetch/$s_!N2Fh!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F19b630dc-3dc6-44c2-89b6-5f256dd88e66_2284x1246.png 1272w, https://substackcdn.com/image/fetch/$s_!N2Fh!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F19b630dc-3dc6-44c2-89b6-5f256dd88e66_2284x1246.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!N2Fh!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F19b630dc-3dc6-44c2-89b6-5f256dd88e66_2284x1246.png" width="1456" height="794" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/19b630dc-3dc6-44c2-89b6-5f256dd88e66_2284x1246.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:794,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1619993,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://aipapertrails.substack.com/i/175856692?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F19b630dc-3dc6-44c2-89b6-5f256dd88e66_2284x1246.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!N2Fh!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F19b630dc-3dc6-44c2-89b6-5f256dd88e66_2284x1246.png 424w, https://substackcdn.com/image/fetch/$s_!N2Fh!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F19b630dc-3dc6-44c2-89b6-5f256dd88e66_2284x1246.png 848w, https://substackcdn.com/image/fetch/$s_!N2Fh!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F19b630dc-3dc6-44c2-89b6-5f256dd88e66_2284x1246.png 1272w, https://substackcdn.com/image/fetch/$s_!N2Fh!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F19b630dc-3dc6-44c2-89b6-5f256dd88e66_2284x1246.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h3>Out-of-distribution generalization &amp; Correctness</h3><p>Even though an LLM is universal <em>in principle</em>, program induction from finite data guarantees there will be inputs just outside what the learned program covers. Schuurmans makes the broader point that machine learning promises are mostly <em>in-distribution</em>, while program execution aims for <em>all instances</em>. If you ignore the difference, you misdiagnose computational failures as &#8220;OOD coverage problems.&#8221; </p><p>Also, how can we characterise CoTs with self-verification &amp; backtracking? Are they verifying the program and choosing a different program to run? Isn&#8217;t the case program correctness is undecidable?</p><p>There are many interesting questions with this framework of treating LLM as a computer. In Dale Schuurmans&#8217;s words:</p><blockquote><p>&#8220;machine learning is awesome&#8230; but computer science matters&#8230; If you ignore those laws, I predict disappointment.&#8221;</p></blockquote><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.aipapertrails.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading AI PaperTrails! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Large reasoning models research at COLM 2025]]></title><description><![CDATA[State of research in scaling reasoning, the current paradigm for improving LLMs]]></description><link>https://www.aipapertrails.com/p/large-reasoning-models-research-at</link><guid isPermaLink="false">https://www.aipapertrails.com/p/large-reasoning-models-research-at</guid><dc:creator><![CDATA[Prakash Kagitha]]></dc:creator><pubDate>Sat, 13 Sep 2025 23:48:38 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/9d36227b-ce8c-4b56-ae0f-ad3b8474ccdf_1024x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Nr3v!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdd069c01-5c3f-49af-bc6c-b1b75760c167_1024x768.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Nr3v!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdd069c01-5c3f-49af-bc6c-b1b75760c167_1024x768.png 424w, https://substackcdn.com/image/fetch/$s_!Nr3v!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdd069c01-5c3f-49af-bc6c-b1b75760c167_1024x768.png 848w, https://substackcdn.com/image/fetch/$s_!Nr3v!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdd069c01-5c3f-49af-bc6c-b1b75760c167_1024x768.png 1272w, https://substackcdn.com/image/fetch/$s_!Nr3v!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdd069c01-5c3f-49af-bc6c-b1b75760c167_1024x768.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Nr3v!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdd069c01-5c3f-49af-bc6c-b1b75760c167_1024x768.png" width="1024" height="768" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/dd069c01-5c3f-49af-bc6c-b1b75760c167_1024x768.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:768,&quot;width&quot;:1024,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1880259,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://aipapertrails.substack.com/i/173451048?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff15f5354-8966-47da-805e-0d560edb701e_1024x1024.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Nr3v!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdd069c01-5c3f-49af-bc6c-b1b75760c167_1024x768.png 424w, https://substackcdn.com/image/fetch/$s_!Nr3v!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdd069c01-5c3f-49af-bc6c-b1b75760c167_1024x768.png 848w, https://substackcdn.com/image/fetch/$s_!Nr3v!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdd069c01-5c3f-49af-bc6c-b1b75760c167_1024x768.png 1272w, https://substackcdn.com/image/fetch/$s_!Nr3v!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdd069c01-5c3f-49af-bc6c-b1b75760c167_1024x768.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Conference on Language Models (COLM) will take place in Montreal from October 7-10, 2025. There are around 70, out of over 400, research papers related to &#8220;reasoning&#8221;<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a>. Find all the accepted papers with a reasoning tag, abstracts, authors, PDF links, and review scores in this <a href="https://docs.google.com/spreadsheets/d/1PHWsM6GWlRPlVEPQlfqJhXn7l5gQlWZg0F0jmxbGnXQ/edit?usp=sharing">Google Sheet</a>.</p><p>Peer-reviewed papers from an upcoming top LLM conference must give a broad sense of the research area. As Large Reasoning Models (LRMs)<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-2" href="#footnote-2" target="_self">2</a> are undeniably the cutting edge of language models, this first post on COLM 2025 papers introduces the latest research on LRMs, a subset of 70 reasoning papers, covering research:</p><ol><li><p>Improving reasoning in LRMs,</p></li><li><p>Applications of LRMs in different domains &amp; tasks,</p></li><li><p>Multi-modal reasoning enabled by LRMs,</p></li><li><p>Many limitations of LRMs, and</p></li><li><p>Understanding the reasoning in LRMs.</p></li></ol><p>I aim to provide clear segues with friendly context to different research papers, particularly, how they tie into the narrative of &#8220;scaling reasoning is the paradigm that pushed the frontier this time around&#8221;.</p><h3>0. LRMs represent the cutting edge of Language Models</h3><p>Large Language Models (LLMs) have been improving rapidly since 2020. The initial progress came from the paradigm of <strong>scaling pretraining</strong>, more data &amp; model size, giving LLMs robust natural language understanding and instruction following, with models like GPT-4. While these models had never-seen-before reasoning capabilities for an AI, they were not human-level in many reasoning and planning tasks. Scaling pretraining didn&#8217;t help with this, as the gains have plateaued.</p><p><strong>Scaling reasoning</strong>/thinking tokens with inference-time compute has emerged as the new paradigm that improved LLMs, evident with pioneering Large Reasoning Models (LRMs) OpenAI O1 and DeepSeek R1. This improved the capability to perform complex multi-step reasoning tasks and extended tool-use, enabling several applications of LLM Agents, including for software engineering and computer use. We still didn&#8217;t see a plateau in scaling reasoning, but we could be racing towards it!</p><p>Unlike speculation, research on LRMs at COLM 2025 will shed some light on where we stand. In summary: Several methods have been proposed to improve reasoning and inference-time compute, suggesting that scaling reasoning might still have enough headroom. Reasoning models have been trained for several tasks in multiple domains, justifying being an effective technique. Also, reasoning was key to enabling a few multi-modal reasoning tasks, suggesting the versatility. Notwithstanding spectacular achievements like <a href="https://deepmind.google/discover/blog/advanced-version-of-gemini-with-deep-think-officially-achieves-gold-medal-standard-at-the-international-mathematical-olympiad/">earning a gold medal at IMO 2025</a>, LRMs have many limitations, suggesting that scaling reasoning might not be a panacea. Fortunately, there are several investigations into scientifically understanding LRMs, which should be a guiding light to understand limitations, improve, or look for the next paradigm.</p><h3>1. Improving the reasoning of LRMs</h3><p>Striking and first evidence with <a href="https://openai.com/index/learning-to-reason-with-llms/">OpenAI O1</a> that scaling reasoning will be the next frontier is self-verification and backtracking in language models&#8217; reasoning chains. Hence, Large Reasoning Models. Also, the aha-moment from DeepSeek R1, which showed how to achieve this long reasoning that others are still following. There are avenues to improve reasoning still.</p><h4>Algorithmic synthetic data for long reasoning</h4><p><a href="https://openreview.net/forum?id=9DCQAGBoII#discussion">Spontaneous self-correction</a> method distills self-verification capability from multi-turn, multi-agent collaboration into a single turn generation. After distillation with Supervised Fine-Tuning (SFT), online RL is used to improve self-verification.</p><p><a href="https://openreview.net/forum?id=6vMRcaYbU7#discussion">Another work</a> generates diverse reasoning traces on graph problems that underscore broad reasoning abilities required for math and scientific reasoning. Training on these reasoning traces also showed ~20% improvement in non-mathematical reasoning tasks like logical and commonsense reasoning.</p><p>Similarly, <a href="https://openreview.net/forum?id=oN9STRYQVa#discussion">Step-Wise RL (SWiRL</a>) applies synthetic data filtering and RL optimisation to improve math reasoning and question answering with retrieval.</p><h4>Verifiers and multi-agent systems for better reasoning</h4><p>R1-type of training incentivises long reasoning with rewards and RL (GRPO). Another way to obtain self-verification capability is to have multi-agent systems with generators &amp; verifiers. This is long reasoning in effect.</p><p><a href="https://openreview.net/forum?id=kVOrGZM5N7#discussion">Putting the value back in RL.</a> This augments RL methods without value functions, like GRPO, with generative verifiers to enable 8-32x efficient test-time compute scaling.</p><p>Another work just does away with RL and gets comparable performance with just pair-wise preference optimisation training, both generator and verifier - <a href="https://openreview.net/forum?id=OgWh4J7bkT#discussion">Iterative DPO</a></p><p><a href="https://openreview.net/forum?id=zLbmsdyTiN#discussion">MeMAD</a> enables self-verification with a separate reflection module and memory bank during the debate of multiple agents.</p><h4>Novel approaches for reasoning</h4><p>Most LRMs reason by generating natural language tokens. We don&#8217;t know whether this is an effective and comprehensive way of thinking in language models. <a href="https://openreview.net/forum?id=Itxz7S4Ip3#discussion">COCONUT</a> proposes a new way of reasoning in latent space: generating a sequence of continuous vectors, but not generating or training to generate natural language tokens. The authors showed that continuous thought is better than NL thought when substantial search is needed in a reasoning task, like finding a valid path in a big graph.</p><p>While reasoning is generally serialized, <a href="https://openreview.net/forum?id=YgwQ7sXPXU#discussion">Adaptive Parallel Reasoning (APR)</a> tries to spawn multiple reasoning chains and learns to coordinate among them with a new end-to-end RL strategy.</p><p>Apparently, hidden states of reasoning tokens in LRMs have information about answer correctness and different phases like execution, reflection, and transition thoughts. <a href="https://openreview.net/forum?id=klPszYDIRT#discussion">SEAL</a> reduces excessive reflection and transition thoughts with reasoning steering vectors, and <a href="https://openreview.net/forum?id=O6I0Av7683#discussion">the other work</a> early exits reasoning by probing correctness from hidden states.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.aipapertrails.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading AI PaperTrails! Subscribe for free to receive future posts on COLM 2025 papers and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><h3>2. Applications of LRMs in different domains &amp; tasks</h3><p>DeepSeek R1-style of training to incentivise reasoning has become the norm to improve complex reasoning. Long thoughts, more reasoning steps, and extended tool calls underlie improved performance in all the following tasks.</p><p><strong>LRMs for Medical knowledge reasoning</strong>. <a href="https://openreview.net/forum?id=7ZwuGZCopw#discussion">FineMedLM-o1</a> is trained on medical long-form reasoning data to enable advanced dialogue and deep reasoning capabilities supporting differential diagnosis and medication recommendations.</p><p><strong>LRMs for Law. </strong>Towards reasoning-aware legal AI systems, <a href="https://openreview.net/forum?id=MsgdEkcLRz#discussion">LawFlow</a> collects lawyers&#8217; thought processes and compares them to LRM&#8217;s thought processes.</p><p><strong>LRMs for <a href="https://openreview.net/forum?id=MPTlWIVSMU#discussion">logical reasoning</a></strong>. Performing algorithmic reasoning for the 3-SAT problem, the authors found that &#8220;Unlike other LLMs (GPT-4o &amp; DeepSeek V3), R1 shows signs of having learned the underlying reasoning&#8220;. LRMs internalized search process required for algorithmic reasoning!</p><p><strong>LRMs for <a href="https://openreview.net/forum?id=dr3eg5ehR2#discussion">story generation</a>. </strong>Trained with RL with verifiable rewards, similar to R1, to reason over a story and plan for the next chapter, enabling the generation of high-quality stories spanning thousands of tokens.</p><p><strong>LRMs for search</strong>. <a href="https://openreview.net/forum?id=Rwhi91ideu#discussion">Search-R1.</a> &#8220;LLM learns to autonomously generate (multiple) search queries during step-by-step reasoning with real-time retrieval.&#8221;</p><p><strong>LRMs for Text-to-SQL. </strong><a href="https://openreview.net/forum?id=HbwkIDWQgN#discussion">Reasoning-SQL</a>. Improving reasoning in Text-to-SQL with rewards tailored to SQL generation, such as syntax checks, schema-linking, and n-gram similarity. RL-trained 14B model outperforms O3-mini by 4%.</p><p><strong>LRMs for summarisation</strong>. This work proposes a long-COT dataset with reflective reasoning, SumFeed-CoT. Trained on this, <a href="https://openreview.net/forum?id=6BGDGKZN7q#discussion">ReFeed</a> enhances summarization refinement in multiple dimensions through reflective reasoning on feedback.</p><h3>3. Multi-modal reasoning enabled by LRMs</h3><p>Long thoughts are a characteristic of system-2 reasoning, and it is shown to improve visual reasoning as well, which is thought to be system-1 heavy. (pun intended :))</p><p><a href="https://openreview.net/forum?id=SrKdi4MsUW#discussion">LongPerceptualThoughts</a> are long-thought traces for perceptual tasks, made verifiable with their three-state framework, starting from multiple-choice questions to CoTs. Training on these thoughts achieves +3.4 points over 5 vision benchmarks.</p><p><a href="https://openreview.net/forum?id=19fydz1QnW#discussion">BigCharts-R1</a> is the SOTA chart reasoning model trained in R1-style. The authors tackle this hard-to-verify task by rendering diverse chart images with real-world charts and data. This, by definition, has ground truths, thus providing rewards.</p><h3>4. Many limitations of LRMs</h3><p>With all the improvements in reasoning and new tasks that it enables, it is easier to intuit that long reasoning is enough for any task. At least not yet. LRMs fall short in several abilities as a language model in general and sometimes as a reasoning model specifically.</p><p><strong>Reasoning is not all you need?!</strong> <a href="https://openreview.net/forum?id=kH6LOHGjEl#discussion">Corrupted by Reasoning</a> finds that LRMs struggle significantly with cooperation compared to traditional LLMs. In Language model as a judge settings, although LRMs are better at fact-checking, LRMs show a &#8220;<a href="https://openreview.net/forum?id=SlRtFwBdzP#discussion">superficial reflection bias</a>&#8220; where phrases mimicking reasoning (e.g., &#8220;wait, let me think&#8230;&#8220;) significantly influence model judgements.</p><p><strong>Limited in uncertainty awareness. </strong>Despite having reasoning tokens, language models struggle with uncertainty-awareness. This leads to unfaithful reasoning traces and divergent conclusions given logically inconsistent knowledge, only to be disambiguated by uncertainty. Two similar works use <a href="https://openreview.net/forum?id=p4wZfBFgyI#discussion">formal reasoning topology</a> and <a href="https://openreview.net/forum?id=9pzNFfgtyk#discussion">weights of logical rules</a> for uncertainty measurement and find that SOTA LLMs/LRMs are limited.</p><p><strong>Limited in mathematical reasoning</strong>. Really?! Despite unparalleled focus on mathematical reasoning in scaling reasoning, <a href="https://openreview.net/forum?id=uXR2KsA4L9#discussion">Brains vs. Bytes</a> found that O3-mini drops from 48.3% to 14.3% if evaluated on complete proofs, not just the  final answer. Through manual expert analysis on 455 IMO problems, they found that accuracy dropped for models including DeepSeek, Gemini, and OpenAI.</p><p><strong>Limited in code reasoning</strong>. O3-mini(high) <a href="https://openreview.net/forum?id=M7cl4Ldw61#discussion">couldn&#8217;t identify counterexamples</a> for over 90% of incorrect solutions where expert humans can. In long-horizon SWE agent tasks, LLMs are not robust to corrupted functions and <a href="https://openreview.net/forum?id=GQNojroNCH#discussion">fail at system-level reasoning</a>.</p><p>Language models are categorically worse than humans in <a href="https://openreview.net/forum?id=78lTuD6wiO#discussion">visual cognition</a>, <a href="https://openreview.net/forum?id=vNJbDhgrM4#discussion">social reasoning</a>, and <a href="https://openreview.net/forum?id=qbWpEufkqk#discussion">spatial reasoning</a>. Also, LLMs are limited in <a href="https://openreview.net/forum?id=fcRcl1EXc4#discussion">linguistic reasoning</a> in low-resource languages.</p><h3>5. Understanding the reasoning in LRMs</h3><p>Soon after the success of R1, several related RL training experiments were spawned to incentivise long reasoning, in pursuit of understanding its essence. Several works found that long reasoning is only possible if base models themselves have properties like verification, subgoal setting, and backtracking, like Qwen 2.5 models as opposed to Llama 3.2 models.</p><p><a href="https://openreview.net/forum?id=vSMCBUgrQj#discussion">SimpleRL-Zoo</a> finds that by adjusting format reward and controlling query difficulty, most model families show long reasoning performance boosts, albeit with distinct patterns during training. </p><p><a href="https://openreview.net/forum?id=5wAfbEs34A#discussion">Style over substance</a> also found that small models learn reasoning style from distillation rather than substance, and thus achieve similar performance during training with correct or incorrect answers. A<a href="https://openreview.net/forum?id=QGJ9ttXLTy#discussion"> framework to investigate factors </a>that enable long reasoning, also found that priming with few-shot examples with reasoning patterns also elicited long reasoning in most models, even when long reasoning patterns led to incorrect answers.</p><p>As long into the past as the content vs form debate goes, so does it into many domains! This might give a plausible hypothesis for other limitations of LLMs at COLM 2025: LRMs <a href="https://openreview.net/forum?id=ufozo2Wc9e#discussion">go on to overthink</a> even with a missing premise. Maybe they are just following the reasoning style. LRMs <a href="https://openreview.net/forum?id=ptmgWRCWmu#discussion">couldn&#8217;t identify</a> inconsistencies in FlawedFictions as well, maybe because the high-level linguistic structure still makes sense.</p><p>With all the findings and conclusions from scientific investigations in this post, it is imperative to give a shout-out to all the reproducible efforts in any paper. Related work, <a href="https://openreview.net/forum?id=90UrTTxp5O#discussion">Sober look at progress</a>, proposed an evaluation framework to understand the sensitivity of findings to subtle implementation choices, including random seeds, prompt formatting, and decoding parameters. It found that SFT shows consistently stronger generalisation than RL, contrary to the popular belief.</p><h3>&#167;. Looking forward</h3><p>Autonomous scientific discovery to win a Nobel prize would be a great goal for superhuman AI. Can LRMs alone take us there? (including multi-agent systems powered by them) Or do we need a new paradigm or two that push the frontier?</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.aipapertrails.com/p/large-reasoning-models-research-at/comments&quot;,&quot;text&quot;:&quot;Leave a comment&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.aipapertrails.com/p/large-reasoning-models-research-at/comments"><span>Leave a comment</span></a></p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.aipapertrails.com/p/large-reasoning-models-research-at?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.aipapertrails.com/p/large-reasoning-models-research-at?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>including: reasoning or thinking in Large Reasoning Models (LRMs), mathematical, algorithmic, social, inductive, visual, and other reasoning capabilities of LLMs.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-2" href="#footnote-anchor-2" class="footnote-number" contenteditable="false" target="_self">2</a><div class="footnote-content"><p>LRMs are also LLMs. I don&#8217;t like this distinction so much, but it helps in this article.</p><p></p></div></div>]]></content:encoded></item><item><title><![CDATA[ICML 2024 Orals (top 10%): Summaries of interesting papers]]></title><description><![CDATA[ICML, one of the top machine learning conferences, is happening this week.]]></description><link>https://www.aipapertrails.com/p/icml-2024-orals-top-10-summaries-of-interesting-papers-53c415a891de</link><guid isPermaLink="false">https://www.aipapertrails.com/p/icml-2024-orals-top-10-summaries-of-interesting-papers-53c415a891de</guid><dc:creator><![CDATA[Prakash Kagitha]]></dc:creator><pubDate>Tue, 23 Jul 2024 16:14:59 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/5e68f981-8c09-4395-b357-d643e40b430a_1024x576.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>ICML, one of the top machine learning conferences, is happening this week. I am excited about a few tutorials, orals &amp; posters, and most importantly workshops towards the end of the week. I found the following oral papers interesting and wrote friendly summaries.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!X5XK!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6bab5bf6-c88e-4882-9a25-7de9ee875744_1024x576.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!X5XK!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6bab5bf6-c88e-4882-9a25-7de9ee875744_1024x576.jpeg 424w, https://substackcdn.com/image/fetch/$s_!X5XK!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6bab5bf6-c88e-4882-9a25-7de9ee875744_1024x576.jpeg 848w, https://substackcdn.com/image/fetch/$s_!X5XK!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6bab5bf6-c88e-4882-9a25-7de9ee875744_1024x576.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!X5XK!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6bab5bf6-c88e-4882-9a25-7de9ee875744_1024x576.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!X5XK!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6bab5bf6-c88e-4882-9a25-7de9ee875744_1024x576.jpeg" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/6bab5bf6-c88e-4882-9a25-7de9ee875744_1024x576.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!X5XK!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6bab5bf6-c88e-4882-9a25-7de9ee875744_1024x576.jpeg 424w, https://substackcdn.com/image/fetch/$s_!X5XK!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6bab5bf6-c88e-4882-9a25-7de9ee875744_1024x576.jpeg 848w, https://substackcdn.com/image/fetch/$s_!X5XK!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6bab5bf6-c88e-4882-9a25-7de9ee875744_1024x576.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!X5XK!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6bab5bf6-c88e-4882-9a25-7de9ee875744_1024x576.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div></div></div></a><figcaption class="image-caption">Created by&nbsp;author</figcaption></figure></div><h4><a href="https://openreview.net/forum?id=ghNRg2mEgN">Weak-to-Strong Generalization: Eliciting Strong Capabilities With Weak Supervision</a> (From&nbsp;OpenAI)</h4><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!hLzZ!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9f8ecd26-5723-4a6b-8513-ab29af88d7ee_1018x472.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!hLzZ!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9f8ecd26-5723-4a6b-8513-ab29af88d7ee_1018x472.png 424w, https://substackcdn.com/image/fetch/$s_!hLzZ!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9f8ecd26-5723-4a6b-8513-ab29af88d7ee_1018x472.png 848w, https://substackcdn.com/image/fetch/$s_!hLzZ!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9f8ecd26-5723-4a6b-8513-ab29af88d7ee_1018x472.png 1272w, https://substackcdn.com/image/fetch/$s_!hLzZ!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9f8ecd26-5723-4a6b-8513-ab29af88d7ee_1018x472.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!hLzZ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9f8ecd26-5723-4a6b-8513-ab29af88d7ee_1018x472.png" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/9f8ecd26-5723-4a6b-8513-ab29af88d7ee_1018x472.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!hLzZ!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9f8ecd26-5723-4a6b-8513-ab29af88d7ee_1018x472.png 424w, https://substackcdn.com/image/fetch/$s_!hLzZ!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9f8ecd26-5723-4a6b-8513-ab29af88d7ee_1018x472.png 848w, https://substackcdn.com/image/fetch/$s_!hLzZ!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9f8ecd26-5723-4a6b-8513-ab29af88d7ee_1018x472.png 1272w, https://substackcdn.com/image/fetch/$s_!hLzZ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9f8ecd26-5723-4a6b-8513-ab29af88d7ee_1018x472.png 1456w" sizes="100vw"></picture><div></div></div></a></figure></div><p>Currently, LLMs are aligned with Reinforcement Learning with Human Feedback (RLHF), i.e. we label whether a response followed the human intent in the query or whether a response is safe. But when LLMs become superhuman we cannot label the responses of superhuman LLMs to make them better or aligned to human values. For instance, if an LLM generated a code repository of 1 million lines, we cannot easily label the code that it is safe or that it followed the user intent. In this scenario, humans, weak agents, have to somehow align strong agents, superhuman LLMs.</p><p>This paper from OpenAI presents empirical investigations in answering this question. Analogical to using human labels to align/train superhuman LLMs, the authors set up their experiments to train a strong model, GPT-4, trained with labels of GPT-2. If GPT-4 learns from the GPT-2 labels, they call it a weak-to-strong generalization/performance.</p><p>The paper starts by defining three concepts. 1, <strong>Weak performance</strong>: GPT-2 trained on ground truth labels for a task 2, <strong>Strong ceiling performance</strong>: GPT-4 trained on ground truth labels 3, <strong>Weak-to-strong performance</strong>: GPT-4 trained with labels from GPT-2 (First, GPT-2 is trained on a task and it labels data points from the held-out dataset. This labeled dataset is used to train&nbsp;GPT-4.)</p><p>Now, the below passage from the paper defines the performance gap recovered (PGR), the extent strong student (GPT-4) recovered the strong ceiling performance with the labels from a weak supervisor (GPT-2). Hig PGR means that the weak supervisor can train strong students, or humans could align superhuman LLMs in the&nbsp;future.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!cAml!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff12f3d96-b3de-45f7-8ee3-3a52f747de60_806x702.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!cAml!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff12f3d96-b3de-45f7-8ee3-3a52f747de60_806x702.png 424w, https://substackcdn.com/image/fetch/$s_!cAml!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff12f3d96-b3de-45f7-8ee3-3a52f747de60_806x702.png 848w, https://substackcdn.com/image/fetch/$s_!cAml!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff12f3d96-b3de-45f7-8ee3-3a52f747de60_806x702.png 1272w, https://substackcdn.com/image/fetch/$s_!cAml!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff12f3d96-b3de-45f7-8ee3-3a52f747de60_806x702.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!cAml!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff12f3d96-b3de-45f7-8ee3-3a52f747de60_806x702.png" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/f12f3d96-b3de-45f7-8ee3-3a52f747de60_806x702.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!cAml!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff12f3d96-b3de-45f7-8ee3-3a52f747de60_806x702.png 424w, https://substackcdn.com/image/fetch/$s_!cAml!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff12f3d96-b3de-45f7-8ee3-3a52f747de60_806x702.png 848w, https://substackcdn.com/image/fetch/$s_!cAml!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff12f3d96-b3de-45f7-8ee3-3a52f747de60_806x702.png 1272w, https://substackcdn.com/image/fetch/$s_!cAml!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff12f3d96-b3de-45f7-8ee3-3a52f747de60_806x702.png 1456w" sizes="100vw"></picture><div></div></div></a></figure></div><p>Fortunately, weak-to-strong generalization is possible as shown across diverse tasks. 80% in NLP tasks and over 60 % in reward modeling tasks as seen in the below figure. Chess Puzzles saw the least weak-to-strong generalization with below 40&nbsp;percent.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!bskS!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F79cb82f2-25ad-421c-a8a0-50f87997c6c7_1024x436.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!bskS!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F79cb82f2-25ad-421c-a8a0-50f87997c6c7_1024x436.png 424w, https://substackcdn.com/image/fetch/$s_!bskS!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F79cb82f2-25ad-421c-a8a0-50f87997c6c7_1024x436.png 848w, https://substackcdn.com/image/fetch/$s_!bskS!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F79cb82f2-25ad-421c-a8a0-50f87997c6c7_1024x436.png 1272w, https://substackcdn.com/image/fetch/$s_!bskS!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F79cb82f2-25ad-421c-a8a0-50f87997c6c7_1024x436.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!bskS!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F79cb82f2-25ad-421c-a8a0-50f87997c6c7_1024x436.png" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/79cb82f2-25ad-421c-a8a0-50f87997c6c7_1024x436.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!bskS!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F79cb82f2-25ad-421c-a8a0-50f87997c6c7_1024x436.png 424w, https://substackcdn.com/image/fetch/$s_!bskS!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F79cb82f2-25ad-421c-a8a0-50f87997c6c7_1024x436.png 848w, https://substackcdn.com/image/fetch/$s_!bskS!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F79cb82f2-25ad-421c-a8a0-50f87997c6c7_1024x436.png 1272w, https://substackcdn.com/image/fetch/$s_!bskS!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F79cb82f2-25ad-421c-a8a0-50f87997c6c7_1024x436.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a></figure></div><p>The paper also proposed approaches to improve weak-to-strong generalization:</p><p><strong>Auxiliary loss based on strong student&#8217;s confidence</strong>: While training with labels from a weak supervisor (GPT-2), sometimes a strong student (GPT-4) might overfit to the mistakes of a weak supervisor. The paper proposed to add a loss term based on the confidence of the strong student helping it to stick to its own prediction when in conflict with the weak supervisor.</p><p><strong>Bootstrapping</strong>: Instead of a weak model (GPT-2) supervising the strong model (GPT-4) directly, this approach utilizes intermediate models in between the weak and strong. For instance, if the weak model is 1B and the strong model is 100B, a sequence of models that are 2B, 4B, 8B, 16B, 32B, and 64B would be used. The bootstrap process starts by training the 2B model with weak labels, the 4B model with 2B model labels, the 8B model with 4B model, and so on. The strong model would be trained with labels of the model before it in the above sequence.</p><p><strong>Unsupervised generative fine-tuning</strong>: Sometimes just training with the next token prediction on the training data without labels would help in learning the right representation or eliciting the required behavior for that task. This approach includes unsupervised fine-tuning with all data points without labels before fine-tuning.</p><p>The paper showed that the above approach improved weak-to-strong generalization compared to standard fine-tuning with weak labels. (Notice the difference between dotted lines and solid&nbsp;lines)</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!i0Im!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe83720a2-318d-44f2-b93e-36a12aa062e2_842x444.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!i0Im!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe83720a2-318d-44f2-b93e-36a12aa062e2_842x444.png 424w, https://substackcdn.com/image/fetch/$s_!i0Im!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe83720a2-318d-44f2-b93e-36a12aa062e2_842x444.png 848w, https://substackcdn.com/image/fetch/$s_!i0Im!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe83720a2-318d-44f2-b93e-36a12aa062e2_842x444.png 1272w, https://substackcdn.com/image/fetch/$s_!i0Im!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe83720a2-318d-44f2-b93e-36a12aa062e2_842x444.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!i0Im!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe83720a2-318d-44f2-b93e-36a12aa062e2_842x444.png" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/e83720a2-318d-44f2-b93e-36a12aa062e2_842x444.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!i0Im!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe83720a2-318d-44f2-b93e-36a12aa062e2_842x444.png 424w, https://substackcdn.com/image/fetch/$s_!i0Im!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe83720a2-318d-44f2-b93e-36a12aa062e2_842x444.png 848w, https://substackcdn.com/image/fetch/$s_!i0Im!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe83720a2-318d-44f2-b93e-36a12aa062e2_842x444.png 1272w, https://substackcdn.com/image/fetch/$s_!i0Im!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe83720a2-318d-44f2-b93e-36a12aa062e2_842x444.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a></figure></div><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!nFH0!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdfab6644-287c-4b92-90b9-55c79b53689b_800x416.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!nFH0!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdfab6644-287c-4b92-90b9-55c79b53689b_800x416.png 424w, https://substackcdn.com/image/fetch/$s_!nFH0!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdfab6644-287c-4b92-90b9-55c79b53689b_800x416.png 848w, https://substackcdn.com/image/fetch/$s_!nFH0!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdfab6644-287c-4b92-90b9-55c79b53689b_800x416.png 1272w, https://substackcdn.com/image/fetch/$s_!nFH0!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdfab6644-287c-4b92-90b9-55c79b53689b_800x416.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!nFH0!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdfab6644-287c-4b92-90b9-55c79b53689b_800x416.png" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/dfab6644-287c-4b92-90b9-55c79b53689b_800x416.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!nFH0!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdfab6644-287c-4b92-90b9-55c79b53689b_800x416.png 424w, https://substackcdn.com/image/fetch/$s_!nFH0!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdfab6644-287c-4b92-90b9-55c79b53689b_800x416.png 848w, https://substackcdn.com/image/fetch/$s_!nFH0!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdfab6644-287c-4b92-90b9-55c79b53689b_800x416.png 1272w, https://substackcdn.com/image/fetch/$s_!nFH0!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdfab6644-287c-4b92-90b9-55c79b53689b_800x416.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a></figure></div><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!tly9!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F39465c8f-b64c-4abc-9252-712d0e91d2e5_816x442.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!tly9!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F39465c8f-b64c-4abc-9252-712d0e91d2e5_816x442.png 424w, https://substackcdn.com/image/fetch/$s_!tly9!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F39465c8f-b64c-4abc-9252-712d0e91d2e5_816x442.png 848w, https://substackcdn.com/image/fetch/$s_!tly9!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F39465c8f-b64c-4abc-9252-712d0e91d2e5_816x442.png 1272w, https://substackcdn.com/image/fetch/$s_!tly9!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F39465c8f-b64c-4abc-9252-712d0e91d2e5_816x442.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!tly9!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F39465c8f-b64c-4abc-9252-712d0e91d2e5_816x442.png" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/39465c8f-b64c-4abc-9252-712d0e91d2e5_816x442.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!tly9!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F39465c8f-b64c-4abc-9252-712d0e91d2e5_816x442.png 424w, https://substackcdn.com/image/fetch/$s_!tly9!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F39465c8f-b64c-4abc-9252-712d0e91d2e5_816x442.png 848w, https://substackcdn.com/image/fetch/$s_!tly9!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F39465c8f-b64c-4abc-9252-712d0e91d2e5_816x442.png 1272w, https://substackcdn.com/image/fetch/$s_!tly9!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F39465c8f-b64c-4abc-9252-712d0e91d2e5_816x442.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a></figure></div><h4><a href="https://openreview.net/forum?id=CfOtiepP8s">Interpreting and Improving Large Language Models in Arithmetic Calculation</a></h4><p>Although Large Language Models (LLMs) show impressive performance in solving math word problems and perform descent arithmetic calculations, we don&#8217;t yet understand how they do mathematical reasoning. This paper shines a light on the mechanisms that underlie mathematical reasoning.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!rsXx!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff07d98b3-0653-4988-b6d2-1ac07e75edf7_1024x704.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!rsXx!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff07d98b3-0653-4988-b6d2-1ac07e75edf7_1024x704.png 424w, https://substackcdn.com/image/fetch/$s_!rsXx!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff07d98b3-0653-4988-b6d2-1ac07e75edf7_1024x704.png 848w, https://substackcdn.com/image/fetch/$s_!rsXx!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff07d98b3-0653-4988-b6d2-1ac07e75edf7_1024x704.png 1272w, https://substackcdn.com/image/fetch/$s_!rsXx!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff07d98b3-0653-4988-b6d2-1ac07e75edf7_1024x704.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!rsXx!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff07d98b3-0653-4988-b6d2-1ac07e75edf7_1024x704.png" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/f07d98b3-0653-4988-b6d2-1ac07e75edf7_1024x704.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!rsXx!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff07d98b3-0653-4988-b6d2-1ac07e75edf7_1024x704.png 424w, https://substackcdn.com/image/fetch/$s_!rsXx!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff07d98b3-0653-4988-b6d2-1ac07e75edf7_1024x704.png 848w, https://substackcdn.com/image/fetch/$s_!rsXx!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff07d98b3-0653-4988-b6d2-1ac07e75edf7_1024x704.png 1272w, https://substackcdn.com/image/fetch/$s_!rsXx!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff07d98b3-0653-4988-b6d2-1ac07e75edf7_1024x704.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a></figure></div><p>For templates of arithmetic calculations shown in the above figure, the authors tested LLaMA2&#8211;7B and 13B models where input, for instance, would be &#8216;3 + 5 =&#8217; and the output would be &#8216;8&#8217;. The first step in their investigations is to understand which attention heads are affecting the prediction of token 8, the&nbsp;result.</p><p>Through a technique called <a href="https://arxiv.org/abs/2304.05969">path patching</a>, they observed that only &lt;5% of attention heads are important in performing arithmetic calculation. You can see in the figure below that there are very few heads or MLPs with darker&nbsp;colors.</p><p>They also observed that removing identified &#8220;arithmetic&#8221; heads affected the performance catastrophically compared to removing the same amount of random attention heads.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!ktxH!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F203193d9-918a-406c-bfcf-d4c74967d3f4_984x910.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!ktxH!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F203193d9-918a-406c-bfcf-d4c74967d3f4_984x910.png 424w, https://substackcdn.com/image/fetch/$s_!ktxH!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F203193d9-918a-406c-bfcf-d4c74967d3f4_984x910.png 848w, https://substackcdn.com/image/fetch/$s_!ktxH!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F203193d9-918a-406c-bfcf-d4c74967d3f4_984x910.png 1272w, https://substackcdn.com/image/fetch/$s_!ktxH!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F203193d9-918a-406c-bfcf-d4c74967d3f4_984x910.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!ktxH!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F203193d9-918a-406c-bfcf-d4c74967d3f4_984x910.png" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/203193d9-918a-406c-bfcf-d4c74967d3f4_984x910.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!ktxH!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F203193d9-918a-406c-bfcf-d4c74967d3f4_984x910.png 424w, https://substackcdn.com/image/fetch/$s_!ktxH!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F203193d9-918a-406c-bfcf-d4c74967d3f4_984x910.png 848w, https://substackcdn.com/image/fetch/$s_!ktxH!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F203193d9-918a-406c-bfcf-d4c74967d3f4_984x910.png 1272w, https://substackcdn.com/image/fetch/$s_!ktxH!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F203193d9-918a-406c-bfcf-d4c74967d3f4_984x910.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a></figure></div><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!-rVV!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7fc1351c-c4f8-49f4-bf32-0306cf1783ba_990x858.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!-rVV!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7fc1351c-c4f8-49f4-bf32-0306cf1783ba_990x858.png 424w, https://substackcdn.com/image/fetch/$s_!-rVV!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7fc1351c-c4f8-49f4-bf32-0306cf1783ba_990x858.png 848w, https://substackcdn.com/image/fetch/$s_!-rVV!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7fc1351c-c4f8-49f4-bf32-0306cf1783ba_990x858.png 1272w, https://substackcdn.com/image/fetch/$s_!-rVV!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7fc1351c-c4f8-49f4-bf32-0306cf1783ba_990x858.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!-rVV!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7fc1351c-c4f8-49f4-bf32-0306cf1783ba_990x858.png" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/7fc1351c-c4f8-49f4-bf32-0306cf1783ba_990x858.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!-rVV!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7fc1351c-c4f8-49f4-bf32-0306cf1783ba_990x858.png 424w, https://substackcdn.com/image/fetch/$s_!-rVV!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7fc1351c-c4f8-49f4-bf32-0306cf1783ba_990x858.png 848w, https://substackcdn.com/image/fetch/$s_!-rVV!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7fc1351c-c4f8-49f4-bf32-0306cf1783ba_990x858.png 1272w, https://substackcdn.com/image/fetch/$s_!-rVV!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7fc1351c-c4f8-49f4-bf32-0306cf1783ba_990x858.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">Figure 2: Only a few attention heads and MLPs affect the prediction. (darker color units) Figure 3: Performance drops when the identified &#8220;arithmetic&#8221; heads are removed as opposed to random attention heads</figcaption></figure></div><p>Interestingly, they observed specific attention heads that attend to operators (+,-,..) and operands (numbers). Figure 5 below shows attention weights being high at positions with operators and operands respectively.</p><p>The authors also investigated where the same &#8220;arithmetic&#8221; heads play a crucial role in mathematical reasoning in other datasets. Figure 4 below shows that after removing (knockout) the &#8220;arithmetic&#8221; heads the LLM produces wrong answers for data points that were predicted correctly before.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!gV3e!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff1392bf3-ba18-4bf5-9e7d-d9ce8bc72f8c_972x976.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!gV3e!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff1392bf3-ba18-4bf5-9e7d-d9ce8bc72f8c_972x976.png 424w, https://substackcdn.com/image/fetch/$s_!gV3e!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff1392bf3-ba18-4bf5-9e7d-d9ce8bc72f8c_972x976.png 848w, https://substackcdn.com/image/fetch/$s_!gV3e!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff1392bf3-ba18-4bf5-9e7d-d9ce8bc72f8c_972x976.png 1272w, https://substackcdn.com/image/fetch/$s_!gV3e!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff1392bf3-ba18-4bf5-9e7d-d9ce8bc72f8c_972x976.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!gV3e!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff1392bf3-ba18-4bf5-9e7d-d9ce8bc72f8c_972x976.png" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/f1392bf3-ba18-4bf5-9e7d-d9ce8bc72f8c_972x976.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!gV3e!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff1392bf3-ba18-4bf5-9e7d-d9ce8bc72f8c_972x976.png 424w, https://substackcdn.com/image/fetch/$s_!gV3e!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff1392bf3-ba18-4bf5-9e7d-d9ce8bc72f8c_972x976.png 848w, https://substackcdn.com/image/fetch/$s_!gV3e!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff1392bf3-ba18-4bf5-9e7d-d9ce8bc72f8c_972x976.png 1272w, https://substackcdn.com/image/fetch/$s_!gV3e!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff1392bf3-ba18-4bf5-9e7d-d9ce8bc72f8c_972x976.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a></figure></div><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!1WWy!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff9043beb-e107-4f42-aeb0-213279401092_966x722.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!1WWy!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff9043beb-e107-4f42-aeb0-213279401092_966x722.png 424w, https://substackcdn.com/image/fetch/$s_!1WWy!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff9043beb-e107-4f42-aeb0-213279401092_966x722.png 848w, https://substackcdn.com/image/fetch/$s_!1WWy!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff9043beb-e107-4f42-aeb0-213279401092_966x722.png 1272w, https://substackcdn.com/image/fetch/$s_!1WWy!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff9043beb-e107-4f42-aeb0-213279401092_966x722.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!1WWy!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff9043beb-e107-4f42-aeb0-213279401092_966x722.png" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/f9043beb-e107-4f42-aeb0-213279401092_966x722.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!1WWy!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff9043beb-e107-4f42-aeb0-213279401092_966x722.png 424w, https://substackcdn.com/image/fetch/$s_!1WWy!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff9043beb-e107-4f42-aeb0-213279401092_966x722.png 848w, https://substackcdn.com/image/fetch/$s_!1WWy!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff9043beb-e107-4f42-aeb0-213279401092_966x722.png 1272w, https://substackcdn.com/image/fetch/$s_!1WWy!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff9043beb-e107-4f42-aeb0-213279401092_966x722.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a></figure></div><p>Having identified the components that perform arithmetic reasoning, they proposed to fine-tune only the &#8220;arithmetic&#8221; heads (10% parameters) and improve the performance on mathematical reasoning tasks like GSM8k and SVAMP. Dubbed as Precise SFT in the below figure, it shows improvements compared to full SFT with 3x training speed. It is interesting to note that precise finetuning on mathematical tasks didn&#8217;t affect performance on generic tasks as much as full fine-tuning.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!KV2d!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1092e72c-e8f1-41e2-bbdb-5f136db07b1c_1024x494.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!KV2d!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1092e72c-e8f1-41e2-bbdb-5f136db07b1c_1024x494.png 424w, https://substackcdn.com/image/fetch/$s_!KV2d!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1092e72c-e8f1-41e2-bbdb-5f136db07b1c_1024x494.png 848w, https://substackcdn.com/image/fetch/$s_!KV2d!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1092e72c-e8f1-41e2-bbdb-5f136db07b1c_1024x494.png 1272w, https://substackcdn.com/image/fetch/$s_!KV2d!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1092e72c-e8f1-41e2-bbdb-5f136db07b1c_1024x494.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!KV2d!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1092e72c-e8f1-41e2-bbdb-5f136db07b1c_1024x494.png" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/1092e72c-e8f1-41e2-bbdb-5f136db07b1c_1024x494.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!KV2d!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1092e72c-e8f1-41e2-bbdb-5f136db07b1c_1024x494.png 424w, https://substackcdn.com/image/fetch/$s_!KV2d!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1092e72c-e8f1-41e2-bbdb-5f136db07b1c_1024x494.png 848w, https://substackcdn.com/image/fetch/$s_!KV2d!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1092e72c-e8f1-41e2-bbdb-5f136db07b1c_1024x494.png 1272w, https://substackcdn.com/image/fetch/$s_!KV2d!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1092e72c-e8f1-41e2-bbdb-5f136db07b1c_1024x494.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a></figure></div><h4><a href="https://openreview.net/forum?id=bX3J7ho18S">Monitoring AI-Modified Content at Scale: A Case Study on the Impact of ChatGPT on AI Conference Peer&nbsp;Reviews</a></h4><p>Motivated by a disproportionate increase in adjectives such as &#8220;intricate&#8221; &amp; &#8220;meticulous&#8221; in peer reviews, the authors devised a statistical parameter inference approach to estimate the ratio of AI-generated text at prominent deep learning conferences and Nature portfolio journals. As suspected, they estimated that 10.6% of ICLR 2024 review sentences and 16.9% of EMNLP are completely AI-generated.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!5f3d!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc68de04d-55b9-46ca-8fda-27514b671981_812x732.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!5f3d!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc68de04d-55b9-46ca-8fda-27514b671981_812x732.png 424w, https://substackcdn.com/image/fetch/$s_!5f3d!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc68de04d-55b9-46ca-8fda-27514b671981_812x732.png 848w, https://substackcdn.com/image/fetch/$s_!5f3d!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc68de04d-55b9-46ca-8fda-27514b671981_812x732.png 1272w, https://substackcdn.com/image/fetch/$s_!5f3d!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc68de04d-55b9-46ca-8fda-27514b671981_812x732.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!5f3d!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc68de04d-55b9-46ca-8fda-27514b671981_812x732.png" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/c68de04d-55b9-46ca-8fda-27514b671981_812x732.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!5f3d!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc68de04d-55b9-46ca-8fda-27514b671981_812x732.png 424w, https://substackcdn.com/image/fetch/$s_!5f3d!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc68de04d-55b9-46ca-8fda-27514b671981_812x732.png 848w, https://substackcdn.com/image/fetch/$s_!5f3d!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc68de04d-55b9-46ca-8fda-27514b671981_812x732.png 1272w, https://substackcdn.com/image/fetch/$s_!5f3d!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc68de04d-55b9-46ca-8fda-27514b671981_812x732.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a></figure></div><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!CSBl!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F62a453af-fc39-4a02-a5de-c70cd58a7297_820x734.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!CSBl!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F62a453af-fc39-4a02-a5de-c70cd58a7297_820x734.png 424w, https://substackcdn.com/image/fetch/$s_!CSBl!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F62a453af-fc39-4a02-a5de-c70cd58a7297_820x734.png 848w, https://substackcdn.com/image/fetch/$s_!CSBl!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F62a453af-fc39-4a02-a5de-c70cd58a7297_820x734.png 1272w, https://substackcdn.com/image/fetch/$s_!CSBl!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F62a453af-fc39-4a02-a5de-c70cd58a7297_820x734.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!CSBl!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F62a453af-fc39-4a02-a5de-c70cd58a7297_820x734.png" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/62a453af-fc39-4a02-a5de-c70cd58a7297_820x734.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!CSBl!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F62a453af-fc39-4a02-a5de-c70cd58a7297_820x734.png 424w, https://substackcdn.com/image/fetch/$s_!CSBl!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F62a453af-fc39-4a02-a5de-c70cd58a7297_820x734.png 848w, https://substackcdn.com/image/fetch/$s_!CSBl!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F62a453af-fc39-4a02-a5de-c70cd58a7297_820x734.png 1272w, https://substackcdn.com/image/fetch/$s_!CSBl!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F62a453af-fc39-4a02-a5de-c70cd58a7297_820x734.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a></figure></div><p>The paper started by explaining the extreme difficulty of identifying LLM-generated content instance by instance. For example, techniques like zero-shot LLM detection, fine-tuned models for classifying LLM-generated content, or LLM watermarking classify each document as LLM-generated or not. However, these techniques are shown to be similar to random predictors or such mechanisms reduce the coherence of LLM generation.</p><p>Alternatively, the authors proposed to estimate the ratio of AI-generated sentences/documents in a corpus rather than classifying them at the instance level. They assumed that a corpus would be generated from the mixture distribution of P (human-generated) and Q (LLM-generated) as in the below figure, with alpha as the ratio of LLM-generated content. The log-likelihood of any corpus could be formulated with an alpha parameter like&nbsp;below.</p><p>The authors estimated the true P and Q distribution with training corpora with a known ratio of human vs AI-generated content (alpha). Knowing the P and Q distribution, we could then estimate the alpha for any corpus by maximum likelihood estimation (MLE). (In simple terms, perturb the alpha many times to get the right alpha that fits the given&nbsp;corpus).</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!ab8p!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fddecebd7-8073-4051-89bb-b64df1bf8cad_800x356.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!ab8p!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fddecebd7-8073-4051-89bb-b64df1bf8cad_800x356.png 424w, https://substackcdn.com/image/fetch/$s_!ab8p!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fddecebd7-8073-4051-89bb-b64df1bf8cad_800x356.png 848w, https://substackcdn.com/image/fetch/$s_!ab8p!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fddecebd7-8073-4051-89bb-b64df1bf8cad_800x356.png 1272w, https://substackcdn.com/image/fetch/$s_!ab8p!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fddecebd7-8073-4051-89bb-b64df1bf8cad_800x356.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!ab8p!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fddecebd7-8073-4051-89bb-b64df1bf8cad_800x356.png" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/ddecebd7-8073-4051-89bb-b64df1bf8cad_800x356.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!ab8p!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fddecebd7-8073-4051-89bb-b64df1bf8cad_800x356.png 424w, https://substackcdn.com/image/fetch/$s_!ab8p!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fddecebd7-8073-4051-89bb-b64df1bf8cad_800x356.png 848w, https://substackcdn.com/image/fetch/$s_!ab8p!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fddecebd7-8073-4051-89bb-b64df1bf8cad_800x356.png 1272w, https://substackcdn.com/image/fetch/$s_!ab8p!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fddecebd7-8073-4051-89bb-b64df1bf8cad_800x356.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a></figure></div><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!W7Fg!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1f93a12b-fc0b-4933-83a4-18579fe3c8f6_804x436.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!W7Fg!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1f93a12b-fc0b-4933-83a4-18579fe3c8f6_804x436.png 424w, https://substackcdn.com/image/fetch/$s_!W7Fg!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1f93a12b-fc0b-4933-83a4-18579fe3c8f6_804x436.png 848w, https://substackcdn.com/image/fetch/$s_!W7Fg!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1f93a12b-fc0b-4933-83a4-18579fe3c8f6_804x436.png 1272w, https://substackcdn.com/image/fetch/$s_!W7Fg!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1f93a12b-fc0b-4933-83a4-18579fe3c8f6_804x436.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!W7Fg!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1f93a12b-fc0b-4933-83a4-18579fe3c8f6_804x436.png" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/1f93a12b-fc0b-4933-83a4-18579fe3c8f6_804x436.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!W7Fg!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1f93a12b-fc0b-4933-83a4-18579fe3c8f6_804x436.png 424w, https://substackcdn.com/image/fetch/$s_!W7Fg!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1f93a12b-fc0b-4933-83a4-18579fe3c8f6_804x436.png 848w, https://substackcdn.com/image/fetch/$s_!W7Fg!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1f93a12b-fc0b-4933-83a4-18579fe3c8f6_804x436.png 1272w, https://substackcdn.com/image/fetch/$s_!W7Fg!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1f93a12b-fc0b-4933-83a4-18579fe3c8f6_804x436.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a></figure></div><p>The entire process is depicted in the following figure: 1, Create human and LLM-generated corpus 2, Temporally split training corpus and validation corpus 3, Estimate the true P and Q distributions with training corpus 4, Validate the estimated distributions on validation corpus where alpha is known 5, Finally, estimate the alpha on the corpus of interest, here ICLR 2024 or NeurIPS 2023 and so on. This gives us the ratio of LLM-generated sentences/documents.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!3EH9!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F197cb820-b83f-4c64-8b08-c868e6a10b17_1024x516.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!3EH9!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F197cb820-b83f-4c64-8b08-c868e6a10b17_1024x516.png 424w, https://substackcdn.com/image/fetch/$s_!3EH9!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F197cb820-b83f-4c64-8b08-c868e6a10b17_1024x516.png 848w, https://substackcdn.com/image/fetch/$s_!3EH9!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F197cb820-b83f-4c64-8b08-c868e6a10b17_1024x516.png 1272w, https://substackcdn.com/image/fetch/$s_!3EH9!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F197cb820-b83f-4c64-8b08-c868e6a10b17_1024x516.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!3EH9!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F197cb820-b83f-4c64-8b08-c868e6a10b17_1024x516.png" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/197cb820-b83f-4c64-8b08-c868e6a10b17_1024x516.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!3EH9!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F197cb820-b83f-4c64-8b08-c868e6a10b17_1024x516.png 424w, https://substackcdn.com/image/fetch/$s_!3EH9!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F197cb820-b83f-4c64-8b08-c868e6a10b17_1024x516.png 848w, https://substackcdn.com/image/fetch/$s_!3EH9!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F197cb820-b83f-4c64-8b08-c868e6a10b17_1024x516.png 1272w, https://substackcdn.com/image/fetch/$s_!3EH9!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F197cb820-b83f-4c64-8b08-c868e6a10b17_1024x516.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a></figure></div><p>The paper also has interesting correlations between the ratio of LLM-generated content and different aspects of peer review. The estimated fraction of LLM-generated text is higher in reviews that don&#8217;t have scholarly citations (reference effect) and from reviewers who are less likely to respond to author rebuttals (lower reply rate effect). The LLM-generated text also tends to be homogeneous, which reduces the value-add of multiple reviews per paper (homogenization effect).</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!JnJh!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8638059d-9e5a-4f3a-a2e9-5bb3f0659a1a_824x534.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!JnJh!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8638059d-9e5a-4f3a-a2e9-5bb3f0659a1a_824x534.png 424w, https://substackcdn.com/image/fetch/$s_!JnJh!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8638059d-9e5a-4f3a-a2e9-5bb3f0659a1a_824x534.png 848w, https://substackcdn.com/image/fetch/$s_!JnJh!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8638059d-9e5a-4f3a-a2e9-5bb3f0659a1a_824x534.png 1272w, https://substackcdn.com/image/fetch/$s_!JnJh!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8638059d-9e5a-4f3a-a2e9-5bb3f0659a1a_824x534.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!JnJh!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8638059d-9e5a-4f3a-a2e9-5bb3f0659a1a_824x534.png" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/8638059d-9e5a-4f3a-a2e9-5bb3f0659a1a_824x534.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!JnJh!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8638059d-9e5a-4f3a-a2e9-5bb3f0659a1a_824x534.png 424w, https://substackcdn.com/image/fetch/$s_!JnJh!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8638059d-9e5a-4f3a-a2e9-5bb3f0659a1a_824x534.png 848w, https://substackcdn.com/image/fetch/$s_!JnJh!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8638059d-9e5a-4f3a-a2e9-5bb3f0659a1a_824x534.png 1272w, https://substackcdn.com/image/fetch/$s_!JnJh!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8638059d-9e5a-4f3a-a2e9-5bb3f0659a1a_824x534.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a></figure></div><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!waCH!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F75295116-82e3-4fe0-b28b-449430c705f4_800x546.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!waCH!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F75295116-82e3-4fe0-b28b-449430c705f4_800x546.png 424w, https://substackcdn.com/image/fetch/$s_!waCH!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F75295116-82e3-4fe0-b28b-449430c705f4_800x546.png 848w, https://substackcdn.com/image/fetch/$s_!waCH!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F75295116-82e3-4fe0-b28b-449430c705f4_800x546.png 1272w, https://substackcdn.com/image/fetch/$s_!waCH!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F75295116-82e3-4fe0-b28b-449430c705f4_800x546.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!waCH!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F75295116-82e3-4fe0-b28b-449430c705f4_800x546.png" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/75295116-82e3-4fe0-b28b-449430c705f4_800x546.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!waCH!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F75295116-82e3-4fe0-b28b-449430c705f4_800x546.png 424w, https://substackcdn.com/image/fetch/$s_!waCH!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F75295116-82e3-4fe0-b28b-449430c705f4_800x546.png 848w, https://substackcdn.com/image/fetch/$s_!waCH!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F75295116-82e3-4fe0-b28b-449430c705f4_800x546.png 1272w, https://substackcdn.com/image/fetch/$s_!waCH!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F75295116-82e3-4fe0-b28b-449430c705f4_800x546.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a></figure></div><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!gbwj!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1d5cfb9d-8901-4567-8ad0-bbcfff02cea6_798x578.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!gbwj!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1d5cfb9d-8901-4567-8ad0-bbcfff02cea6_798x578.png 424w, https://substackcdn.com/image/fetch/$s_!gbwj!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1d5cfb9d-8901-4567-8ad0-bbcfff02cea6_798x578.png 848w, https://substackcdn.com/image/fetch/$s_!gbwj!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1d5cfb9d-8901-4567-8ad0-bbcfff02cea6_798x578.png 1272w, https://substackcdn.com/image/fetch/$s_!gbwj!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1d5cfb9d-8901-4567-8ad0-bbcfff02cea6_798x578.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!gbwj!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1d5cfb9d-8901-4567-8ad0-bbcfff02cea6_798x578.png" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/1d5cfb9d-8901-4567-8ad0-bbcfff02cea6_798x578.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!gbwj!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1d5cfb9d-8901-4567-8ad0-bbcfff02cea6_798x578.png 424w, https://substackcdn.com/image/fetch/$s_!gbwj!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1d5cfb9d-8901-4567-8ad0-bbcfff02cea6_798x578.png 848w, https://substackcdn.com/image/fetch/$s_!gbwj!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1d5cfb9d-8901-4567-8ad0-bbcfff02cea6_798x578.png 1272w, https://substackcdn.com/image/fetch/$s_!gbwj!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1d5cfb9d-8901-4567-8ad0-bbcfff02cea6_798x578.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a></figure></div><p>The paper also has a nice Box with all the main findings.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!8a3B!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcdd8d579-eef3-4cd7-9930-6edc3af4ee96_798x904.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!8a3B!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcdd8d579-eef3-4cd7-9930-6edc3af4ee96_798x904.png 424w, https://substackcdn.com/image/fetch/$s_!8a3B!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcdd8d579-eef3-4cd7-9930-6edc3af4ee96_798x904.png 848w, https://substackcdn.com/image/fetch/$s_!8a3B!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcdd8d579-eef3-4cd7-9930-6edc3af4ee96_798x904.png 1272w, https://substackcdn.com/image/fetch/$s_!8a3B!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcdd8d579-eef3-4cd7-9930-6edc3af4ee96_798x904.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!8a3B!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcdd8d579-eef3-4cd7-9930-6edc3af4ee96_798x904.png" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/cdd8d579-eef3-4cd7-9930-6edc3af4ee96_798x904.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!8a3B!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcdd8d579-eef3-4cd7-9930-6edc3af4ee96_798x904.png 424w, https://substackcdn.com/image/fetch/$s_!8a3B!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcdd8d579-eef3-4cd7-9930-6edc3af4ee96_798x904.png 848w, https://substackcdn.com/image/fetch/$s_!8a3B!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcdd8d579-eef3-4cd7-9930-6edc3af4ee96_798x904.png 1272w, https://substackcdn.com/image/fetch/$s_!8a3B!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcdd8d579-eef3-4cd7-9930-6edc3af4ee96_798x904.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a></figure></div><h4>Other interesting papers:</h4><h4><a href="https://openreview.net/forum?id=frA0NNBS1n">Probabilistic Inference in Language Models via Twisted Sequential Monte&nbsp;Carlo</a></h4><h4><a href="https://openreview.net/forum?id=QBj7Uurdwf">Learning Useful Representations of Recurrent Neural Network Weight&nbsp;Matrices</a></h4><h4><a href="https://openreview.net/forum?id=BwAkaxqiLB">Evolution of Heuristics: Towards Efficient Automatic Algorithm Design Using Large Language&nbsp;Model</a></h4><h4><a href="https://openreview.net/forum?id=jKYyFbH8ap">Evaluation of LLMs on Syntax-Aware Code Fill-in-the-Middle Tasks</a></h4><h4><a href="https://openreview.net/forum?id=hYHsrKDiX7">GaLore: Memory-Efficient LLM Training by Gradient Low-Rank Projection</a></h4><h4><a href="https://openreview.net/forum?id=uDkXoZMzBv">Compressible Dynamics in Deep Overparameterized Low-Rank Learning &amp; Adaptation</a></h4><h4><a href="https://openreview.net/forum?id=MFPYCvWsNR">Bottleneck-Minimal Indexing for Generative Document Retrieval</a></h4><h4><a href="https://openreview.net/forum?id=685vj0lC9z">How do Large Language Models Navigate Conflicts between Honesty and Helpfulness?</a></h4><h4><a href="https://openreview.net/forum?id=oowQ8LPA12">Theoretical Analysis of Learned Database Operations under Distribution Shift through Distribution Learnability</a></h4><h4><a href="https://openreview.net/forum?id=6XH8R7YrSk">Is DPO Superior to PPO for LLM Alignment? A Comprehensive Study</a></h4><h4><a href="https://openreview.net/forum?id=gPStP3FSY9">Discovering Environments with&nbsp;XRM</a></h4><h4><a href="https://openreview.net/forum?id=MdPBVWTfwG">I/O Complexity of Attention, or How Optimal is FlashAttention?</a></h4><h4><a href="https://openreview.net/forum?id=byxXa99PtF">Decomposing Uncertainty for Large Language Models through Input Clarification Ensembling</a></h4><h4><a href="https://openreview.net/forum?id=hlvKd7Vdxm">ExCP: Extreme LLM Checkpoint Compression via Weight-Momentum Joint Shrinking</a></h4><h4><a href="https://openreview.net/forum?id=wGtzp4ZT1n">CompeteAI: Understanding the Competition Dynamics of Large Language Model-based Agents</a></h4><h4><a href="https://openreview.net/forum?id=moyG54Okrj">Repoformer: Selective Retrieval for Repository-Level Code Completion</a></h4><h4><a href="https://openreview.net/forum?id=FPnUhsQJ5B">Scaling Rectified Flow Transformers for High-Resolution Image Synthesis</a></h4><h4><a href="https://openreview.net/forum?id=4jqOV6NlUz">Automated Evaluation of Retrieval-Augmented Language Models with Task-Specific Exam Generation</a></h4><h4><a href="https://openreview.net/forum?id=p225Od0aYt">PRISE: LLM-Style Sequence Compression for Learning Temporal Action Abstractions in&nbsp;Control</a></h4><h4><a href="https://openreview.net/forum?id=HPXRzM9BYZ">LCA-on-the-Line: Benchmarking Out of Distribution Generalization with Class Taxonomies</a></h4><h4><a href="https://openreview.net/forum?id=7dP6Yq9Uwv">Learning to Model the World With&nbsp;Language</a></h4><h4><a href="https://openreview.net/forum?id=KviM5k8pcP">AI Control: Improving Safety Despite Intentional Subversion</a></h4><h4><a href="https://openreview.net/forum?id=tl2qmO5kpD">Offline Actor-Critic Reinforcement Learning Scales to Large&nbsp;Models</a></h4><h4><a href="https://openreview.net/forum?id=Zc22RDtsvP">MagicLens: Self-Supervised Image Retrieval with Open-Ended Instructions</a></h4><h4><a href="https://openreview.net/forum?id=VE3yWXt3KB">Stealing part of a production language&nbsp;model</a></h4><h4><a href="https://openreview.net/forum?id=iLCZtl7FTa">Debating with More Persuasive LLMs Leads to More Truthful&nbsp;Answers</a></h4><h4><a href="https://openreview.net/forum?id=JSYN891WnB">Image Clustering with External&nbsp;Guidance</a></h4><h4><a href="https://openreview.net/forum?id=dBqHGZPGZI">A Mechanistic Understanding of Alignment Algorithms: A Case Study on DPO and&nbsp;Toxicity</a></h4><h4><a href="https://openreview.net/forum?id=bJbSbJskOS">Genie: Generative Interactive Environments</a></h4><h4><a href="https://openreview.net/forum?id=qz1Vx1v9iK">Test-Time Model Adaptation with Only Forward&nbsp;Passes</a></h4><h4><a href="https://openreview.net/forum?id=uTC9AFXIhg">GPTSwarm: Language Agents as Optimizable Graphs</a></h4>]]></content:encoded></item><item><title><![CDATA[Knowledge Distillation Research Review]]></title><description><![CDATA[The latest advances and insights for Neural Network model compression with Knowledge Distillation]]></description><link>https://www.aipapertrails.com/p/knowledge-distillation-research-review-2020-e0333f554863</link><guid isPermaLink="false">https://www.aipapertrails.com/p/knowledge-distillation-research-review-2020-e0333f554863</guid><dc:creator><![CDATA[Prakash Kagitha]]></dc:creator><pubDate>Fri, 05 Feb 2021 14:46:44 GMT</pubDate><enclosure url="https://cdn-images-1.medium.com/max/934/1*hJT_wRUdsPrL0WKpoVcSBg.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<h4>The latest advances and insights for Neural Network model compression with Knowledge Distillation</h4><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!NtfD!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff68f1e0c-4ccd-4be7-b14e-38f21a097859_934x343.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!NtfD!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff68f1e0c-4ccd-4be7-b14e-38f21a097859_934x343.png 424w, https://substackcdn.com/image/fetch/$s_!NtfD!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff68f1e0c-4ccd-4be7-b14e-38f21a097859_934x343.png 848w, https://substackcdn.com/image/fetch/$s_!NtfD!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff68f1e0c-4ccd-4be7-b14e-38f21a097859_934x343.png 1272w, https://substackcdn.com/image/fetch/$s_!NtfD!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff68f1e0c-4ccd-4be7-b14e-38f21a097859_934x343.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!NtfD!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff68f1e0c-4ccd-4be7-b14e-38f21a097859_934x343.png" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/f68f1e0c-4ccd-4be7-b14e-38f21a097859_934x343.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!NtfD!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff68f1e0c-4ccd-4be7-b14e-38f21a097859_934x343.png 424w, https://substackcdn.com/image/fetch/$s_!NtfD!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff68f1e0c-4ccd-4be7-b14e-38f21a097859_934x343.png 848w, https://substackcdn.com/image/fetch/$s_!NtfD!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff68f1e0c-4ccd-4be7-b14e-38f21a097859_934x343.png 1272w, https://substackcdn.com/image/fetch/$s_!NtfD!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff68f1e0c-4ccd-4be7-b14e-38f21a097859_934x343.png 1456w" sizes="100vw" fetchpriority="high"></picture><div></div></div></a><figcaption class="image-caption">Image reproduced from <a href="https://nips.cc/virtual/2020/public/poster_912d2b1c7b2826caf99687388d2e8f7c.html">the&nbsp;paper.</a></figcaption></figure></div><p>Knowledge Distillation is a process where a smaller/less complex model is trained to imitate the behavior of a larger/more complex&nbsp;model.</p><p>Particularly when deploying NN models on mobiles or edge devices, Pruning, and model compression in general, is desirable and often the only plausible way to deploy as the memory and computational budget of these devices is very&nbsp;limited.</p><p>Why not use potential infinite virtual memory and computational power from cloud machines? While a lot of NN models are running on the cloud even now, latency is not low enough for mobile/edge devices, which hinders utility and requires data to be transferred to the cloud which rises a lot of privacy concerns.</p><h4>Preliminaries of Neural Network Knowledge Distillation</h4><p><strong>Response (Output logits) Distillation vs. Feature Distillation vs. Relation Distillation:</strong> This distinction is based on the knowledge that is transferred from a teacher network to a student network. If a student network is trained to reproduce the output logits distribution it is called Response Distillation. Instead, if a student network is trained to reproduce the intermediate representations, then it is called Feature Distillation. If a student network is trained to reproduce the relative response on a pair of inputs it is called Relation Distillation.</p><p><strong>Offline Distillation vs. Online Distillation:</strong> If the teacher model is already trained and a student model is trained after transferring the knowledge from the teacher, then it is called Offline Distillation. If both the teacher and student models are trained in parallel, then it is called Online Distillation. (Online Distillation could be between models of the same size and same complexity where there are no teachers or students).</p><p><strong>Distilling from ensembles of models:</strong> In this paradigm, a single model is obtained distilling the knowledge from an ensemble of models to lower the inference costs.</p><p><strong>Self-distillation:</strong> Sometimes a model is trained to mimic the feature outputs of the Nth-layer at (N/3)th-layer or at (N/2)th-layer. Now, only N/2 or N/3 layers of the model are sufficient to produce the results of a full N-layer network. This can be seen as distilling a smaller model from a bigger model which is&nbsp;itself.</p><h4><a href="https://iclr.cc/virtual_2020/poster_SkgpBJrtvS.html">Contrastive Representation Distillation</a> (ICLR&nbsp;2020)</h4><p><strong>Outstanding problem:</strong> Instead of training a student model to imitate just the class independent output probabilities of a teacher model, training to imitate the representations of the teacher model transfers more knowledge from the teacher model to a student&nbsp;model.</p><p>Even though previous methods dealt with transferring representations of the teacher model, their loss functions are not designed to model the correlations and higher-order dependencies in the representational space using just a dot product between representations. This constrains the feature vectors of the student and the teacher to be of the same size, which doesn&#8217;t allow the possibility of having smaller student networks without any constraints on their architecture, which is most&nbsp;desired.</p><p>This work employs a contrastive learning approach based on mutual information to learn correlations in the representational space. The similarity between representations is estimated with a critic model that is trained along with the student model which takes two representations and outputs a similarity between 0 to 1, making it possible to have different sizes of feature vectors in teacher and student networks.</p><p><strong>Proposed solution</strong>: Instead of just maximizing the dot product between the representations of a teacher model and a student model or minimizing L1 between them, this work proposes that we use mutual information as the criterion to maximize.</p><p>Mutual information is a measure between two variables that tells how much information is present in one variable about the&nbsp;other.</p><p>Adopting a contrastive learning framework, the authors train positive pairs which are student and teacher representations for the same input by increasing the mutual information between those representations. They also trained negative pairs which are student and teacher representations for different inputs by decreasing the mutual information between those representations.</p><p><strong>Results and Conclusions:</strong> It is shown that the output probabilities of the student network trained with the proposed methods achieve higher correlations with the output probabilities of the teacher&nbsp;network.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!JzgM!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e3bb186-2d55-4b8b-b216-50af71a90ad5_1024x504.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!JzgM!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e3bb186-2d55-4b8b-b216-50af71a90ad5_1024x504.png 424w, https://substackcdn.com/image/fetch/$s_!JzgM!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e3bb186-2d55-4b8b-b216-50af71a90ad5_1024x504.png 848w, https://substackcdn.com/image/fetch/$s_!JzgM!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e3bb186-2d55-4b8b-b216-50af71a90ad5_1024x504.png 1272w, https://substackcdn.com/image/fetch/$s_!JzgM!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e3bb186-2d55-4b8b-b216-50af71a90ad5_1024x504.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!JzgM!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e3bb186-2d55-4b8b-b216-50af71a90ad5_1024x504.png" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/5e3bb186-2d55-4b8b-b216-50af71a90ad5_1024x504.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!JzgM!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e3bb186-2d55-4b8b-b216-50af71a90ad5_1024x504.png 424w, https://substackcdn.com/image/fetch/$s_!JzgM!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e3bb186-2d55-4b8b-b216-50af71a90ad5_1024x504.png 848w, https://substackcdn.com/image/fetch/$s_!JzgM!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e3bb186-2d55-4b8b-b216-50af71a90ad5_1024x504.png 1272w, https://substackcdn.com/image/fetch/$s_!JzgM!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e3bb186-2d55-4b8b-b216-50af71a90ad5_1024x504.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">Image reproduced from <a href="https://iclr.cc/virtual_2020/poster_SkgpBJrtvS.html">the&nbsp;paper.</a></figcaption></figure></div><p>The proposed method outperforms many of the recently proposed Knowledge Distillations methods.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://cdn-images-1.medium.com/max/1024/1*TeRjTiXrCn0Gb6xIBDSijw.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://cdn-images-1.medium.com/max/1024/1*TeRjTiXrCn0Gb6xIBDSijw.png 424w, https://cdn-images-1.medium.com/max/1024/1*TeRjTiXrCn0Gb6xIBDSijw.png 848w, https://cdn-images-1.medium.com/max/1024/1*TeRjTiXrCn0Gb6xIBDSijw.png 1272w, https://cdn-images-1.medium.com/max/1024/1*TeRjTiXrCn0Gb6xIBDSijw.png 1456w" sizes="100vw"><img src="https://cdn-images-1.medium.com/max/1024/1*TeRjTiXrCn0Gb6xIBDSijw.png" data-attrs="{&quot;src&quot;:&quot;https://cdn-images-1.medium.com/max/1024/1*TeRjTiXrCn0Gb6xIBDSijw.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://cdn-images-1.medium.com/max/1024/1*TeRjTiXrCn0Gb6xIBDSijw.png 424w, https://cdn-images-1.medium.com/max/1024/1*TeRjTiXrCn0Gb6xIBDSijw.png 848w, https://cdn-images-1.medium.com/max/1024/1*TeRjTiXrCn0Gb6xIBDSijw.png 1272w, https://cdn-images-1.medium.com/max/1024/1*TeRjTiXrCn0Gb6xIBDSijw.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">Image reproduced from <a href="https://iclr.cc/virtual_2020/poster_SkgpBJrtvS.html">the&nbsp;paper</a>.</figcaption></figure></div><p>Also, the proposed method outperforms all other methods in cases where the architectures of the teacher and student models are very different.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://cdn-images-1.medium.com/max/1024/1*ncJ432YLBiAshwWVFYe6FQ.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://cdn-images-1.medium.com/max/1024/1*ncJ432YLBiAshwWVFYe6FQ.png 424w, https://cdn-images-1.medium.com/max/1024/1*ncJ432YLBiAshwWVFYe6FQ.png 848w, https://cdn-images-1.medium.com/max/1024/1*ncJ432YLBiAshwWVFYe6FQ.png 1272w, https://cdn-images-1.medium.com/max/1024/1*ncJ432YLBiAshwWVFYe6FQ.png 1456w" sizes="100vw"><img src="https://cdn-images-1.medium.com/max/1024/1*ncJ432YLBiAshwWVFYe6FQ.png" data-attrs="{&quot;src&quot;:&quot;https://cdn-images-1.medium.com/max/1024/1*ncJ432YLBiAshwWVFYe6FQ.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://cdn-images-1.medium.com/max/1024/1*ncJ432YLBiAshwWVFYe6FQ.png 424w, https://cdn-images-1.medium.com/max/1024/1*ncJ432YLBiAshwWVFYe6FQ.png 848w, https://cdn-images-1.medium.com/max/1024/1*ncJ432YLBiAshwWVFYe6FQ.png 1272w, https://cdn-images-1.medium.com/max/1024/1*ncJ432YLBiAshwWVFYe6FQ.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">Image reproduced from <a href="https://iclr.cc/virtual_2020/poster_SkgpBJrtvS.html">the&nbsp;paper</a>.</figcaption></figure></div><h4><a href="https://icml.cc/virtual/2020/poster/5785">Feature-map-level Online Adversarial Knowledge Distillation</a> (ICML&nbsp;2020)</h4><p><strong>Outstanding problem:</strong> Even though it is established that transferring knowledge from feature maps is more efficient than just transferring the knowledge of output logit distributions, no method exists to transfer the feature-map-level in an online knowledge distillation setting (where multiple models are trained simultaneously to predict the ground truth label and also to mimic each other&#8217;s behavior).</p><p>It is particularly hard to transfer the feature-map-level knowledge because the feature maps change more rapidly than the output logits and pose a major problem to distillation methods having to transfer representations from this moving&nbsp;target.</p><p>The current method overcomes this challenge with adversarial training by trying to make the representational distributions similar rather than individual representations.</p><p><strong>Proposed solution:</strong> Given two networks, this work proposes three types of losses in the entire objective. 1. Cross-entropy between predicted and ground truth labels for each network. 2. KL divergence losses between the output logits of the two networks (both ways). 3. Discriminator loss where the representations of a particular network are treated as fake and representations from the other network are treated as&nbsp;real.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://cdn-images-1.medium.com/max/1024/1*YOE3xhyVQyDLwVoVJ-Gqmw.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://cdn-images-1.medium.com/max/1024/1*YOE3xhyVQyDLwVoVJ-Gqmw.png 424w, https://cdn-images-1.medium.com/max/1024/1*YOE3xhyVQyDLwVoVJ-Gqmw.png 848w, https://cdn-images-1.medium.com/max/1024/1*YOE3xhyVQyDLwVoVJ-Gqmw.png 1272w, https://cdn-images-1.medium.com/max/1024/1*YOE3xhyVQyDLwVoVJ-Gqmw.png 1456w" sizes="100vw"><img src="https://cdn-images-1.medium.com/max/1024/1*YOE3xhyVQyDLwVoVJ-Gqmw.png" data-attrs="{&quot;src&quot;:&quot;https://cdn-images-1.medium.com/max/1024/1*YOE3xhyVQyDLwVoVJ-Gqmw.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://cdn-images-1.medium.com/max/1024/1*YOE3xhyVQyDLwVoVJ-Gqmw.png 424w, https://cdn-images-1.medium.com/max/1024/1*YOE3xhyVQyDLwVoVJ-Gqmw.png 848w, https://cdn-images-1.medium.com/max/1024/1*YOE3xhyVQyDLwVoVJ-Gqmw.png 1272w, https://cdn-images-1.medium.com/max/1024/1*YOE3xhyVQyDLwVoVJ-Gqmw.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">Image reproduced from <a href="https://icml.cc/virtual/2020/poster/5785">the&nbsp;paper</a>.</figcaption></figure></div><p>Cross-entropy loss brings in the information from the dataset. KL divergence acts as the standard knowledge distillation of output logits. And, finally, the discriminator losses would help the network to learn from other networks even though the actual representations are changing by pushing for the entire distributions to be&nbsp;similar.</p><p><strong>Results and Conclusions:</strong> The proposed method outperforms the direct alignment-based methods&nbsp;(L1).</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://cdn-images-1.medium.com/max/1024/1*9O0DNcn8hlq2Tai0-WTANw.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://cdn-images-1.medium.com/max/1024/1*9O0DNcn8hlq2Tai0-WTANw.png 424w, https://cdn-images-1.medium.com/max/1024/1*9O0DNcn8hlq2Tai0-WTANw.png 848w, https://cdn-images-1.medium.com/max/1024/1*9O0DNcn8hlq2Tai0-WTANw.png 1272w, https://cdn-images-1.medium.com/max/1024/1*9O0DNcn8hlq2Tai0-WTANw.png 1456w" sizes="100vw"><img src="https://cdn-images-1.medium.com/max/1024/1*9O0DNcn8hlq2Tai0-WTANw.png" data-attrs="{&quot;src&quot;:&quot;https://cdn-images-1.medium.com/max/1024/1*9O0DNcn8hlq2Tai0-WTANw.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://cdn-images-1.medium.com/max/1024/1*9O0DNcn8hlq2Tai0-WTANw.png 424w, https://cdn-images-1.medium.com/max/1024/1*9O0DNcn8hlq2Tai0-WTANw.png 848w, https://cdn-images-1.medium.com/max/1024/1*9O0DNcn8hlq2Tai0-WTANw.png 1272w, https://cdn-images-1.medium.com/max/1024/1*9O0DNcn8hlq2Tai0-WTANw.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">Image reproduced from <a href="https://icml.cc/virtual/2020/poster/5785">the&nbsp;paper</a>.</figcaption></figure></div><p>The proposed method also outperforms the previous online distillation methods in both cases with the same networks and different networks.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://cdn-images-1.medium.com/max/1024/1*Mh0S5Z7LOfA14ziCbaNy8Q.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://cdn-images-1.medium.com/max/1024/1*Mh0S5Z7LOfA14ziCbaNy8Q.png 424w, https://cdn-images-1.medium.com/max/1024/1*Mh0S5Z7LOfA14ziCbaNy8Q.png 848w, https://cdn-images-1.medium.com/max/1024/1*Mh0S5Z7LOfA14ziCbaNy8Q.png 1272w, https://cdn-images-1.medium.com/max/1024/1*Mh0S5Z7LOfA14ziCbaNy8Q.png 1456w" sizes="100vw"><img src="https://cdn-images-1.medium.com/max/1024/1*Mh0S5Z7LOfA14ziCbaNy8Q.png" data-attrs="{&quot;src&quot;:&quot;https://cdn-images-1.medium.com/max/1024/1*Mh0S5Z7LOfA14ziCbaNy8Q.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://cdn-images-1.medium.com/max/1024/1*Mh0S5Z7LOfA14ziCbaNy8Q.png 424w, https://cdn-images-1.medium.com/max/1024/1*Mh0S5Z7LOfA14ziCbaNy8Q.png 848w, https://cdn-images-1.medium.com/max/1024/1*Mh0S5Z7LOfA14ziCbaNy8Q.png 1272w, https://cdn-images-1.medium.com/max/1024/1*Mh0S5Z7LOfA14ziCbaNy8Q.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">Image reproduced from <a href="https://icml.cc/virtual/2020/poster/5785">the&nbsp;paper</a>.</figcaption></figure></div><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://cdn-images-1.medium.com/max/1024/1*cL_22SPlCBo8LOWfNEokbg.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://cdn-images-1.medium.com/max/1024/1*cL_22SPlCBo8LOWfNEokbg.png 424w, https://cdn-images-1.medium.com/max/1024/1*cL_22SPlCBo8LOWfNEokbg.png 848w, https://cdn-images-1.medium.com/max/1024/1*cL_22SPlCBo8LOWfNEokbg.png 1272w, https://cdn-images-1.medium.com/max/1024/1*cL_22SPlCBo8LOWfNEokbg.png 1456w" sizes="100vw"><img src="https://cdn-images-1.medium.com/max/1024/1*cL_22SPlCBo8LOWfNEokbg.png" data-attrs="{&quot;src&quot;:&quot;https://cdn-images-1.medium.com/max/1024/1*cL_22SPlCBo8LOWfNEokbg.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://cdn-images-1.medium.com/max/1024/1*cL_22SPlCBo8LOWfNEokbg.png 424w, https://cdn-images-1.medium.com/max/1024/1*cL_22SPlCBo8LOWfNEokbg.png 848w, https://cdn-images-1.medium.com/max/1024/1*cL_22SPlCBo8LOWfNEokbg.png 1272w, https://cdn-images-1.medium.com/max/1024/1*cL_22SPlCBo8LOWfNEokbg.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">Image reproduced from<a href="https://icml.cc/virtual/2020/poster/5785"> the&nbsp;paper</a>.</figcaption></figure></div><h4><a href="https://nips.cc/virtual/2020/public/poster_3f5ee243547dee91fbd053c1c4a845aa.html">MiniLM: Deep Self-Attention Distillation for Task-Agnostic Compression of Pre-Trained Transformers</a> (NeurIPS&nbsp;2020)</h4><p><strong>Outstanding problem:</strong> Existing methods dealt with distilling the task-specific models where the pre-trained transformer-based language models are fine-tuned, but no work exists which deals with task-agnostic distillations of pre-trained models which are then fine-tuned with fewer resources compared to the large pre-trained models.</p><p>The existing distillation methods for transformers either constrain the number of layers, or the hidden representation size of the student model, or adopts layer-to-layer distillation which needs extra parameters.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://cdn-images-1.medium.com/max/795/1*_lA4l7HakUv_YRnrtWtfYw.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://cdn-images-1.medium.com/max/795/1*_lA4l7HakUv_YRnrtWtfYw.png 424w, https://cdn-images-1.medium.com/max/795/1*_lA4l7HakUv_YRnrtWtfYw.png 848w, https://cdn-images-1.medium.com/max/795/1*_lA4l7HakUv_YRnrtWtfYw.png 1272w, https://cdn-images-1.medium.com/max/795/1*_lA4l7HakUv_YRnrtWtfYw.png 1456w" sizes="100vw"><img src="https://cdn-images-1.medium.com/max/795/1*_lA4l7HakUv_YRnrtWtfYw.png" data-attrs="{&quot;src&quot;:&quot;https://cdn-images-1.medium.com/max/795/1*_lA4l7HakUv_YRnrtWtfYw.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://cdn-images-1.medium.com/max/795/1*_lA4l7HakUv_YRnrtWtfYw.png 424w, https://cdn-images-1.medium.com/max/795/1*_lA4l7HakUv_YRnrtWtfYw.png 848w, https://cdn-images-1.medium.com/max/795/1*_lA4l7HakUv_YRnrtWtfYw.png 1272w, https://cdn-images-1.medium.com/max/795/1*_lA4l7HakUv_YRnrtWtfYw.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">Image reproduced from <a href="https://nips.cc/virtual/2020/public/poster_3f5ee243547dee91fbd053c1c4a845aa.html">the&nbsp;paper</a>.</figcaption></figure></div><p><strong>Proposed solution:</strong> This proposes a distillation method that mimics the behavior of the last layer of a transformer which removes the constraint on the number of layers of a student&nbsp;network.</p><p>The novelty of this work is that it also distills the so-called value-relational knowledge which mimics the value-value scaled dot product of a teacher model along with attention weights distribution of the self-attention layer as previous methods do. See the figure&nbsp;below.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://cdn-images-1.medium.com/max/775/1*t8dea2kfbOdQflbeiHbcFw.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://cdn-images-1.medium.com/max/775/1*t8dea2kfbOdQflbeiHbcFw.png 424w, https://cdn-images-1.medium.com/max/775/1*t8dea2kfbOdQflbeiHbcFw.png 848w, https://cdn-images-1.medium.com/max/775/1*t8dea2kfbOdQflbeiHbcFw.png 1272w, https://cdn-images-1.medium.com/max/775/1*t8dea2kfbOdQflbeiHbcFw.png 1456w" sizes="100vw"><img src="https://cdn-images-1.medium.com/max/775/1*t8dea2kfbOdQflbeiHbcFw.png" data-attrs="{&quot;src&quot;:&quot;https://cdn-images-1.medium.com/max/775/1*t8dea2kfbOdQflbeiHbcFw.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://cdn-images-1.medium.com/max/775/1*t8dea2kfbOdQflbeiHbcFw.png 424w, https://cdn-images-1.medium.com/max/775/1*t8dea2kfbOdQflbeiHbcFw.png 848w, https://cdn-images-1.medium.com/max/775/1*t8dea2kfbOdQflbeiHbcFw.png 1272w, https://cdn-images-1.medium.com/max/775/1*t8dea2kfbOdQflbeiHbcFw.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">Image reproduced from <a href="https://nips.cc/virtual/2020/public/poster_3f5ee243547dee91fbd053c1c4a845aa.html">the&nbsp;paper</a>.</figcaption></figure></div><p>As the value-relational knowledge is a pair-wise relational knowledge, it removes the constraint on the hidden representation size of a student&nbsp;model.</p><p><strong>Results and Conclusions: </strong>The student model trained with this model, when fine-tuned, outperforms the task-specific distilled methods for a range of downstream tasks where all of the models are six-layer transformer models with a hidden representation size of&nbsp;768.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://cdn-images-1.medium.com/max/800/1*IbKU8wA3usaW6W6k7d17CA.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://cdn-images-1.medium.com/max/800/1*IbKU8wA3usaW6W6k7d17CA.png 424w, https://cdn-images-1.medium.com/max/800/1*IbKU8wA3usaW6W6k7d17CA.png 848w, https://cdn-images-1.medium.com/max/800/1*IbKU8wA3usaW6W6k7d17CA.png 1272w, https://cdn-images-1.medium.com/max/800/1*IbKU8wA3usaW6W6k7d17CA.png 1456w" sizes="100vw"><img src="https://cdn-images-1.medium.com/max/800/1*IbKU8wA3usaW6W6k7d17CA.png" data-attrs="{&quot;src&quot;:&quot;https://cdn-images-1.medium.com/max/800/1*IbKU8wA3usaW6W6k7d17CA.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://cdn-images-1.medium.com/max/800/1*IbKU8wA3usaW6W6k7d17CA.png 424w, https://cdn-images-1.medium.com/max/800/1*IbKU8wA3usaW6W6k7d17CA.png 848w, https://cdn-images-1.medium.com/max/800/1*IbKU8wA3usaW6W6k7d17CA.png 1272w, https://cdn-images-1.medium.com/max/800/1*IbKU8wA3usaW6W6k7d17CA.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">Image reproduced from <a href="https://nips.cc/virtual/2020/public/poster_3f5ee243547dee91fbd053c1c4a845aa.html">the&nbsp;paper</a>.</figcaption></figure></div><p>The proposed method also outperforms different methods for 3-, 4-, and 6-layer transformers for small hidden representation sizes as&nbsp;well.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://cdn-images-1.medium.com/max/975/1*FaMQTFJdbxn84xR354GaAw.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://cdn-images-1.medium.com/max/975/1*FaMQTFJdbxn84xR354GaAw.png 424w, https://cdn-images-1.medium.com/max/975/1*FaMQTFJdbxn84xR354GaAw.png 848w, https://cdn-images-1.medium.com/max/975/1*FaMQTFJdbxn84xR354GaAw.png 1272w, https://cdn-images-1.medium.com/max/975/1*FaMQTFJdbxn84xR354GaAw.png 1456w" sizes="100vw"><img src="https://cdn-images-1.medium.com/max/975/1*FaMQTFJdbxn84xR354GaAw.png" data-attrs="{&quot;src&quot;:&quot;https://cdn-images-1.medium.com/max/975/1*FaMQTFJdbxn84xR354GaAw.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://cdn-images-1.medium.com/max/975/1*FaMQTFJdbxn84xR354GaAw.png 424w, https://cdn-images-1.medium.com/max/975/1*FaMQTFJdbxn84xR354GaAw.png 848w, https://cdn-images-1.medium.com/max/975/1*FaMQTFJdbxn84xR354GaAw.png 1272w, https://cdn-images-1.medium.com/max/975/1*FaMQTFJdbxn84xR354GaAw.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">Image reproduced from <a href="https://nips.cc/virtual/2020/public/poster_3f5ee243547dee91fbd053c1c4a845aa.html">the&nbsp;paper</a>.</figcaption></figure></div><h4><a href="https://nips.cc/virtual/2020/public/poster_657b96f0592803e25a4f07166fff289a.html">Residual Distillation: Towards Portable Deep Neural Networks without Shortcuts</a> (NeurIPS&nbsp;2020)</h4><p><strong>Outstanding problem:</strong> Residual connections (a layer takes the output of the last layer as well as the input of the last layer, i.e. F(x)+x) avoid vanishing gradient problem and allow us to train very deep neural networks (ResNet-152!). But these residual connections occupy 40% of the total memory usage because of the need to store more activations and gradients.</p><p>Plain-CNN would have low latency and a low memory footprint due to the removal of residual connections with a 20&#8211;30% reduction. See figure&nbsp;below.</p><p>A distillation method that learns a network without these residual connections, but with the same accuracy, is very desirable.</p><p><strong>Proposed solution:</strong> This work proposes to distill the knowledge from a teacher ResNet with residual connections into a plain-CNN student with all the residual connections removed.</p><p>To avoid a vanishing gradient problem, this method trains the plain-CNN jointly with the corresponding ResNet. Specifically, during the forward pass, the output of the initial layers are passed both into the deeper layers of plain-CNN along with the deeper layers of the&nbsp;ResNet.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://cdn-images-1.medium.com/max/1024/1*EDzNwuZpkO-PuFF6HLrJNw.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://cdn-images-1.medium.com/max/1024/1*EDzNwuZpkO-PuFF6HLrJNw.png 424w, https://cdn-images-1.medium.com/max/1024/1*EDzNwuZpkO-PuFF6HLrJNw.png 848w, https://cdn-images-1.medium.com/max/1024/1*EDzNwuZpkO-PuFF6HLrJNw.png 1272w, https://cdn-images-1.medium.com/max/1024/1*EDzNwuZpkO-PuFF6HLrJNw.png 1456w" sizes="100vw"><img src="https://cdn-images-1.medium.com/max/1024/1*EDzNwuZpkO-PuFF6HLrJNw.png" data-attrs="{&quot;src&quot;:&quot;https://cdn-images-1.medium.com/max/1024/1*EDzNwuZpkO-PuFF6HLrJNw.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://cdn-images-1.medium.com/max/1024/1*EDzNwuZpkO-PuFF6HLrJNw.png 424w, https://cdn-images-1.medium.com/max/1024/1*EDzNwuZpkO-PuFF6HLrJNw.png 848w, https://cdn-images-1.medium.com/max/1024/1*EDzNwuZpkO-PuFF6HLrJNw.png 1272w, https://cdn-images-1.medium.com/max/1024/1*EDzNwuZpkO-PuFF6HLrJNw.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">Image reproduced from<a href="https://nips.cc/virtual/2020/public/poster_657b96f0592803e25a4f07166fff289a.html"> the&nbsp;paper</a>.</figcaption></figure></div><p>Thus, the gradients flow through the initial layer of the plain-CNN because of the residual connections of the ResNet. But the inference after training is only done with the plain-CNN, thus avoiding memory and computations bandwidth of having residual networks.</p><p><strong>Results and Conclusions:</strong> The plain-CNN trained with this method shows a performance competitive with the corresponding ResNet and outperforms others knowledge distillation methods.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://cdn-images-1.medium.com/max/1024/1*-wVL3SKVIULHhc65aKci8w.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://cdn-images-1.medium.com/max/1024/1*-wVL3SKVIULHhc65aKci8w.png 424w, https://cdn-images-1.medium.com/max/1024/1*-wVL3SKVIULHhc65aKci8w.png 848w, https://cdn-images-1.medium.com/max/1024/1*-wVL3SKVIULHhc65aKci8w.png 1272w, https://cdn-images-1.medium.com/max/1024/1*-wVL3SKVIULHhc65aKci8w.png 1456w" sizes="100vw"><img src="https://cdn-images-1.medium.com/max/1024/1*-wVL3SKVIULHhc65aKci8w.png" data-attrs="{&quot;src&quot;:&quot;https://cdn-images-1.medium.com/max/1024/1*-wVL3SKVIULHhc65aKci8w.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://cdn-images-1.medium.com/max/1024/1*-wVL3SKVIULHhc65aKci8w.png 424w, https://cdn-images-1.medium.com/max/1024/1*-wVL3SKVIULHhc65aKci8w.png 848w, https://cdn-images-1.medium.com/max/1024/1*-wVL3SKVIULHhc65aKci8w.png 1272w, https://cdn-images-1.medium.com/max/1024/1*-wVL3SKVIULHhc65aKci8w.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">Image reproduced from<a href="https://nips.cc/virtual/2020/public/poster_657b96f0592803e25a4f07166fff289a.html"> the&nbsp;paper</a>.</figcaption></figure></div><p>Even with pruning on top of knowledge distillation, plain-CNN performs competitively and some times better than the corresponding ResNet with pruning ratios of 30 and 60&nbsp;percent.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://cdn-images-1.medium.com/max/729/1*n6Nay83pLvlwoZ9For5RUg.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://cdn-images-1.medium.com/max/729/1*n6Nay83pLvlwoZ9For5RUg.png 424w, https://cdn-images-1.medium.com/max/729/1*n6Nay83pLvlwoZ9For5RUg.png 848w, https://cdn-images-1.medium.com/max/729/1*n6Nay83pLvlwoZ9For5RUg.png 1272w, https://cdn-images-1.medium.com/max/729/1*n6Nay83pLvlwoZ9For5RUg.png 1456w" sizes="100vw"><img src="https://cdn-images-1.medium.com/max/729/1*n6Nay83pLvlwoZ9For5RUg.png" data-attrs="{&quot;src&quot;:&quot;https://cdn-images-1.medium.com/max/729/1*n6Nay83pLvlwoZ9For5RUg.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://cdn-images-1.medium.com/max/729/1*n6Nay83pLvlwoZ9For5RUg.png 424w, https://cdn-images-1.medium.com/max/729/1*n6Nay83pLvlwoZ9For5RUg.png 848w, https://cdn-images-1.medium.com/max/729/1*n6Nay83pLvlwoZ9For5RUg.png 1272w, https://cdn-images-1.medium.com/max/729/1*n6Nay83pLvlwoZ9For5RUg.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">Image reproduced from <a href="https://nips.cc/virtual/2020/public/poster_657b96f0592803e25a4f07166fff289a.html">the&nbsp;paper</a>.</figcaption></figure></div><h4><a href="https://nips.cc/virtual/2020/public/poster_912d2b1c7b2826caf99687388d2e8f7c.html">Kernel-Based Progressive Distillation for Adder Neural Networks</a> (NeurIPS&nbsp;2020)</h4><p><strong>Outstanding problem:</strong> Adder Neural Networks (ANNs) replace the convolutional operation in CNNs with additions (L1 operator) thus reducing the computational resources amenable to deploying in low-resource environments such as mobile phones and&nbsp;cameras.</p><p>In practice it is observed that ANNs are difficult to optimize and thus the performance of ANNs is still lower than CNNs. This calls for methods that can train ANNs efficiently and increase their performance to be competitive with traditional CNNs.</p><p><strong>Proposed solution:</strong> This work proposes a method to distill the knowledge from CNN into an ANN, both the knowledge of output logit distributions and intermediate representations.</p><p>First, the authors observe that the optimal weight initialization of ANNs is based on Laplace distribution, but Gaussian distribution is used for CNNs and that the representation distributions in ANNs and CNNs are of different characteristics.</p><p>Thus, they devise a kernel method to transform the intermediate features from ANNs and CNNs into one vector space to enable distillation. They call this method a progressive kernel-based knowledge distillation (PKKD).</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://cdn-images-1.medium.com/max/934/1*hJT_wRUdsPrL0WKpoVcSBg.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://cdn-images-1.medium.com/max/934/1*hJT_wRUdsPrL0WKpoVcSBg.png 424w, https://cdn-images-1.medium.com/max/934/1*hJT_wRUdsPrL0WKpoVcSBg.png 848w, https://cdn-images-1.medium.com/max/934/1*hJT_wRUdsPrL0WKpoVcSBg.png 1272w, https://cdn-images-1.medium.com/max/934/1*hJT_wRUdsPrL0WKpoVcSBg.png 1456w" sizes="100vw"><img src="https://cdn-images-1.medium.com/max/934/1*hJT_wRUdsPrL0WKpoVcSBg.png" data-attrs="{&quot;src&quot;:&quot;https://cdn-images-1.medium.com/max/934/1*hJT_wRUdsPrL0WKpoVcSBg.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://cdn-images-1.medium.com/max/934/1*hJT_wRUdsPrL0WKpoVcSBg.png 424w, https://cdn-images-1.medium.com/max/934/1*hJT_wRUdsPrL0WKpoVcSBg.png 848w, https://cdn-images-1.medium.com/max/934/1*hJT_wRUdsPrL0WKpoVcSBg.png 1272w, https://cdn-images-1.medium.com/max/934/1*hJT_wRUdsPrL0WKpoVcSBg.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">Image reproduced from <a href="https://nips.cc/virtual/2020/public/poster_912d2b1c7b2826caf99687388d2e8f7c.html">the&nbsp;paper.</a></figcaption></figure></div><p><strong>Results and Conclusions:</strong> PKKD ANN outperforms Binary Neural Networks (BNNs) and other variants of ANNs on CIFAR-10, CIFAR-100.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://cdn-images-1.medium.com/max/846/1*G8R3cdGh6liz6Kh-uBWFcQ.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://cdn-images-1.medium.com/max/846/1*G8R3cdGh6liz6Kh-uBWFcQ.png 424w, https://cdn-images-1.medium.com/max/846/1*G8R3cdGh6liz6Kh-uBWFcQ.png 848w, https://cdn-images-1.medium.com/max/846/1*G8R3cdGh6liz6Kh-uBWFcQ.png 1272w, https://cdn-images-1.medium.com/max/846/1*G8R3cdGh6liz6Kh-uBWFcQ.png 1456w" sizes="100vw"><img src="https://cdn-images-1.medium.com/max/846/1*G8R3cdGh6liz6Kh-uBWFcQ.png" data-attrs="{&quot;src&quot;:&quot;https://cdn-images-1.medium.com/max/846/1*G8R3cdGh6liz6Kh-uBWFcQ.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://cdn-images-1.medium.com/max/846/1*G8R3cdGh6liz6Kh-uBWFcQ.png 424w, https://cdn-images-1.medium.com/max/846/1*G8R3cdGh6liz6Kh-uBWFcQ.png 848w, https://cdn-images-1.medium.com/max/846/1*G8R3cdGh6liz6Kh-uBWFcQ.png 1272w, https://cdn-images-1.medium.com/max/846/1*G8R3cdGh6liz6Kh-uBWFcQ.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">Image reproduced from <a href="https://nips.cc/virtual/2020/public/poster_912d2b1c7b2826caf99687388d2e8f7c.html">the&nbsp;paper</a>.</figcaption></figure></div><p>And also on the ImageNet&nbsp;dataset.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://cdn-images-1.medium.com/max/786/1*IJ4dMuuHQ7UixgtPnwNkng.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://cdn-images-1.medium.com/max/786/1*IJ4dMuuHQ7UixgtPnwNkng.png 424w, https://cdn-images-1.medium.com/max/786/1*IJ4dMuuHQ7UixgtPnwNkng.png 848w, https://cdn-images-1.medium.com/max/786/1*IJ4dMuuHQ7UixgtPnwNkng.png 1272w, https://cdn-images-1.medium.com/max/786/1*IJ4dMuuHQ7UixgtPnwNkng.png 1456w" sizes="100vw"><img src="https://cdn-images-1.medium.com/max/786/1*IJ4dMuuHQ7UixgtPnwNkng.png" data-attrs="{&quot;src&quot;:&quot;https://cdn-images-1.medium.com/max/786/1*IJ4dMuuHQ7UixgtPnwNkng.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://cdn-images-1.medium.com/max/786/1*IJ4dMuuHQ7UixgtPnwNkng.png 424w, https://cdn-images-1.medium.com/max/786/1*IJ4dMuuHQ7UixgtPnwNkng.png 848w, https://cdn-images-1.medium.com/max/786/1*IJ4dMuuHQ7UixgtPnwNkng.png 1272w, https://cdn-images-1.medium.com/max/786/1*IJ4dMuuHQ7UixgtPnwNkng.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">Image reproduced from <a href="https://nips.cc/virtual/2020/public/poster_912d2b1c7b2826caf99687388d2e8f7c.html">the&nbsp;paper</a>.</figcaption></figure></div><h4>Conclusion</h4><p>Research on Neural Network Distillation, and, more generally, NN compression, is evolving to be more scientific and rigorous. One of the reasons is, undoubtedly, the interaction between the wide adoption of deep learning methods in computer vision &amp; NLP and elsewhere and the increasing amount of memory, energy, and computational resources required for state-of-the-art methods.</p><p>Due to the research in 2020, we learned that imitating the feature representations are more efficient than imitating logits alone. We can now make a ResNet into a feed-forward network without much degradation in performance and we know a more efficient way to self-distill large language&nbsp;models.</p><p>Going into 2021, it would be great to see approaches that combine distillation with pruning and quantization to be more efficient than simply performing two or three of them in sequence.</p><p><em>Editor&#8217;s Note: <a href="https://heartbeat.comet.ml/">Heartbeat</a> is a contributor-driven online publication and community dedicated to providing premier educational resources for data science, machine learning, and deep learning practitioners. We&#8217;re committed to supporting and inspiring developers and engineers from all walks of&nbsp;life.</em></p><p><em>Editorially independent, Heartbeat is sponsored and published by <a href="http://comet.ml/?utm_campaign=heartbeat-statement&amp;utm_source=blog&amp;utm_medium=medium">Comet</a>, an MLOps platform that enables data scientists &amp; ML teams to track, compare, explain, &amp; optimize their experiments. We pay our contributors, and we don&#8217;t sell&nbsp;ads.</em></p><p><em>If you&#8217;d like to contribute, head on over to our<a href="https://heartbeat.fritz.ai/call-for-contributors-october-2018-update-fee7f5b80f3e"> call for contributors</a>. You can also sign up to receive our weekly newsletters (<a href="https://www.deeplearningweekly.com/">Deep Learning Weekly</a> and the <a href="https://info.comet.ml/newsletter-signup/">Comet Newsletter</a>), join us on<a href="https://join.slack.com/t/fritz-ai-community/shared_invite/enQtNTY5NDM2MTQwMTgwLWU4ZDEwNTAxYWE2YjIxZDllMTcxMWE4MGFhNDk5Y2QwNTcxYzEyNWZmZWEwMzE4NTFkOWY2NTM0OGQwYjM5Y2U"> </a><a href="https://join.slack.com/t/cometml/shared_invite/zt-49v4zxxz-qHcTeyrMEzqZc5lQb9hgvw">Slack</a>, and follow Comet on <a href="https://twitter.com/Cometml">Twitter</a> and <a href="https://www.linkedin.com/company/comet-ml/">LinkedIn</a> for resources, events, and much more that will help you build better ML models,&nbsp;faster.</em></p><div><hr></div><p><a href="https://heartbeat.comet.ml/knowledge-distillation-research-review-2020-e0333f554863">Knowledge Distillation Research Review</a> was originally published in <a href="https://heartbeat.comet.ml">Heartbeat</a> on Medium, where people are continuing the conversation by highlighting and responding to this story.</p>]]></content:encoded></item><item><title><![CDATA[Neural Network Quantization Research Review]]></title><description><![CDATA[The latest advances and insights for Neural Network model compression with Quantization]]></description><link>https://www.aipapertrails.com/p/neural-network-quantization-research-review-2020-6d72b06f09b1</link><guid isPermaLink="false">https://www.aipapertrails.com/p/neural-network-quantization-research-review-2020-6d72b06f09b1</guid><dc:creator><![CDATA[Prakash Kagitha]]></dc:creator><pubDate>Tue, 19 Jan 2021 13:49:45 GMT</pubDate><enclosure url="https://cdn-images-1.medium.com/max/763/1*WiOvCAqNnUj0pgNLr1OXeQ.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<h4>The latest advances and insights for Neural Network model compression with Quantization</h4><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://cdn-images-1.medium.com/max/763/1*WiOvCAqNnUj0pgNLr1OXeQ.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://cdn-images-1.medium.com/max/763/1*WiOvCAqNnUj0pgNLr1OXeQ.png 424w, https://cdn-images-1.medium.com/max/763/1*WiOvCAqNnUj0pgNLr1OXeQ.png 848w, https://cdn-images-1.medium.com/max/763/1*WiOvCAqNnUj0pgNLr1OXeQ.png 1272w, https://cdn-images-1.medium.com/max/763/1*WiOvCAqNnUj0pgNLr1OXeQ.png 1456w" sizes="100vw"><img src="https://cdn-images-1.medium.com/max/763/1*WiOvCAqNnUj0pgNLr1OXeQ.png" data-attrs="{&quot;src&quot;:&quot;https://cdn-images-1.medium.com/max/763/1*WiOvCAqNnUj0pgNLr1OXeQ.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://cdn-images-1.medium.com/max/763/1*WiOvCAqNnUj0pgNLr1OXeQ.png 424w, https://cdn-images-1.medium.com/max/763/1*WiOvCAqNnUj0pgNLr1OXeQ.png 848w, https://cdn-images-1.medium.com/max/763/1*WiOvCAqNnUj0pgNLr1OXeQ.png 1272w, https://cdn-images-1.medium.com/max/763/1*WiOvCAqNnUj0pgNLr1OXeQ.png 1456w" sizes="100vw" fetchpriority="high"></picture><div></div></div></a></figure></div><p>Neural network quantization is a process of reducing the precision of the weights in the neural network, thus reducing the memory, computation, and energy bandwidths.</p><p>Particularly when deploying NN models on mobile or edge devices, quantization, and model compression in general, is desirable and often the only plausible way to deploy a mobile model as the memory and computational budget of these devices is very&nbsp;limited.</p><p>Why not use potential infinite virtual memory and computational power from cloud machines? While a lot of NN models are running on the cloud even now, latency is not low enough for mobile/edge devices, which hinders utility and requires data to be transferred to the cloud which rises a lot of privacy concerns.</p><h4>Preliminaries of Neural Network Quantization</h4><p>There are a lot of quantization methods in the literature today and to best understand the pros and cons of each method it helps to classify quantization methods. Quantization in a general sense is assigning a discrete value from a pre-specified set of values for a given&nbsp;input.</p><p><strong>Quantization-aware training vs. post-training quantization:</strong> When the quantization operations are embedded into a neural network and then trained it is called quantization-aware training. And if quantization is performed on a model after training, for example with rounding mechanisms, it is called post-training quantization.</p><p><strong>Scalar quantization vs. vector quantization vs. product quantization:</strong> If the input is scalar and the pre-specified set is also a scalar, then it is called scalar quantization (for example, making FP32 floating point into INT8 integer). If the input is a vector and it is assigned to a vector, it is called vector quantization. Product quantization is a form of vector quantization with increased granularity as it works with sub-vectors rather than the entire&nbsp;vector.</p><p><strong>Unified quantization or non-unified quantization: </strong>During quantization, if there is a restriction that discrete values that can be assigned should be at equal step size, then it is called unified quantization. Converting FP32 to INT8 is an example of unified quantization. Alternately, if the discrete set of values to be assigned doesn&#8217;t have any restrictions it is called non-unified quantization. Vector quantization and product quantization are examples of this&nbsp;type.</p><p><strong>Fixed-precision quantization vs. mixed-precision quantization:</strong> If different layers or channels have different precisions or bit-widths then it is called mixed-precision quantization. If the entire network is quantized to values of one bit-width, then it is called fixed-precision quantization.</p><p><strong>Layer-wise quantization or channel-wise quantization:</strong> This is a fine-grained classification within mixed-precision quantization depending on whether we have different bit-widths for each layer or for each&nbsp;channel.</p><p>Now we&#8217;ll dive into six research papers that addressed these problems in&nbsp;2020.</p><h4><a href="https://iclr.cc/virtual_2020/poster_rJehVyrKwH.html">And the Bit Goes Down: Revisiting the Quantization of Neural Networks</a> (ICLR&nbsp;2020)</h4><p><strong>Outstanding problem:</strong> State-of-the-art deep learning models take a lot of memory. Even ResNet-50, which is not quite state-of-the-art takes ~100MB and the Faster-RCNN takes&nbsp;~200MB.</p><p>Compressing the model alone is desirable to deploy deep learning solutions in low-memory environments.</p><p><strong>Proposed solution:</strong> This work proposes applying Product Quantization (PQ) to the weights of the network which represent the whole network with very few floating-point numbers.</p><p><strong>Product Quantization</strong><em>:</em> First, the columns of the M x N matrix are each split into d sub-vectors. It results in a total of M*d sub-vectors. These vectors are grouped into clusters, sometimes while minimizing a pre-specified loss. Then, to compress the matrix, these sub-vectors are replaced with the centroid of the cluster in which it falls. This achieves a stark compression as we store only cluster centroid vectors and cluster labels for sub-vectors rather than storing all the sub-vectors in a&nbsp;matrix.</p><p>This work applies the same process for weight matrix at each convolutional layer which is of size (<strong>C_out</strong> x<strong> C_in</strong> x <strong>K</strong> x <strong>K</strong>) where C_out and C_in are the number of channels going out of a layer and the number of channels coming into a layer and K is the height and width of a square&nbsp;kernel.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://cdn-images-1.medium.com/max/1024/1*D6HYeeo1EHKVGoIB3rLjMA.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://cdn-images-1.medium.com/max/1024/1*D6HYeeo1EHKVGoIB3rLjMA.png 424w, https://cdn-images-1.medium.com/max/1024/1*D6HYeeo1EHKVGoIB3rLjMA.png 848w, https://cdn-images-1.medium.com/max/1024/1*D6HYeeo1EHKVGoIB3rLjMA.png 1272w, https://cdn-images-1.medium.com/max/1024/1*D6HYeeo1EHKVGoIB3rLjMA.png 1456w" sizes="100vw"><img src="https://cdn-images-1.medium.com/max/1024/1*D6HYeeo1EHKVGoIB3rLjMA.png" data-attrs="{&quot;src&quot;:&quot;https://cdn-images-1.medium.com/max/1024/1*D6HYeeo1EHKVGoIB3rLjMA.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://cdn-images-1.medium.com/max/1024/1*D6HYeeo1EHKVGoIB3rLjMA.png 424w, https://cdn-images-1.medium.com/max/1024/1*D6HYeeo1EHKVGoIB3rLjMA.png 848w, https://cdn-images-1.medium.com/max/1024/1*D6HYeeo1EHKVGoIB3rLjMA.png 1272w, https://cdn-images-1.medium.com/max/1024/1*D6HYeeo1EHKVGoIB3rLjMA.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">Figure reproduced from <a href="https://iclr.cc/virtual_2020/poster_rJehVyrKwH.html">the&nbsp;paper</a>.</figcaption></figure></div><p>First, the weight matrix is reshaped to a 2d matrix where the size becomes (<strong>C_out</strong> x <strong>C_in*K*K</strong>). Columns of the weight matrix of length C_in*K*K are transformed into C_in sub-vectors each of K*K length. This results in C_out*C_in sub-vectors.</p><p>The labels of the sub-vectors are learned with weighted k-means with an objective of reducing the offset (reconstruction loss) between activations produced by the original weight matrix and the activations produced by a matrix created by replacing all the sub-vectors with their cluster centroids.</p><p>(They also show that reconstructing the activations of a layer is more efficient than reconstructing the actual weights of the network.)</p><p><strong>Results and conclusion:</strong> It is shown that models could be compressed efficiently with this approach compared to the existing&nbsp;methods.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://cdn-images-1.medium.com/max/1024/1*pt5HrSd-bp1CPx2V61dMNg.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://cdn-images-1.medium.com/max/1024/1*pt5HrSd-bp1CPx2V61dMNg.png 424w, https://cdn-images-1.medium.com/max/1024/1*pt5HrSd-bp1CPx2V61dMNg.png 848w, https://cdn-images-1.medium.com/max/1024/1*pt5HrSd-bp1CPx2V61dMNg.png 1272w, https://cdn-images-1.medium.com/max/1024/1*pt5HrSd-bp1CPx2V61dMNg.png 1456w" sizes="100vw"><img src="https://cdn-images-1.medium.com/max/1024/1*pt5HrSd-bp1CPx2V61dMNg.png" data-attrs="{&quot;src&quot;:&quot;https://cdn-images-1.medium.com/max/1024/1*pt5HrSd-bp1CPx2V61dMNg.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://cdn-images-1.medium.com/max/1024/1*pt5HrSd-bp1CPx2V61dMNg.png 424w, https://cdn-images-1.medium.com/max/1024/1*pt5HrSd-bp1CPx2V61dMNg.png 848w, https://cdn-images-1.medium.com/max/1024/1*pt5HrSd-bp1CPx2V61dMNg.png 1272w, https://cdn-images-1.medium.com/max/1024/1*pt5HrSd-bp1CPx2V61dMNg.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">Figure reproduced from <a href="https://iclr.cc/virtual_2020/poster_rJehVyrKwH.html">the&nbsp;paper</a>.</figcaption></figure></div><p>Also, a ResNet-50 model is quantized to a size of 5 MB (with 20x compression factor) and Mask R-CNN to a size of ~6MB (26x compression factor) both with competitive performances compared to the original&nbsp;models.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://cdn-images-1.medium.com/max/1024/1*nKXlA7jnAZ2N1vWtcwe3Kg.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://cdn-images-1.medium.com/max/1024/1*nKXlA7jnAZ2N1vWtcwe3Kg.png 424w, https://cdn-images-1.medium.com/max/1024/1*nKXlA7jnAZ2N1vWtcwe3Kg.png 848w, https://cdn-images-1.medium.com/max/1024/1*nKXlA7jnAZ2N1vWtcwe3Kg.png 1272w, https://cdn-images-1.medium.com/max/1024/1*nKXlA7jnAZ2N1vWtcwe3Kg.png 1456w" sizes="100vw"><img src="https://cdn-images-1.medium.com/max/1024/1*nKXlA7jnAZ2N1vWtcwe3Kg.png" data-attrs="{&quot;src&quot;:&quot;https://cdn-images-1.medium.com/max/1024/1*nKXlA7jnAZ2N1vWtcwe3Kg.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://cdn-images-1.medium.com/max/1024/1*nKXlA7jnAZ2N1vWtcwe3Kg.png 424w, https://cdn-images-1.medium.com/max/1024/1*nKXlA7jnAZ2N1vWtcwe3Kg.png 848w, https://cdn-images-1.medium.com/max/1024/1*nKXlA7jnAZ2N1vWtcwe3Kg.png 1272w, https://cdn-images-1.medium.com/max/1024/1*nKXlA7jnAZ2N1vWtcwe3Kg.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">Figure reproduced from <a href="https://iclr.cc/virtual_2020/poster_rJehVyrKwH.html">the&nbsp;paper</a>.</figcaption></figure></div><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://cdn-images-1.medium.com/max/1024/1*9VpUbfS0q49tDKMTIhGQ-Q.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://cdn-images-1.medium.com/max/1024/1*9VpUbfS0q49tDKMTIhGQ-Q.png 424w, https://cdn-images-1.medium.com/max/1024/1*9VpUbfS0q49tDKMTIhGQ-Q.png 848w, https://cdn-images-1.medium.com/max/1024/1*9VpUbfS0q49tDKMTIhGQ-Q.png 1272w, https://cdn-images-1.medium.com/max/1024/1*9VpUbfS0q49tDKMTIhGQ-Q.png 1456w" sizes="100vw"><img src="https://cdn-images-1.medium.com/max/1024/1*9VpUbfS0q49tDKMTIhGQ-Q.png" data-attrs="{&quot;src&quot;:&quot;https://cdn-images-1.medium.com/max/1024/1*9VpUbfS0q49tDKMTIhGQ-Q.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://cdn-images-1.medium.com/max/1024/1*9VpUbfS0q49tDKMTIhGQ-Q.png 424w, https://cdn-images-1.medium.com/max/1024/1*9VpUbfS0q49tDKMTIhGQ-Q.png 848w, https://cdn-images-1.medium.com/max/1024/1*9VpUbfS0q49tDKMTIhGQ-Q.png 1272w, https://cdn-images-1.medium.com/max/1024/1*9VpUbfS0q49tDKMTIhGQ-Q.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">Figure reproduced from <a href="https://iclr.cc/virtual_2020/poster_rJehVyrKwH.html">the current&nbsp;paper</a>.</figcaption></figure></div><h4><a href="https://iclr.cc/virtual_2020/poster_rygfnn4twS.html">AutoQ: Automated Kernel-Wise Neural Network Quantization</a></h4><p><strong>Outstanding problem:</strong> Kernel-wise quantization (different bit-widths for each kernel) is shown to be more effective than layer-wise quantization which is in turn shown to be more effective than network-wise quantization.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://cdn-images-1.medium.com/max/763/1*WiOvCAqNnUj0pgNLr1OXeQ.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://cdn-images-1.medium.com/max/763/1*WiOvCAqNnUj0pgNLr1OXeQ.png 424w, https://cdn-images-1.medium.com/max/763/1*WiOvCAqNnUj0pgNLr1OXeQ.png 848w, https://cdn-images-1.medium.com/max/763/1*WiOvCAqNnUj0pgNLr1OXeQ.png 1272w, https://cdn-images-1.medium.com/max/763/1*WiOvCAqNnUj0pgNLr1OXeQ.png 1456w" sizes="100vw"><img src="https://cdn-images-1.medium.com/max/763/1*WiOvCAqNnUj0pgNLr1OXeQ.png" data-attrs="{&quot;src&quot;:&quot;https://cdn-images-1.medium.com/max/763/1*WiOvCAqNnUj0pgNLr1OXeQ.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://cdn-images-1.medium.com/max/763/1*WiOvCAqNnUj0pgNLr1OXeQ.png 424w, https://cdn-images-1.medium.com/max/763/1*WiOvCAqNnUj0pgNLr1OXeQ.png 848w, https://cdn-images-1.medium.com/max/763/1*WiOvCAqNnUj0pgNLr1OXeQ.png 1272w, https://cdn-images-1.medium.com/max/763/1*WiOvCAqNnUj0pgNLr1OXeQ.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">Figure reproduced from <a href="https://iclr.cc/virtual_2020/poster_rygfnn4twS.html">the&nbsp;paper</a>.</figcaption></figure></div><p>As the search for space of all possible combinations is huge, manual engineering of kernel-wise bit-widths is sub-optimal. It is shown that even a generic RL agent like DDPG can&#8217;t find a good-enough policy. This work proposes a method that learns kernel-wise bit-widths more efficiently than any other existing&nbsp;method.</p><p><strong>Proposed solution:</strong> This work proposed a Hierarchical Reinforcement Learning method to first choose layer-wise bit-widths with a High-Level Controller and then choose kernel-wise bit-widths with a Low-Level Controller within that&nbsp;layer.</p><p>HLC chooses bit-width for activations of a particular layer and also the mean bit-widths of all the kernels (termed as &#8216;goal for LLC&#8217;) of that layer. LLC then chooses bit-widths for all the kernels in that particular layer.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://cdn-images-1.medium.com/max/1024/1*qGJfYLrPe7y-r2-gWxgluQ.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://cdn-images-1.medium.com/max/1024/1*qGJfYLrPe7y-r2-gWxgluQ.png 424w, https://cdn-images-1.medium.com/max/1024/1*qGJfYLrPe7y-r2-gWxgluQ.png 848w, https://cdn-images-1.medium.com/max/1024/1*qGJfYLrPe7y-r2-gWxgluQ.png 1272w, https://cdn-images-1.medium.com/max/1024/1*qGJfYLrPe7y-r2-gWxgluQ.png 1456w" sizes="100vw"><img src="https://cdn-images-1.medium.com/max/1024/1*qGJfYLrPe7y-r2-gWxgluQ.png" data-attrs="{&quot;src&quot;:&quot;https://cdn-images-1.medium.com/max/1024/1*qGJfYLrPe7y-r2-gWxgluQ.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://cdn-images-1.medium.com/max/1024/1*qGJfYLrPe7y-r2-gWxgluQ.png 424w, https://cdn-images-1.medium.com/max/1024/1*qGJfYLrPe7y-r2-gWxgluQ.png 848w, https://cdn-images-1.medium.com/max/1024/1*qGJfYLrPe7y-r2-gWxgluQ.png 1272w, https://cdn-images-1.medium.com/max/1024/1*qGJfYLrPe7y-r2-gWxgluQ.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">Figure reproduced from<a href="https://iclr.cc/virtual_2020/poster_rygfnn4twS.html"> the&nbsp;paper</a>.</figcaption></figure></div><p>The reward for this HRL agent in training is devised such that the number of FLOPs and memory required is lower and test accuracy is&nbsp;higher.</p><p>To accelerate training, LLC is given an additional intrinsic reward, when it completes the goal set by HLC or chooses the bid-widths for kernels such that the mean is equal to the predicted mean of the HLC for that particular layer.</p><p><strong>Results and conclusion:</strong> The proposed method outperforms previous RL methods like DDPG. It outperforms HIRO as well, even though their implementation is based on HIRO because of the intrinsic reward they designed for the&nbsp;LLC.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://cdn-images-1.medium.com/max/610/1*38l3UJxt6Kh1BMo_LBmxbA.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://cdn-images-1.medium.com/max/610/1*38l3UJxt6Kh1BMo_LBmxbA.png 424w, https://cdn-images-1.medium.com/max/610/1*38l3UJxt6Kh1BMo_LBmxbA.png 848w, https://cdn-images-1.medium.com/max/610/1*38l3UJxt6Kh1BMo_LBmxbA.png 1272w, https://cdn-images-1.medium.com/max/610/1*38l3UJxt6Kh1BMo_LBmxbA.png 1456w" sizes="100vw"><img src="https://cdn-images-1.medium.com/max/610/1*38l3UJxt6Kh1BMo_LBmxbA.png" data-attrs="{&quot;src&quot;:&quot;https://cdn-images-1.medium.com/max/610/1*38l3UJxt6Kh1BMo_LBmxbA.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://cdn-images-1.medium.com/max/610/1*38l3UJxt6Kh1BMo_LBmxbA.png 424w, https://cdn-images-1.medium.com/max/610/1*38l3UJxt6Kh1BMo_LBmxbA.png 848w, https://cdn-images-1.medium.com/max/610/1*38l3UJxt6Kh1BMo_LBmxbA.png 1272w, https://cdn-images-1.medium.com/max/610/1*38l3UJxt6Kh1BMo_LBmxbA.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">Figure reproduced from <a href="https://iclr.cc/virtual_2020/poster_rygfnn4twS.html">the&nbsp;paper</a>.</figcaption></figure></div><p>They also show that kernel-wise quantization is the right granularity to choose bit-widths with an experiment showing that sub-kernel-wise quantization doesn&#8217;t improve latency (inference time) beyond kernel-wise quantization.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://cdn-images-1.medium.com/max/1024/1*XB3M75UnyrYWmsc7oA1vVQ.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://cdn-images-1.medium.com/max/1024/1*XB3M75UnyrYWmsc7oA1vVQ.png 424w, https://cdn-images-1.medium.com/max/1024/1*XB3M75UnyrYWmsc7oA1vVQ.png 848w, https://cdn-images-1.medium.com/max/1024/1*XB3M75UnyrYWmsc7oA1vVQ.png 1272w, https://cdn-images-1.medium.com/max/1024/1*XB3M75UnyrYWmsc7oA1vVQ.png 1456w" sizes="100vw"><img src="https://cdn-images-1.medium.com/max/1024/1*XB3M75UnyrYWmsc7oA1vVQ.png" data-attrs="{&quot;src&quot;:&quot;https://cdn-images-1.medium.com/max/1024/1*XB3M75UnyrYWmsc7oA1vVQ.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://cdn-images-1.medium.com/max/1024/1*XB3M75UnyrYWmsc7oA1vVQ.png 424w, https://cdn-images-1.medium.com/max/1024/1*XB3M75UnyrYWmsc7oA1vVQ.png 848w, https://cdn-images-1.medium.com/max/1024/1*XB3M75UnyrYWmsc7oA1vVQ.png 1272w, https://cdn-images-1.medium.com/max/1024/1*XB3M75UnyrYWmsc7oA1vVQ.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">Figure reproduced from <a href="https://iclr.cc/virtual_2020/poster_rygfnn4twS.html">the&nbsp;paper</a>.</figcaption></figure></div><h4><a href="https://icml.cc/virtual/2020/poster/6429">Differentiable Product Quantization for End-to-End Embedding Compression</a></h4><p><strong>Outstanding problem:</strong> The embedding layer alone has 95% of all parameters in a medium-sized LSTM with the vocabulary of a language modeling dataset&nbsp;PTB.</p><p>Compressing the embedding layers without loss in performance of a model would make them mobile and edge-device friendly.</p><p><strong>Proposed solution:</strong> This method adopts the Product Quantization (PQ) method where columns of a matrix are split into sub-vectors, clustered, and finally are replaced with the centroid sub-vectors of the cluster they belong&nbsp;to.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://cdn-images-1.medium.com/max/1024/1*khjTW1tG21SsOCc-_2UjOA.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://cdn-images-1.medium.com/max/1024/1*khjTW1tG21SsOCc-_2UjOA.png 424w, https://cdn-images-1.medium.com/max/1024/1*khjTW1tG21SsOCc-_2UjOA.png 848w, https://cdn-images-1.medium.com/max/1024/1*khjTW1tG21SsOCc-_2UjOA.png 1272w, https://cdn-images-1.medium.com/max/1024/1*khjTW1tG21SsOCc-_2UjOA.png 1456w" sizes="100vw"><img src="https://cdn-images-1.medium.com/max/1024/1*khjTW1tG21SsOCc-_2UjOA.png" data-attrs="{&quot;src&quot;:&quot;https://cdn-images-1.medium.com/max/1024/1*khjTW1tG21SsOCc-_2UjOA.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://cdn-images-1.medium.com/max/1024/1*khjTW1tG21SsOCc-_2UjOA.png 424w, https://cdn-images-1.medium.com/max/1024/1*khjTW1tG21SsOCc-_2UjOA.png 848w, https://cdn-images-1.medium.com/max/1024/1*khjTW1tG21SsOCc-_2UjOA.png 1272w, https://cdn-images-1.medium.com/max/1024/1*khjTW1tG21SsOCc-_2UjOA.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">Figure reproduced from <a href="https://icml.cc/virtual/2020/poster/6429">the&nbsp;paper</a>.</figcaption></figure></div><p>The notable aspect of this method is that it makes PQ differentiable by approximating the non-differentiable operators with Stop Gradient operators.</p><p>Given a weight matrix to compress (Query Q) the proposed method learns a codebook that contains the centroid vectors for each cluster and also the cluster labels for each sub-vector. At inference, this codebook is used to revert back the discrete representation (a list of cluster labels for each column) to a continuous vector representation.</p><p><strong>Results and conclusion:</strong> The proposed method outperforms recently proposed embedding compression methods.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://cdn-images-1.medium.com/max/1024/1*5Axjs4qQ9TPFmiHTd4KfKA.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://cdn-images-1.medium.com/max/1024/1*5Axjs4qQ9TPFmiHTd4KfKA.png 424w, https://cdn-images-1.medium.com/max/1024/1*5Axjs4qQ9TPFmiHTd4KfKA.png 848w, https://cdn-images-1.medium.com/max/1024/1*5Axjs4qQ9TPFmiHTd4KfKA.png 1272w, https://cdn-images-1.medium.com/max/1024/1*5Axjs4qQ9TPFmiHTd4KfKA.png 1456w" sizes="100vw"><img src="https://cdn-images-1.medium.com/max/1024/1*5Axjs4qQ9TPFmiHTd4KfKA.png" data-attrs="{&quot;src&quot;:&quot;https://cdn-images-1.medium.com/max/1024/1*5Axjs4qQ9TPFmiHTd4KfKA.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://cdn-images-1.medium.com/max/1024/1*5Axjs4qQ9TPFmiHTd4KfKA.png 424w, https://cdn-images-1.medium.com/max/1024/1*5Axjs4qQ9TPFmiHTd4KfKA.png 848w, https://cdn-images-1.medium.com/max/1024/1*5Axjs4qQ9TPFmiHTd4KfKA.png 1272w, https://cdn-images-1.medium.com/max/1024/1*5Axjs4qQ9TPFmiHTd4KfKA.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">Figure reproduced from<a href="https://icml.cc/virtual/2020/poster/6429"> the&nbsp;paper</a>.</figcaption></figure></div><p>Also, it outperforms traditional embedding compression approaches like scalar quantization, product quantization (standard), and low-rank approximation.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://cdn-images-1.medium.com/max/745/1*rHvATuLPf4R7Pz6tqVWDjQ.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://cdn-images-1.medium.com/max/745/1*rHvATuLPf4R7Pz6tqVWDjQ.png 424w, https://cdn-images-1.medium.com/max/745/1*rHvATuLPf4R7Pz6tqVWDjQ.png 848w, https://cdn-images-1.medium.com/max/745/1*rHvATuLPf4R7Pz6tqVWDjQ.png 1272w, https://cdn-images-1.medium.com/max/745/1*rHvATuLPf4R7Pz6tqVWDjQ.png 1456w" sizes="100vw"><img src="https://cdn-images-1.medium.com/max/745/1*rHvATuLPf4R7Pz6tqVWDjQ.png" data-attrs="{&quot;src&quot;:&quot;https://cdn-images-1.medium.com/max/745/1*rHvATuLPf4R7Pz6tqVWDjQ.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://cdn-images-1.medium.com/max/745/1*rHvATuLPf4R7Pz6tqVWDjQ.png 424w, https://cdn-images-1.medium.com/max/745/1*rHvATuLPf4R7Pz6tqVWDjQ.png 848w, https://cdn-images-1.medium.com/max/745/1*rHvATuLPf4R7Pz6tqVWDjQ.png 1272w, https://cdn-images-1.medium.com/max/745/1*rHvATuLPf4R7Pz6tqVWDjQ.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">Figure reproduced from <a href="https://icml.cc/virtual/2020/poster/6429">the&nbsp;paper.</a></figcaption></figure></div><h4><a href="https://icml.cc/virtual/2020/poster/5787">Towards Accurate Post-training Network Quantization via Bit-Split and Stitching</a></h4><p><strong>Outstanding problem: </strong>Quantizing the neural network before training or while training (Quantization-aware training; with quantized operations embedded in the network) has to device special mechanisms for back-propagating through discrete quantized entities and needs domain expertise for hyperparameter tuning and also can&#8217;t work with models that are already&nbsp;trained.</p><p>On the other hand, post-training quantization avoids all of that but is shown to be ineffective for bit-widths lower than 8. TF-Lite is the previous state-of-the-art method that only uses 8-bit post-training quantization.</p><p><strong>Proposed solution:</strong> This work proposes an optimization process to calibrate the network with a few unlabeled data and find the optimal low-bit integers to replace the floating-point values in the weight tensors. If the bit-width desired is m, the task is to find the best possible m-bit integer [-2^m-1, +2^m-1] for every floating-point FP32 weight in the weight matrix of a layer (so that it doesn&#8217;t result in performance degradation).</p><p>As this calibration process has a lot of search-space to optimize over, this work proposes to break the m-bit optimization process into m bits and then perform the more tractable optimization step of these individual bits.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://cdn-images-1.medium.com/max/1024/1*TKSe6mrCQTCTmoBGR__Ecg.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://cdn-images-1.medium.com/max/1024/1*TKSe6mrCQTCTmoBGR__Ecg.png 424w, https://cdn-images-1.medium.com/max/1024/1*TKSe6mrCQTCTmoBGR__Ecg.png 848w, https://cdn-images-1.medium.com/max/1024/1*TKSe6mrCQTCTmoBGR__Ecg.png 1272w, https://cdn-images-1.medium.com/max/1024/1*TKSe6mrCQTCTmoBGR__Ecg.png 1456w" sizes="100vw"><img src="https://cdn-images-1.medium.com/max/1024/1*TKSe6mrCQTCTmoBGR__Ecg.png" data-attrs="{&quot;src&quot;:&quot;https://cdn-images-1.medium.com/max/1024/1*TKSe6mrCQTCTmoBGR__Ecg.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://cdn-images-1.medium.com/max/1024/1*TKSe6mrCQTCTmoBGR__Ecg.png 424w, https://cdn-images-1.medium.com/max/1024/1*TKSe6mrCQTCTmoBGR__Ecg.png 848w, https://cdn-images-1.medium.com/max/1024/1*TKSe6mrCQTCTmoBGR__Ecg.png 1272w, https://cdn-images-1.medium.com/max/1024/1*TKSe6mrCQTCTmoBGR__Ecg.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">Figure reproduced from <a href="https://icml.cc/virtual/2020/poster/5787">the&nbsp;paper</a>.</figcaption></figure></div><p>After the optimization, which could be analytically performed after sensible assumptions, the bits are stitched back by adding all the bits multiplied with their associated power 2 terms. Hence, this process is called bit-split and stitching.</p><p><strong>Results and conclusion:</strong> Compared to the previous state-of-the-art methodology of TF-Lite, the proposed method is effective in the 3-bit quantization of a network on the ImageNet&nbsp;dataset.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://cdn-images-1.medium.com/max/1024/1*b2PGvffvmjuplJyaTohr0Q.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://cdn-images-1.medium.com/max/1024/1*b2PGvffvmjuplJyaTohr0Q.png 424w, https://cdn-images-1.medium.com/max/1024/1*b2PGvffvmjuplJyaTohr0Q.png 848w, https://cdn-images-1.medium.com/max/1024/1*b2PGvffvmjuplJyaTohr0Q.png 1272w, https://cdn-images-1.medium.com/max/1024/1*b2PGvffvmjuplJyaTohr0Q.png 1456w" sizes="100vw"><img src="https://cdn-images-1.medium.com/max/1024/1*b2PGvffvmjuplJyaTohr0Q.png" data-attrs="{&quot;src&quot;:&quot;https://cdn-images-1.medium.com/max/1024/1*b2PGvffvmjuplJyaTohr0Q.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://cdn-images-1.medium.com/max/1024/1*b2PGvffvmjuplJyaTohr0Q.png 424w, https://cdn-images-1.medium.com/max/1024/1*b2PGvffvmjuplJyaTohr0Q.png 848w, https://cdn-images-1.medium.com/max/1024/1*b2PGvffvmjuplJyaTohr0Q.png 1272w, https://cdn-images-1.medium.com/max/1024/1*b2PGvffvmjuplJyaTohr0Q.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">Figure reproduced from <a href="https://icml.cc/virtual/2020/poster/5787">the&nbsp;paper</a>.</figcaption></figure></div><p>The proposed method also outperforms many recent methods for post-training quantization.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://cdn-images-1.medium.com/max/1024/1*78mOZ0m-8INzCffhxbA71w.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://cdn-images-1.medium.com/max/1024/1*78mOZ0m-8INzCffhxbA71w.png 424w, https://cdn-images-1.medium.com/max/1024/1*78mOZ0m-8INzCffhxbA71w.png 848w, https://cdn-images-1.medium.com/max/1024/1*78mOZ0m-8INzCffhxbA71w.png 1272w, https://cdn-images-1.medium.com/max/1024/1*78mOZ0m-8INzCffhxbA71w.png 1456w" sizes="100vw"><img src="https://cdn-images-1.medium.com/max/1024/1*78mOZ0m-8INzCffhxbA71w.png" data-attrs="{&quot;src&quot;:&quot;https://cdn-images-1.medium.com/max/1024/1*78mOZ0m-8INzCffhxbA71w.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://cdn-images-1.medium.com/max/1024/1*78mOZ0m-8INzCffhxbA71w.png 424w, https://cdn-images-1.medium.com/max/1024/1*78mOZ0m-8INzCffhxbA71w.png 848w, https://cdn-images-1.medium.com/max/1024/1*78mOZ0m-8INzCffhxbA71w.png 1272w, https://cdn-images-1.medium.com/max/1024/1*78mOZ0m-8INzCffhxbA71w.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">Figure reproduced from <a href="https://icml.cc/virtual/2020/poster/5787">the&nbsp;paper</a>.</figcaption></figure></div><p>Also, the proposed method achieves 4-bit weight quantization and 8-bit activation quantization of RetinaNet and Mask-RCNN for object detection and segmentation with only 0.8&#8211;1.2% mAP degradation.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://cdn-images-1.medium.com/max/645/1*8JtTmt0P_EkYUD3b_rnfcw.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://cdn-images-1.medium.com/max/645/1*8JtTmt0P_EkYUD3b_rnfcw.png 424w, https://cdn-images-1.medium.com/max/645/1*8JtTmt0P_EkYUD3b_rnfcw.png 848w, https://cdn-images-1.medium.com/max/645/1*8JtTmt0P_EkYUD3b_rnfcw.png 1272w, https://cdn-images-1.medium.com/max/645/1*8JtTmt0P_EkYUD3b_rnfcw.png 1456w" sizes="100vw"><img src="https://cdn-images-1.medium.com/max/645/1*8JtTmt0P_EkYUD3b_rnfcw.png" data-attrs="{&quot;src&quot;:&quot;https://cdn-images-1.medium.com/max/645/1*8JtTmt0P_EkYUD3b_rnfcw.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://cdn-images-1.medium.com/max/645/1*8JtTmt0P_EkYUD3b_rnfcw.png 424w, https://cdn-images-1.medium.com/max/645/1*8JtTmt0P_EkYUD3b_rnfcw.png 848w, https://cdn-images-1.medium.com/max/645/1*8JtTmt0P_EkYUD3b_rnfcw.png 1272w, https://cdn-images-1.medium.com/max/645/1*8JtTmt0P_EkYUD3b_rnfcw.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">Figure reproduced from<a href="https://icml.cc/virtual/2020/poster/5787"> the&nbsp;paper</a>.</figcaption></figure></div><h4><a href="https://neurips.cc/virtual/2020/public/poster_3f13cf4ddf6fc50c0d39a1d5aeb57dd8.html">Bayesian Bits: Unifying Quantization and&nbsp;Pruning</a></h4><p><strong>Outstanding problem:</strong> Learning channel-wise bit-widths to perform adaptive quantization is very hard. Reinforcement learning and hierarchical reinforcement learning approaches have tried to learn effective strategies but they are not effective enough. This work learns channel-wise bit-widths effectively.</p><p>Also, if one needs to perform both quantization and pruning, methods that unify both are more effective than applying separate methods one-by-one. This work unifies both quantization and&nbsp;pruning.</p><p>While a lot of mixed-precision quantization methods find arbitrary bit-widths for each weight tensor, thus making the network incompatible with standard GPU/TPU accelerators, the proposed method comes up with strategies with only powers-of-two bit-widths, making it hardware friendly.</p><p><strong>Proposed solution:</strong> First, this work proposes a formulation of M-bit quantization value as the addition of (M/2)-bit quantization value and M-bit residual error between (M/2)-bit quantization value and M-bit quantization value. Similarly, the (M/2)-bit quantization value can also be decomposed in terms of (M/4)-bit quantization value and the residual error between (M/2)-bit value and (M/4)-bit value.</p><p>Starting with 2-bit quantization, this decomposition allows for having binary gates that control whether to go for a higher bit-width or not, thus creating a mixed-precision strategy. This method learns this mixed-precision strategy, i.e., binary gates for each weight tensor in a variational inference framework with a few sensible approximations.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://cdn-images-1.medium.com/max/1024/1*_tL4-3jIYsKJfrrNS4G7bA.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://cdn-images-1.medium.com/max/1024/1*_tL4-3jIYsKJfrrNS4G7bA.png 424w, https://cdn-images-1.medium.com/max/1024/1*_tL4-3jIYsKJfrrNS4G7bA.png 848w, https://cdn-images-1.medium.com/max/1024/1*_tL4-3jIYsKJfrrNS4G7bA.png 1272w, https://cdn-images-1.medium.com/max/1024/1*_tL4-3jIYsKJfrrNS4G7bA.png 1456w" sizes="100vw"><img src="https://cdn-images-1.medium.com/max/1024/1*_tL4-3jIYsKJfrrNS4G7bA.png" data-attrs="{&quot;src&quot;:&quot;https://cdn-images-1.medium.com/max/1024/1*_tL4-3jIYsKJfrrNS4G7bA.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://cdn-images-1.medium.com/max/1024/1*_tL4-3jIYsKJfrrNS4G7bA.png 424w, https://cdn-images-1.medium.com/max/1024/1*_tL4-3jIYsKJfrrNS4G7bA.png 848w, https://cdn-images-1.medium.com/max/1024/1*_tL4-3jIYsKJfrrNS4G7bA.png 1272w, https://cdn-images-1.medium.com/max/1024/1*_tL4-3jIYsKJfrrNS4G7bA.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">Figure reproduced from <a href="https://neurips.cc/virtual/2020/public/poster_3f13cf4ddf6fc50c0d39a1d5aeb57dd8.html">the&nbsp;paper</a>.</figcaption></figure></div><p>By starting with 0-bit quantization (practically pruning) this method unifies quantization and pruning in an effective way.</p><p><strong>Results and conclusion:</strong> The proposed method outperforms many existing fixed-precision and mixed-precision methods on MNIST and CIFAR-10 datasets.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://cdn-images-1.medium.com/max/1024/1*tmG1SK6hhEsSEs864pbRWg.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://cdn-images-1.medium.com/max/1024/1*tmG1SK6hhEsSEs864pbRWg.png 424w, https://cdn-images-1.medium.com/max/1024/1*tmG1SK6hhEsSEs864pbRWg.png 848w, https://cdn-images-1.medium.com/max/1024/1*tmG1SK6hhEsSEs864pbRWg.png 1272w, https://cdn-images-1.medium.com/max/1024/1*tmG1SK6hhEsSEs864pbRWg.png 1456w" sizes="100vw"><img src="https://cdn-images-1.medium.com/max/1024/1*tmG1SK6hhEsSEs864pbRWg.png" data-attrs="{&quot;src&quot;:&quot;https://cdn-images-1.medium.com/max/1024/1*tmG1SK6hhEsSEs864pbRWg.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://cdn-images-1.medium.com/max/1024/1*tmG1SK6hhEsSEs864pbRWg.png 424w, https://cdn-images-1.medium.com/max/1024/1*tmG1SK6hhEsSEs864pbRWg.png 848w, https://cdn-images-1.medium.com/max/1024/1*tmG1SK6hhEsSEs864pbRWg.png 1272w, https://cdn-images-1.medium.com/max/1024/1*tmG1SK6hhEsSEs864pbRWg.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">Figure reproduced from <a href="https://neurips.cc/virtual/2020/public/poster_3f13cf4ddf6fc50c0d39a1d5aeb57dd8.html">the&nbsp;paper.</a></figcaption></figure></div><p>Also, it outperforms the existing state-of-the-art methods on the ImageNet dataset with ResNet-18 and MobileNet V2 architecture.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://cdn-images-1.medium.com/max/1024/1*ja8GwIw0LRr8jQsofRHJNw.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://cdn-images-1.medium.com/max/1024/1*ja8GwIw0LRr8jQsofRHJNw.png 424w, https://cdn-images-1.medium.com/max/1024/1*ja8GwIw0LRr8jQsofRHJNw.png 848w, https://cdn-images-1.medium.com/max/1024/1*ja8GwIw0LRr8jQsofRHJNw.png 1272w, https://cdn-images-1.medium.com/max/1024/1*ja8GwIw0LRr8jQsofRHJNw.png 1456w" sizes="100vw"><img src="https://cdn-images-1.medium.com/max/1024/1*ja8GwIw0LRr8jQsofRHJNw.png" data-attrs="{&quot;src&quot;:&quot;https://cdn-images-1.medium.com/max/1024/1*ja8GwIw0LRr8jQsofRHJNw.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://cdn-images-1.medium.com/max/1024/1*ja8GwIw0LRr8jQsofRHJNw.png 424w, https://cdn-images-1.medium.com/max/1024/1*ja8GwIw0LRr8jQsofRHJNw.png 848w, https://cdn-images-1.medium.com/max/1024/1*ja8GwIw0LRr8jQsofRHJNw.png 1272w, https://cdn-images-1.medium.com/max/1024/1*ja8GwIw0LRr8jQsofRHJNw.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">Figure reproduced from <a href="https://neurips.cc/virtual/2020/public/poster_3f13cf4ddf6fc50c0d39a1d5aeb57dd8.html">the&nbsp;paper</a>.</figcaption></figure></div><h4><a href="https://icml.cc/virtual/2020/poster/6482">Up or Down? Adaptive Rounding for Post-Training Quantization</a></h4><p><strong>Outstanding problem:</strong> It has been previously shown that when quantizing a value from higher precision to lower precision, the rounding mechanism used has a lot of impact on performance.</p><p>Specifically, the performances of quantization with the nearest-value rounding mechanism, always ceiling or always flooring rounding mechanism, and stochastic rounding mechanism show a lot of variances. See the figure&nbsp;below.</p><p>(Stochastic rounding is a mechanism of rounding a value to a lower-precision value with the probability of (1- px) where px is the proximity of that particular value and the lower-precision value.)</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://cdn-images-1.medium.com/max/666/1*rxi3B7zjorscTW7AQSix5g.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://cdn-images-1.medium.com/max/666/1*rxi3B7zjorscTW7AQSix5g.png 424w, https://cdn-images-1.medium.com/max/666/1*rxi3B7zjorscTW7AQSix5g.png 848w, https://cdn-images-1.medium.com/max/666/1*rxi3B7zjorscTW7AQSix5g.png 1272w, https://cdn-images-1.medium.com/max/666/1*rxi3B7zjorscTW7AQSix5g.png 1456w" sizes="100vw"><img src="https://cdn-images-1.medium.com/max/666/1*rxi3B7zjorscTW7AQSix5g.png" data-attrs="{&quot;src&quot;:&quot;https://cdn-images-1.medium.com/max/666/1*rxi3B7zjorscTW7AQSix5g.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://cdn-images-1.medium.com/max/666/1*rxi3B7zjorscTW7AQSix5g.png 424w, https://cdn-images-1.medium.com/max/666/1*rxi3B7zjorscTW7AQSix5g.png 848w, https://cdn-images-1.medium.com/max/666/1*rxi3B7zjorscTW7AQSix5g.png 1272w, https://cdn-images-1.medium.com/max/666/1*rxi3B7zjorscTW7AQSix5g.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">Figure reproduced from <a href="https://icml.cc/virtual/2020/poster/6482">the&nbsp;paper</a>.</figcaption></figure></div><p>This work proposes a better rounding mechanism post-training with just a small calibration on a few unlabelled examples.</p><p><strong>Proposed solution:</strong> This work investigated whether to round up or round down for each weight by formulating it as an optimization problem of reducing the reconstruction loss of weights of a&nbsp;network.</p><p>This work set out to make the inherently discrete optimization problem into a continuous optimization problem with better formulations and sensible approximations based on Taylor series expansion.</p><p><strong>Results and conclusion:</strong> The proposed (seemingly simple) method shows considerable effectiveness at post-training quantization and outperforms existing post-training quantization methods.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://cdn-images-1.medium.com/max/1024/1*YLBLnxUh8iudQ8v1IXB8TQ.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://cdn-images-1.medium.com/max/1024/1*YLBLnxUh8iudQ8v1IXB8TQ.png 424w, https://cdn-images-1.medium.com/max/1024/1*YLBLnxUh8iudQ8v1IXB8TQ.png 848w, https://cdn-images-1.medium.com/max/1024/1*YLBLnxUh8iudQ8v1IXB8TQ.png 1272w, https://cdn-images-1.medium.com/max/1024/1*YLBLnxUh8iudQ8v1IXB8TQ.png 1456w" sizes="100vw"><img src="https://cdn-images-1.medium.com/max/1024/1*YLBLnxUh8iudQ8v1IXB8TQ.png" data-attrs="{&quot;src&quot;:&quot;https://cdn-images-1.medium.com/max/1024/1*YLBLnxUh8iudQ8v1IXB8TQ.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://cdn-images-1.medium.com/max/1024/1*YLBLnxUh8iudQ8v1IXB8TQ.png 424w, https://cdn-images-1.medium.com/max/1024/1*YLBLnxUh8iudQ8v1IXB8TQ.png 848w, https://cdn-images-1.medium.com/max/1024/1*YLBLnxUh8iudQ8v1IXB8TQ.png 1272w, https://cdn-images-1.medium.com/max/1024/1*YLBLnxUh8iudQ8v1IXB8TQ.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">Figure reproduced from <a href="https://icml.cc/virtual/2020/poster/6482">the&nbsp;paper</a>.</figcaption></figure></div><p>The authors also show that the effectiveness of their method increases with the number of unlabeled examples used for calibration across different datasets.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://cdn-images-1.medium.com/max/642/1*n7wczONkzkTSZ7fJet51AQ.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://cdn-images-1.medium.com/max/642/1*n7wczONkzkTSZ7fJet51AQ.png 424w, https://cdn-images-1.medium.com/max/642/1*n7wczONkzkTSZ7fJet51AQ.png 848w, https://cdn-images-1.medium.com/max/642/1*n7wczONkzkTSZ7fJet51AQ.png 1272w, https://cdn-images-1.medium.com/max/642/1*n7wczONkzkTSZ7fJet51AQ.png 1456w" sizes="100vw"><img src="https://cdn-images-1.medium.com/max/642/1*n7wczONkzkTSZ7fJet51AQ.png" data-attrs="{&quot;src&quot;:&quot;https://cdn-images-1.medium.com/max/642/1*n7wczONkzkTSZ7fJet51AQ.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://cdn-images-1.medium.com/max/642/1*n7wczONkzkTSZ7fJet51AQ.png 424w, https://cdn-images-1.medium.com/max/642/1*n7wczONkzkTSZ7fJet51AQ.png 848w, https://cdn-images-1.medium.com/max/642/1*n7wczONkzkTSZ7fJet51AQ.png 1272w, https://cdn-images-1.medium.com/max/642/1*n7wczONkzkTSZ7fJet51AQ.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">Figure reproduced from <a href="https://icml.cc/virtual/2020/poster/6482">the&nbsp;paper</a>.</figcaption></figure></div><h4>Conclusion</h4><p>Research on Neural Network Quantization, more generally NN compression, is evolving to be more scientific and rigorous. One of the reasons is, undoubtedly, the interaction between the wide adoption of deep learning methods in Computer Vision &amp; NLP and elsewhere and the increasing amount of memory, energy, and computational resources required for the state-of-the-art methods.</p><p>Because of the research in NN Quantization in 2020, we have different methods for layer-wise, channel-wise quantization methods rather than network-wise quantization methods, we have better post-training quantization techniques which are better than methods offered by TF-Lite, and we have approaches that combined pruning and quantization under one framework and give the advantages of both the techniques.</p><p>Going into the future we will see these advanced and efficient methods made more accessible to developers and researchers by including them into tools like TF-Lite and other NN compression libraries. And it is exciting to see the research on this topic in 2021 which has to beat already great efficient approaches introduced this&nbsp;year.</p><p><em>Editor&#8217;s Note: <a href="https://heartbeat.comet.ml/">Heartbeat</a> is a contributor-driven online publication and community dedicated to providing premier educational resources for data science, machine learning, and deep learning practitioners. We&#8217;re committed to supporting and inspiring developers and engineers from all walks of&nbsp;life.</em></p><p><em>Editorially independent, Heartbeat is sponsored and published by <a href="http://comet.ml/?utm_campaign=heartbeat-statement&amp;utm_source=blog&amp;utm_medium=medium">Comet</a>, an MLOps platform that enables data scientists &amp; ML teams to track, compare, explain, &amp; optimize their experiments. We pay our contributors, and we don&#8217;t sell&nbsp;ads.</em></p><p><em>If you&#8217;d like to contribute, head on over to our<a href="https://heartbeat.fritz.ai/call-for-contributors-october-2018-update-fee7f5b80f3e"> call for contributors</a>. You can also sign up to receive our weekly newsletters (<a href="https://www.deeplearningweekly.com/">Deep Learning Weekly</a> and the <a href="https://info.comet.ml/newsletter-signup/">Comet Newsletter</a>), join us on<a href="https://join.slack.com/t/fritz-ai-community/shared_invite/enQtNTY5NDM2MTQwMTgwLWU4ZDEwNTAxYWE2YjIxZDllMTcxMWE4MGFhNDk5Y2QwNTcxYzEyNWZmZWEwMzE4NTFkOWY2NTM0OGQwYjM5Y2U"> </a><a href="https://join.slack.com/t/cometml/shared_invite/zt-49v4zxxz-qHcTeyrMEzqZc5lQb9hgvw">Slack</a>, and follow Comet on <a href="https://twitter.com/Cometml">Twitter</a> and <a href="https://www.linkedin.com/company/comet-ml/">LinkedIn</a> for resources, events, and much more that will help you build better ML models,&nbsp;faster.</em></p><div><hr></div><p><a href="https://heartbeat.comet.ml/neural-network-quantization-research-review-2020-6d72b06f09b1">Neural Network Quantization Research Review</a> was originally published in <a href="https://heartbeat.comet.ml">Heartbeat</a> on Medium, where people are continuing the conversation by highlighting and responding to this story.</p>]]></content:encoded></item><item><title><![CDATA[Neural Network Pruning Research Review 2020]]></title><description><![CDATA[The latest advances and insights towards Neural Network model compression with Pruning]]></description><link>https://www.aipapertrails.com/p/neural-network-pruning-research-review-2020-bc21a77f0295</link><guid isPermaLink="false">https://www.aipapertrails.com/p/neural-network-pruning-research-review-2020-bc21a77f0295</guid><dc:creator><![CDATA[Prakash Kagitha]]></dc:creator><pubDate>Thu, 24 Dec 2020 14:13:34 GMT</pubDate><enclosure url="https://cdn-images-1.medium.com/max/685/1*p2AMGLzaDaFI4eBLvCTKRg.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<h4>The latest advances and insights towards Neural Network model compression with&nbsp;Pruning</h4><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://cdn-images-1.medium.com/max/685/1*p2AMGLzaDaFI4eBLvCTKRg.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://cdn-images-1.medium.com/max/685/1*p2AMGLzaDaFI4eBLvCTKRg.png 424w, https://cdn-images-1.medium.com/max/685/1*p2AMGLzaDaFI4eBLvCTKRg.png 848w, https://cdn-images-1.medium.com/max/685/1*p2AMGLzaDaFI4eBLvCTKRg.png 1272w, https://cdn-images-1.medium.com/max/685/1*p2AMGLzaDaFI4eBLvCTKRg.png 1456w" sizes="100vw"><img src="https://cdn-images-1.medium.com/max/685/1*p2AMGLzaDaFI4eBLvCTKRg.png" data-attrs="{&quot;src&quot;:&quot;https://cdn-images-1.medium.com/max/685/1*p2AMGLzaDaFI4eBLvCTKRg.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://cdn-images-1.medium.com/max/685/1*p2AMGLzaDaFI4eBLvCTKRg.png 424w, https://cdn-images-1.medium.com/max/685/1*p2AMGLzaDaFI4eBLvCTKRg.png 848w, https://cdn-images-1.medium.com/max/685/1*p2AMGLzaDaFI4eBLvCTKRg.png 1272w, https://cdn-images-1.medium.com/max/685/1*p2AMGLzaDaFI4eBLvCTKRg.png 1456w" sizes="100vw" fetchpriority="high"></picture><div></div></div></a></figure></div><p>Neural Network (NN) Pruning is a task of reducing the size of a Neural Network by removing some of its parameters/weights.</p><p>Pruning is often performed with the objective of reducing the memory, computational, and energy bandwidths required for training and deploying NN models which are notorious for their large model size, computational expense, and energy consumption.</p><p>Particularly when deploying NN models on mobiles or edge devices, Pruning, and model compression in general, is desirable and often the only plausible way to deploy as the memory, energy, and computational bandwidths are very&nbsp;limited.</p><p>But, one can ask, why not use potentially infinite virtual memory and computational power from the cloud? While a lot of NN models are running on the cloud even now, latency is not low enough for mobile/edge devices which hinders utility and requires data to be transferred to the cloud, which raises several privacy concerns.</p><p>The year 2020 has been a great year for NN Pruning research and this article discusses six new approaches/insights published at premier peer-reviewed conferences ICLR, ICML, and NeurIPS in&nbsp;2020.</p><h4>Preliminaries of NN Pruning&nbsp;research</h4><p><em>The workflow of a typical Pruning method:</em> A set of parameters of a trained model are zeroed out according to a pre-defined score (usually absolute magnitude) and the network remained is trained again (retraining), thus approximating the accuracy of the original model with this sparse, pruned model requiring drastically lower, when compared to the original model, memory, computational and energy bandwidths because of the fewer parameters.</p><p>To understand the advances in Pruning research it helps to place new techniques as different types across different dimensions (ways to classify) and their unique strengths and weaknesses. The following are a few standard dimensions Pruning methods could be classified into:</p><ol><li><p><strong>Structured vs. Unstructured Pruning</strong></p></li></ol><p>If NN weights are pruned individually in the pruning process, then it is called Unstructured Pruning.</p><p>Zeroing out the parameters randomly gives memory efficiency (model stored in sparse matrices) but may not always give a better computational performance, as we end up doing the same number of matrix multiplications as before. Because matrix dimensions didn&#8217;t change, they are just sparse. Although we could get computational advantages by replacing dense matrix multiplications with sparse matrix multiplications, it is not trivial to accelerate sparse operations on traditional GPUs/TPUs.</p><p>Alternately, in Structured Pruning, the parameters are removed in a structured way to reduce the overall computation required. For example, some of the channels of CNN or neurons in a feedforward layer are removed, which directly reduces the computation.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://cdn-images-1.medium.com/max/685/1*p2AMGLzaDaFI4eBLvCTKRg.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://cdn-images-1.medium.com/max/685/1*p2AMGLzaDaFI4eBLvCTKRg.png 424w, https://cdn-images-1.medium.com/max/685/1*p2AMGLzaDaFI4eBLvCTKRg.png 848w, https://cdn-images-1.medium.com/max/685/1*p2AMGLzaDaFI4eBLvCTKRg.png 1272w, https://cdn-images-1.medium.com/max/685/1*p2AMGLzaDaFI4eBLvCTKRg.png 1456w" sizes="100vw"><img src="https://cdn-images-1.medium.com/max/685/1*p2AMGLzaDaFI4eBLvCTKRg.png" data-attrs="{&quot;src&quot;:&quot;https://cdn-images-1.medium.com/max/685/1*p2AMGLzaDaFI4eBLvCTKRg.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://cdn-images-1.medium.com/max/685/1*p2AMGLzaDaFI4eBLvCTKRg.png 424w, https://cdn-images-1.medium.com/max/685/1*p2AMGLzaDaFI4eBLvCTKRg.png 848w, https://cdn-images-1.medium.com/max/685/1*p2AMGLzaDaFI4eBLvCTKRg.png 1272w, https://cdn-images-1.medium.com/max/685/1*p2AMGLzaDaFI4eBLvCTKRg.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">Structured Pruning (pruning neurons) and Unstructured Pruning (pruning connections) illustrated. Image reproduced from <a href="https://www.researchgate.net/publication/334489669_Deep_Learning_With_Edge_Computing_A_Review">the&nbsp;source</a>.</figcaption></figure></div><p><strong>2. Scoring of parameters</strong></p><p>Pruning methods may differ in the method used to score each parameter which is used to choose one parameter over another. The absolute magnitude scoring method is the standard but one could come up with a new scoring method that would increase the efficiency of&nbsp;Pruning.</p><p><strong>3. One-shot Vs Iterative</strong></p><p>While performing Pruning, Rather than pruning the desired amount at once, which is called One-shot Pruning, some approaches repeat the process of pruning the network to some extent and retraining it until the desired pruning rate is reached, which is called Iterative Pruning.</p><p><strong>4. Scheduling of pruning and fine-tuning</strong></p><p>In the paradigm of Iterative Pruning, previous work showed that learning or designing when to prune and how much helps. A Schedule of pruning specifies the ratio of pruning after each epoch along with the number of epochs to retrain the model whenever the model is&nbsp;pruned.</p><h4><a href="https://nips.cc/virtual/2020/public/poster_703957b6dd9e3a7980e040bee50ded65.html">Neuron-level Structured Pruning using Polarization Regularizer</a> (NeurIPS&nbsp;2020)</h4><p><strong>Outstanding problem:</strong> In one of the variants of Structured Pruning, to remove entire filters in CNN or all weights associated with a particular neuron, a learnable scaling factor is used for each filter/neuron which is multiplied by the weights of that filter/neuron in the forward&nbsp;pass.</p><p>And to prune the model, all the filters/neurons with scaling factor less than a threshold are removed, thus accomplishing the Structured Pruning.</p><p>While training, an L1 regularizer is added to the main training loss to push all the scaling factors to zero with the aim of having scaling factors above the threshold only for filters/neurons that are absolutely necessary.</p><p>This work shows that the L1 regularizer is not efficient<strong> </strong>in contrasting between useful and not useful filters/neurons and tries to address this inefficiency.</p><p><strong>Proposed solution: </strong>L1 regularizer tries to push all scaling factor values to zero which is not consistent with the intuition that only some of the filters are not useful while others are useful. It is hard to pick a threshold value that would clearly separate useful filters among other filters as seen in the left plot of the figure&nbsp;below.</p><p>This work proposes Polarization Regularizer, which pushes the scaling factors towards the extremes, either 0 or 1, which aligns better with the objective of removing some of the filters and while retaining others.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://cdn-images-1.medium.com/max/1024/1*9uCB0RMGwO5fE30gCRSxvw.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://cdn-images-1.medium.com/max/1024/1*9uCB0RMGwO5fE30gCRSxvw.png 424w, https://cdn-images-1.medium.com/max/1024/1*9uCB0RMGwO5fE30gCRSxvw.png 848w, https://cdn-images-1.medium.com/max/1024/1*9uCB0RMGwO5fE30gCRSxvw.png 1272w, https://cdn-images-1.medium.com/max/1024/1*9uCB0RMGwO5fE30gCRSxvw.png 1456w" sizes="100vw"><img src="https://cdn-images-1.medium.com/max/1024/1*9uCB0RMGwO5fE30gCRSxvw.png" data-attrs="{&quot;src&quot;:&quot;https://cdn-images-1.medium.com/max/1024/1*9uCB0RMGwO5fE30gCRSxvw.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://cdn-images-1.medium.com/max/1024/1*9uCB0RMGwO5fE30gCRSxvw.png 424w, https://cdn-images-1.medium.com/max/1024/1*9uCB0RMGwO5fE30gCRSxvw.png 848w, https://cdn-images-1.medium.com/max/1024/1*9uCB0RMGwO5fE30gCRSxvw.png 1272w, https://cdn-images-1.medium.com/max/1024/1*9uCB0RMGwO5fE30gCRSxvw.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">It is clear from the above plot that L1 Regularization yields scaling factors which makes it very hard to come up with an optimal threshold. While Polarization yields scaling factors which makes it easy to find an optimal threshold. Image reproduced from <a href="https://nips.cc/virtual/2020/public/poster_703957b6dd9e3a7980e040bee50ded65.html">the&nbsp;source</a>.</figcaption></figure></div><p>Specifically, this work adds a term to L1 regularizer which is maximum when all scaling factors are near to each other, pushing scaling factors towards two extremes.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://cdn-images-1.medium.com/max/1024/1*D1xUPIh5DBcXJJfgEjIr_g.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://cdn-images-1.medium.com/max/1024/1*D1xUPIh5DBcXJJfgEjIr_g.png 424w, https://cdn-images-1.medium.com/max/1024/1*D1xUPIh5DBcXJJfgEjIr_g.png 848w, https://cdn-images-1.medium.com/max/1024/1*D1xUPIh5DBcXJJfgEjIr_g.png 1272w, https://cdn-images-1.medium.com/max/1024/1*D1xUPIh5DBcXJJfgEjIr_g.png 1456w" sizes="100vw"><img src="https://cdn-images-1.medium.com/max/1024/1*D1xUPIh5DBcXJJfgEjIr_g.png" data-attrs="{&quot;src&quot;:&quot;https://cdn-images-1.medium.com/max/1024/1*D1xUPIh5DBcXJJfgEjIr_g.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://cdn-images-1.medium.com/max/1024/1*D1xUPIh5DBcXJJfgEjIr_g.png 424w, https://cdn-images-1.medium.com/max/1024/1*D1xUPIh5DBcXJJfgEjIr_g.png 848w, https://cdn-images-1.medium.com/max/1024/1*D1xUPIh5DBcXJJfgEjIr_g.png 1272w, https://cdn-images-1.medium.com/max/1024/1*D1xUPIh5DBcXJJfgEjIr_g.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">The second term in equation two pushes the scaling factors to either of the extremes because its value is higher when all the scaling factors are the same. Image reproduced from <a href="https://nips.cc/virtual/2020/public/poster_703957b6dd9e3a7980e040bee50ded65.html">the&nbsp;source</a>.</figcaption></figure></div><p><strong>Results and conclusions:</strong> The proposed method is better than a lot of methods by reducing the FLOPs required by<strong> </strong>ResNet-50 by 54% with just a 0.52 drop in accuracy. The authors even show that they can prune already efficient MobileNet V2 by 28% with just a 0.2 reduction in accuracy.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://cdn-images-1.medium.com/max/1024/1*1Q0WpqQw08QSeiRegot84g.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://cdn-images-1.medium.com/max/1024/1*1Q0WpqQw08QSeiRegot84g.png 424w, https://cdn-images-1.medium.com/max/1024/1*1Q0WpqQw08QSeiRegot84g.png 848w, https://cdn-images-1.medium.com/max/1024/1*1Q0WpqQw08QSeiRegot84g.png 1272w, https://cdn-images-1.medium.com/max/1024/1*1Q0WpqQw08QSeiRegot84g.png 1456w" sizes="100vw"><img src="https://cdn-images-1.medium.com/max/1024/1*1Q0WpqQw08QSeiRegot84g.png" data-attrs="{&quot;src&quot;:&quot;https://cdn-images-1.medium.com/max/1024/1*1Q0WpqQw08QSeiRegot84g.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://cdn-images-1.medium.com/max/1024/1*1Q0WpqQw08QSeiRegot84g.png 424w, https://cdn-images-1.medium.com/max/1024/1*1Q0WpqQw08QSeiRegot84g.png 848w, https://cdn-images-1.medium.com/max/1024/1*1Q0WpqQw08QSeiRegot84g.png 1272w, https://cdn-images-1.medium.com/max/1024/1*1Q0WpqQw08QSeiRegot84g.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">FLOPs required for ResNet-50 are reduced by 54% and by 28% for MobileNet V2. Image reproduced from <a href="https://nips.cc/virtual/2020/public/poster_703957b6dd9e3a7980e040bee50ded65.html">the&nbsp;source</a>.</figcaption></figure></div><p>It also allows for aggressive pruning, where structured pruning methods are not effective, without compromising on the accuracy compared to existing&nbsp;models.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://cdn-images-1.medium.com/max/1024/1*3P8TwvMXHyy-OHi-Z3JNkg.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://cdn-images-1.medium.com/max/1024/1*3P8TwvMXHyy-OHi-Z3JNkg.png 424w, https://cdn-images-1.medium.com/max/1024/1*3P8TwvMXHyy-OHi-Z3JNkg.png 848w, https://cdn-images-1.medium.com/max/1024/1*3P8TwvMXHyy-OHi-Z3JNkg.png 1272w, https://cdn-images-1.medium.com/max/1024/1*3P8TwvMXHyy-OHi-Z3JNkg.png 1456w" sizes="100vw"><img src="https://cdn-images-1.medium.com/max/1024/1*3P8TwvMXHyy-OHi-Z3JNkg.png" data-attrs="{&quot;src&quot;:&quot;https://cdn-images-1.medium.com/max/1024/1*3P8TwvMXHyy-OHi-Z3JNkg.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://cdn-images-1.medium.com/max/1024/1*3P8TwvMXHyy-OHi-Z3JNkg.png 424w, https://cdn-images-1.medium.com/max/1024/1*3P8TwvMXHyy-OHi-Z3JNkg.png 848w, https://cdn-images-1.medium.com/max/1024/1*3P8TwvMXHyy-OHi-Z3JNkg.png 1272w, https://cdn-images-1.medium.com/max/1024/1*3P8TwvMXHyy-OHi-Z3JNkg.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">Results of aggressive pruning of ResNet variants and MobileNet V2. Image reproduced from <a href="https://nips.cc/virtual/2020/public/poster_703957b6dd9e3a7980e040bee50ded65.html">the&nbsp;source</a>.</figcaption></figure></div><h4><a href="https://iclr.cc/virtual_2020/poster_S1gSj0NKvB.html">Comparing Rewinding and Fine-tuning in Neural Network Pruning</a> (ICLR&nbsp;2020)</h4><p><strong>Outstanding problem:</strong> After training and pruning, there is no evidence from the literature as to which form of retraining is most efficient: Rewinding or Fine-tuning.</p><p><strong>Proposed solution</strong>: This work performs comprehensive empirical studies evaluating the efficiency of Rewinding and Fine-tuning and finds that Rewinding is more efficient.</p><p>After training and pruning, if the remaining network is retrained with the trained network, then it is called Fine-tuning. Note that while fine-tuning we will continue using the same learning rate schedule as though it is further training.</p><p>If the remaining network is retrained with the weights at initialization (by rewinding) rather than trained weights, it is called Rewinding. Note that while Rewinding, as the weights are actual initialization weights of the network before training, the<strong> </strong>learning rate schedule starts afresh like it is the start of the training, only now with the small pruned&nbsp;network.</p><p>Also, the authors propose a new retraining technique called Learning Rate Rewinding to replace standard weight rewinding and is shown to out-perform both Rewinding and Fine-tuning. Learning Rate Rewinding is a hybrid of Fine-tuning and rewinding, where trained weights are used to retrain as in Fine-tuning but the learning rate schedule is rewinded as in Rewinding.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://cdn-images-1.medium.com/max/957/1*50OJHnfBlhGcf1og3RyyVg.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://cdn-images-1.medium.com/max/957/1*50OJHnfBlhGcf1og3RyyVg.png 424w, https://cdn-images-1.medium.com/max/957/1*50OJHnfBlhGcf1og3RyyVg.png 848w, https://cdn-images-1.medium.com/max/957/1*50OJHnfBlhGcf1og3RyyVg.png 1272w, https://cdn-images-1.medium.com/max/957/1*50OJHnfBlhGcf1og3RyyVg.png 1456w" sizes="100vw"><img src="https://cdn-images-1.medium.com/max/957/1*50OJHnfBlhGcf1og3RyyVg.png" data-attrs="{&quot;src&quot;:&quot;https://cdn-images-1.medium.com/max/957/1*50OJHnfBlhGcf1og3RyyVg.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://cdn-images-1.medium.com/max/957/1*50OJHnfBlhGcf1og3RyyVg.png 424w, https://cdn-images-1.medium.com/max/957/1*50OJHnfBlhGcf1og3RyyVg.png 848w, https://cdn-images-1.medium.com/max/957/1*50OJHnfBlhGcf1og3RyyVg.png 1272w, https://cdn-images-1.medium.com/max/957/1*50OJHnfBlhGcf1og3RyyVg.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">The step-by-step process of how to perform Learning Rate Rewinding. Image reproduced from <a href="https://iclr.cc/virtual_2020/poster_S1gSj0NKvB.html">the&nbsp;source</a>.</figcaption></figure></div><p><strong>Results and conclusion:</strong> It is shown that learning rate schedule and Rewinding is more efficient than Fine-tuning on CIFAR-10, ImageNet, and WMT16 datasets with standard architecture. This behavior is shown to be consistent whether it is structured or unstructured pruning, and whether it is one-shot or iterative pruning.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://cdn-images-1.medium.com/max/933/1*DSOs2zeVbhJ8IlrIKLepBw.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://cdn-images-1.medium.com/max/933/1*DSOs2zeVbhJ8IlrIKLepBw.png 424w, https://cdn-images-1.medium.com/max/933/1*DSOs2zeVbhJ8IlrIKLepBw.png 848w, https://cdn-images-1.medium.com/max/933/1*DSOs2zeVbhJ8IlrIKLepBw.png 1272w, https://cdn-images-1.medium.com/max/933/1*DSOs2zeVbhJ8IlrIKLepBw.png 1456w" sizes="100vw"><img src="https://cdn-images-1.medium.com/max/933/1*DSOs2zeVbhJ8IlrIKLepBw.png" data-attrs="{&quot;src&quot;:&quot;https://cdn-images-1.medium.com/max/933/1*DSOs2zeVbhJ8IlrIKLepBw.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://cdn-images-1.medium.com/max/933/1*DSOs2zeVbhJ8IlrIKLepBw.png 424w, https://cdn-images-1.medium.com/max/933/1*DSOs2zeVbhJ8IlrIKLepBw.png 848w, https://cdn-images-1.medium.com/max/933/1*DSOs2zeVbhJ8IlrIKLepBw.png 1272w, https://cdn-images-1.medium.com/max/933/1*DSOs2zeVbhJ8IlrIKLepBw.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">Learning Rate Rewinding proposed in this work outperforms both weight Rewinding and Fine-tuning while Weight Rewinding outperforms Fine-tuning. Image reproduced from <a href="https://iclr.cc/virtual_2020/poster_S1gSj0NKvB.html">the&nbsp;source</a>.</figcaption></figure></div><h4><a href="https://nips.cc/virtual/2020/public/poster_ccb1d45fb76f7c5a0bf619f979c6cf36.html">Pruning Filter in Filter</a> (NeurIPS&nbsp;2020)</h4><p><strong>Outstanding problem:</strong> Unstructured Pruning achieves more sparsity with minimal accuracy drop but achieving computational efficiency with this method is not straightforward, as sparse weights don&#8217;t necessarily reduce the number of matrix multiplications and sparse matrix multiplications are not usually accelerated by the standard GPUs/TPUs.</p><p>On the other hand, Structured Pruning, which is removing entire filters or channels, reduces computation but doesn't afford enough sparsity compared to Unstructured Pruning&nbsp;methods.</p><p>Balancing the tradeoff between structure and sparsity is not fully exploited. This work shows the importance of this with an observation that only some parts of a particular filter are useful. The authors prove that taking away the entire filter is not efficient.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://cdn-images-1.medium.com/max/1024/1*H_cJbtzk2bmzbCmTmFhqhQ.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://cdn-images-1.medium.com/max/1024/1*H_cJbtzk2bmzbCmTmFhqhQ.png 424w, https://cdn-images-1.medium.com/max/1024/1*H_cJbtzk2bmzbCmTmFhqhQ.png 848w, https://cdn-images-1.medium.com/max/1024/1*H_cJbtzk2bmzbCmTmFhqhQ.png 1272w, https://cdn-images-1.medium.com/max/1024/1*H_cJbtzk2bmzbCmTmFhqhQ.png 1456w" sizes="100vw"><img src="https://cdn-images-1.medium.com/max/1024/1*H_cJbtzk2bmzbCmTmFhqhQ.png" data-attrs="{&quot;src&quot;:&quot;https://cdn-images-1.medium.com/max/1024/1*H_cJbtzk2bmzbCmTmFhqhQ.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://cdn-images-1.medium.com/max/1024/1*H_cJbtzk2bmzbCmTmFhqhQ.png 424w, https://cdn-images-1.medium.com/max/1024/1*H_cJbtzk2bmzbCmTmFhqhQ.png 848w, https://cdn-images-1.medium.com/max/1024/1*H_cJbtzk2bmzbCmTmFhqhQ.png 1272w, https://cdn-images-1.medium.com/max/1024/1*H_cJbtzk2bmzbCmTmFhqhQ.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">Optimal pruning of the weights in each filter in VGG19. It shows that removing or retaining the entire filter is not optimal. The current work explores this optimality while retaining the computational efficiency of structured pruning methods. Image reproduced from <a href="https://nips.cc/virtual/2020/public/poster_ccb1d45fb76f7c5a0bf619f979c6cf36.html">the&nbsp;source</a>.</figcaption></figure></div><p><strong>Proposed solution: </strong>This work proposes a method called Filter Skeleton which is a strip-wise pruning method that is more granular than Structured Pruning, but could be reduced into fewer matrix multiplications, thus giving more sparsity while retaining the computational advantages of Structured Pruning&nbsp;methods.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://cdn-images-1.medium.com/max/788/1*QBWxO_aDKvDWXxVjS-t8iQ.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://cdn-images-1.medium.com/max/788/1*QBWxO_aDKvDWXxVjS-t8iQ.png 424w, https://cdn-images-1.medium.com/max/788/1*QBWxO_aDKvDWXxVjS-t8iQ.png 848w, https://cdn-images-1.medium.com/max/788/1*QBWxO_aDKvDWXxVjS-t8iQ.png 1272w, https://cdn-images-1.medium.com/max/788/1*QBWxO_aDKvDWXxVjS-t8iQ.png 1456w" sizes="100vw"><img src="https://cdn-images-1.medium.com/max/788/1*QBWxO_aDKvDWXxVjS-t8iQ.png" data-attrs="{&quot;src&quot;:&quot;https://cdn-images-1.medium.com/max/788/1*QBWxO_aDKvDWXxVjS-t8iQ.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://cdn-images-1.medium.com/max/788/1*QBWxO_aDKvDWXxVjS-t8iQ.png 424w, https://cdn-images-1.medium.com/max/788/1*QBWxO_aDKvDWXxVjS-t8iQ.png 848w, https://cdn-images-1.medium.com/max/788/1*QBWxO_aDKvDWXxVjS-t8iQ.png 1272w, https://cdn-images-1.medium.com/max/788/1*QBWxO_aDKvDWXxVjS-t8iQ.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">Visualization of different pruning methods. The method proposed in this work is Stripe-wise pruning which is intuitive and has computational advantages. Image reproduced from<a href="https://nips.cc/virtual/2020/public/poster_ccb1d45fb76f7c5a0bf619f979c6cf36.html"> the&nbsp;source</a>.</figcaption></figure></div><p><strong>Results and conclusion: </strong>Stripe-wise pruning outperforms an existing method of Group-wise pruning which is also a Structured Pruning method with more granularity. See fig.&nbsp;below.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://cdn-images-1.medium.com/max/1024/1*1vdL9HK24q_V7w_MNUR7zw.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://cdn-images-1.medium.com/max/1024/1*1vdL9HK24q_V7w_MNUR7zw.png 424w, https://cdn-images-1.medium.com/max/1024/1*1vdL9HK24q_V7w_MNUR7zw.png 848w, https://cdn-images-1.medium.com/max/1024/1*1vdL9HK24q_V7w_MNUR7zw.png 1272w, https://cdn-images-1.medium.com/max/1024/1*1vdL9HK24q_V7w_MNUR7zw.png 1456w" sizes="100vw"><img src="https://cdn-images-1.medium.com/max/1024/1*1vdL9HK24q_V7w_MNUR7zw.png" data-attrs="{&quot;src&quot;:&quot;https://cdn-images-1.medium.com/max/1024/1*1vdL9HK24q_V7w_MNUR7zw.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://cdn-images-1.medium.com/max/1024/1*1vdL9HK24q_V7w_MNUR7zw.png 424w, https://cdn-images-1.medium.com/max/1024/1*1vdL9HK24q_V7w_MNUR7zw.png 848w, https://cdn-images-1.medium.com/max/1024/1*1vdL9HK24q_V7w_MNUR7zw.png 1272w, https://cdn-images-1.medium.com/max/1024/1*1vdL9HK24q_V7w_MNUR7zw.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">Stripe-wise pruning outperforms Group-wise pruning. Image reproduced from <a href="https://nips.cc/virtual/2020/public/poster_ccb1d45fb76f7c5a0bf619f979c6cf36.html">the&nbsp;source</a>.</figcaption></figure></div><p>The method proposed reduces the required FLOPs of ResNet-18 on ImageNet dataset by 50.4% while increasing the accuracy by&nbsp;0.23%.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://cdn-images-1.medium.com/max/1024/1*kGTSZdzu7_NiAnalg4TtyA.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://cdn-images-1.medium.com/max/1024/1*kGTSZdzu7_NiAnalg4TtyA.png 424w, https://cdn-images-1.medium.com/max/1024/1*kGTSZdzu7_NiAnalg4TtyA.png 848w, https://cdn-images-1.medium.com/max/1024/1*kGTSZdzu7_NiAnalg4TtyA.png 1272w, https://cdn-images-1.medium.com/max/1024/1*kGTSZdzu7_NiAnalg4TtyA.png 1456w" sizes="100vw"><img src="https://cdn-images-1.medium.com/max/1024/1*kGTSZdzu7_NiAnalg4TtyA.png" data-attrs="{&quot;src&quot;:&quot;https://cdn-images-1.medium.com/max/1024/1*kGTSZdzu7_NiAnalg4TtyA.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://cdn-images-1.medium.com/max/1024/1*kGTSZdzu7_NiAnalg4TtyA.png 424w, https://cdn-images-1.medium.com/max/1024/1*kGTSZdzu7_NiAnalg4TtyA.png 848w, https://cdn-images-1.medium.com/max/1024/1*kGTSZdzu7_NiAnalg4TtyA.png 1272w, https://cdn-images-1.medium.com/max/1024/1*kGTSZdzu7_NiAnalg4TtyA.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">The FLOPs required for ResNet-18 on the ImageNet dataset is reduced by 50.4% while <strong>increasing</strong> the accuracy by 0.23%. Image reproduced from <a href="https://nips.cc/virtual/2020/public/poster_ccb1d45fb76f7c5a0bf619f979c6cf36.html">the&nbsp;source</a>.</figcaption></figure></div><h4><a href="https://nips.cc/virtual/2020/public/poster_46a4378f835dc8040c8057beb6a2da52.html">Pruning neural networks without any data by iteratively conserving synaptic flow</a> (NeurIPS&nbsp;2020)</h4><p><strong>Outstanding problem:</strong> Methods for Pruning at initialization without any training what-so-ever are not efficient.</p><p>This work identifies that the problem with current methods for Pruning at initialization is<strong> </strong>Layer-dropping, where a method prunes a particular layer completely, making the networks untrainable. This will occur when the sparsity we want to achieve is higher than the layer-drop threshold. In the figure below, layer-drop happens in all the methods except in the proposed method, Syn-Flow.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://cdn-images-1.medium.com/max/501/1*yPKNx8lxrKXNghmeotmSHw.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://cdn-images-1.medium.com/max/501/1*yPKNx8lxrKXNghmeotmSHw.png 424w, https://cdn-images-1.medium.com/max/501/1*yPKNx8lxrKXNghmeotmSHw.png 848w, https://cdn-images-1.medium.com/max/501/1*yPKNx8lxrKXNghmeotmSHw.png 1272w, https://cdn-images-1.medium.com/max/501/1*yPKNx8lxrKXNghmeotmSHw.png 1456w" sizes="100vw"><img src="https://cdn-images-1.medium.com/max/501/1*yPKNx8lxrKXNghmeotmSHw.png" data-attrs="{&quot;src&quot;:&quot;https://cdn-images-1.medium.com/max/501/1*yPKNx8lxrKXNghmeotmSHw.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://cdn-images-1.medium.com/max/501/1*yPKNx8lxrKXNghmeotmSHw.png 424w, https://cdn-images-1.medium.com/max/501/1*yPKNx8lxrKXNghmeotmSHw.png 848w, https://cdn-images-1.medium.com/max/501/1*yPKNx8lxrKXNghmeotmSHw.png 1272w, https://cdn-images-1.medium.com/max/501/1*yPKNx8lxrKXNghmeotmSHw.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">For all the previous methods, accuracy drops at a particular compression ratio because the method removes a layer entirely (Layer-drop). But the proposed method SynFLow achieves Max compression avoiding the Layer drop. Image reproduced from <a href="https://nips.cc/virtual/2020/public/poster_46a4378f835dc8040c8057beb6a2da52.html">the&nbsp;source</a>.</figcaption></figure></div><p><strong>Proposed solution:</strong> This work theoretically understands that Iterative Magnitude Pruning avoids Layer-drop with iterative training and pruning by achieving a property which they call Maximal Critical Compression. With these insights, they propose a method that<strong> </strong>achieves Maximal Critical Compression without any&nbsp;data.</p><p><strong>Results and conclusion:</strong> Syn-Flow consistently outperforms previous methods for Pruning at initialization in 12 combinations of datasets and models. See the figure&nbsp;below.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://cdn-images-1.medium.com/max/959/1*-O5rWaQoC0L-CV46IPRwEg.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://cdn-images-1.medium.com/max/959/1*-O5rWaQoC0L-CV46IPRwEg.png 424w, https://cdn-images-1.medium.com/max/959/1*-O5rWaQoC0L-CV46IPRwEg.png 848w, https://cdn-images-1.medium.com/max/959/1*-O5rWaQoC0L-CV46IPRwEg.png 1272w, https://cdn-images-1.medium.com/max/959/1*-O5rWaQoC0L-CV46IPRwEg.png 1456w" sizes="100vw"><img src="https://cdn-images-1.medium.com/max/959/1*-O5rWaQoC0L-CV46IPRwEg.png" data-attrs="{&quot;src&quot;:&quot;https://cdn-images-1.medium.com/max/959/1*-O5rWaQoC0L-CV46IPRwEg.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://cdn-images-1.medium.com/max/959/1*-O5rWaQoC0L-CV46IPRwEg.png 424w, https://cdn-images-1.medium.com/max/959/1*-O5rWaQoC0L-CV46IPRwEg.png 848w, https://cdn-images-1.medium.com/max/959/1*-O5rWaQoC0L-CV46IPRwEg.png 1272w, https://cdn-images-1.medium.com/max/959/1*-O5rWaQoC0L-CV46IPRwEg.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">SynFlow outperforms previous methods in 12 combinations of datasets and methods. Image reproduced from <a href="https://nips.cc/virtual/2020/public/poster_46a4378f835dc8040c8057beb6a2da52.html">the&nbsp;source</a>.</figcaption></figure></div><h4><a href="https://nips.cc/virtual/2020/public/poster_b6af2c9703f203a2794be03d443af2e3.html">The Lottery Ticket Hypothesis for Pre-trained BERT Networks</a> (NeurIPS&nbsp;2020)</h4><p><strong>Outstanding problem:</strong> Lottery ticket hypothesis says that there is a sub-network within the original, complete network that could be trained to achieve the same performance as the entire network. These sub-networks are called Lottery tickets. These lottery tickets are represented by binary masks that tell which of the weights should be zeroed&nbsp;out.</p><p><strong>Proposed solution:</strong> The authors used Iterative Magnitude Pruning (IMP) and tried to find lottery tickets from the pre-trained BERT&nbsp;model.</p><p>IMP trains a network, removes weights with lower magnitude, and then repeats this process until the required sparsity is achieved. It uses Rewinding where the weights of the pruned network are reverted back to their initialization weights, here pre-trained BERT&nbsp;weights.</p><p>The lottery ticket here is the subnetwork of the original pre-trained BERT model which shows similar performance. Note that the weights of the lottery ticket and the pre-trained BERT model are the same where the lottery ticket is not pruned because of the Rewinding in&nbsp;IMP.</p><p><strong>Results and conclusion: </strong>This work shows that the lottery ticket hypothesis holds for pre-trained BERT models as well. It also found subnetworks at 40% to 90% sparsity for a range of downstream tasks.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://cdn-images-1.medium.com/max/1024/0*z50lKatd3WuSV-AR.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://cdn-images-1.medium.com/max/1024/0*z50lKatd3WuSV-AR.png 424w, https://cdn-images-1.medium.com/max/1024/0*z50lKatd3WuSV-AR.png 848w, https://cdn-images-1.medium.com/max/1024/0*z50lKatd3WuSV-AR.png 1272w, https://cdn-images-1.medium.com/max/1024/0*z50lKatd3WuSV-AR.png 1456w" sizes="100vw"><img src="https://cdn-images-1.medium.com/max/1024/0*z50lKatd3WuSV-AR.png" data-attrs="{&quot;src&quot;:&quot;https://cdn-images-1.medium.com/max/1024/0*z50lKatd3WuSV-AR.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://cdn-images-1.medium.com/max/1024/0*z50lKatd3WuSV-AR.png 424w, https://cdn-images-1.medium.com/max/1024/0*z50lKatd3WuSV-AR.png 848w, https://cdn-images-1.medium.com/max/1024/0*z50lKatd3WuSV-AR.png 1272w, https://cdn-images-1.medium.com/max/1024/0*z50lKatd3WuSV-AR.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">The last row corresponds to the approach introduced in this paper. Even though it is 40%-90% sparse, performance is comparable to a Full BERT base. Image reproduced from <a href="https://nips.cc/virtual/2020/public/poster_b6af2c9703f203a2794be03d443af2e3.html">the&nbsp;source</a>.</figcaption></figure></div><p>In a work similar to this one, it is shown that lottery tickets exist for NLP and RL as well. See the paper <a href="https://iclr.cc/virtual_2020/poster_S1xnXRVFwH.html">Playing the lottery with rewards and multiple languages: lottery tickets in RL and&nbsp;NLP</a>.</p><h4><a href="https://nips.cc/virtual/2020/public/poster_e3a72c791a69f87b05ea7742e04430ed.html">HYDRA: Pruning Adversarially Robust Neural Networks</a> (NeurIPS&nbsp;2020)</h4><p><strong>Outstanding problem:</strong> In resource-constrained and safety-critical environments, both robustness and compactness are required at the same time. It is shown that to increase the robustness of a Neural Network, we have to increase the size of the network. So, robustness with fewer parameters is very desirable.</p><p>This work conjectures and shows evidence that even low-magnitude weights are important for robustness. As most of the dedicated pruning techniques remove low-magnitude weights, this could hinder robustness accuracy. This shows that dealing with robustness and pruning at the same time leads to more efficient methods, of which the current method is a great&nbsp;example.</p><p><strong>Proposed solution: </strong>This work proposed a pruning method, HYDRA, which decides which weights to remove while decreasing the desired loss, here the loss derives from the robustness training objective.</p><p>Finding an optimal pruning mask is formulated as importance-score-based optimization where score values for different weights could have a floating-point importance score between 0 and 1. After the optimization, weights with lower importance scores are&nbsp;removed.</p><p><strong>Results and conclusion:</strong> The proposed pruning technique prunes the network while showing less degradation in robustness accuracy compared to the dedicated pruning technique LVM.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://cdn-images-1.medium.com/max/793/1*d6p3Z8tt97B6mBBeZ99PXQ.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://cdn-images-1.medium.com/max/793/1*d6p3Z8tt97B6mBBeZ99PXQ.png 424w, https://cdn-images-1.medium.com/max/793/1*d6p3Z8tt97B6mBBeZ99PXQ.png 848w, https://cdn-images-1.medium.com/max/793/1*d6p3Z8tt97B6mBBeZ99PXQ.png 1272w, https://cdn-images-1.medium.com/max/793/1*d6p3Z8tt97B6mBBeZ99PXQ.png 1456w" sizes="100vw"><img src="https://cdn-images-1.medium.com/max/793/1*d6p3Z8tt97B6mBBeZ99PXQ.png" data-attrs="{&quot;src&quot;:&quot;https://cdn-images-1.medium.com/max/793/1*d6p3Z8tt97B6mBBeZ99PXQ.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://cdn-images-1.medium.com/max/793/1*d6p3Z8tt97B6mBBeZ99PXQ.png 424w, https://cdn-images-1.medium.com/max/793/1*d6p3Z8tt97B6mBBeZ99PXQ.png 848w, https://cdn-images-1.medium.com/max/793/1*d6p3Z8tt97B6mBBeZ99PXQ.png 1272w, https://cdn-images-1.medium.com/max/793/1*d6p3Z8tt97B6mBBeZ99PXQ.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">For the method proposed robust accuracy doesn&#8217;t drastically decrease with pruning which is the case with the previous method LWM. Image reproduced from<a href="https://nips.cc/virtual/2020/public/poster_e3a72c791a69f87b05ea7742e04430ed.html"> the&nbsp;source</a>.</figcaption></figure></div><p>The authors also observe that the proposed method retains low-magnitude weights while retaining better robustness accuracy which shows the importance of low-magnitude weights towards robustness.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://cdn-images-1.medium.com/max/1024/1*hhpLTcONRgMk3RpRgVPIzg.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://cdn-images-1.medium.com/max/1024/1*hhpLTcONRgMk3RpRgVPIzg.png 424w, https://cdn-images-1.medium.com/max/1024/1*hhpLTcONRgMk3RpRgVPIzg.png 848w, https://cdn-images-1.medium.com/max/1024/1*hhpLTcONRgMk3RpRgVPIzg.png 1272w, https://cdn-images-1.medium.com/max/1024/1*hhpLTcONRgMk3RpRgVPIzg.png 1456w" sizes="100vw"><img src="https://cdn-images-1.medium.com/max/1024/1*hhpLTcONRgMk3RpRgVPIzg.png" data-attrs="{&quot;src&quot;:&quot;https://cdn-images-1.medium.com/max/1024/1*hhpLTcONRgMk3RpRgVPIzg.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://cdn-images-1.medium.com/max/1024/1*hhpLTcONRgMk3RpRgVPIzg.png 424w, https://cdn-images-1.medium.com/max/1024/1*hhpLTcONRgMk3RpRgVPIzg.png 848w, https://cdn-images-1.medium.com/max/1024/1*hhpLTcONRgMk3RpRgVPIzg.png 1272w, https://cdn-images-1.medium.com/max/1024/1*hhpLTcONRgMk3RpRgVPIzg.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">The proposed method retains low-magnitude weights (shown as black) and shows good robust accuracy while the existing pruning methods prune low-magnitude weights whether they are useful for robustness or not. Image reproduced from <a href="https://nips.cc/virtual/2020/public/poster_e3a72c791a69f87b05ea7742e04430ed.html">the&nbsp;source</a>.</figcaption></figure></div><p>This method shows state-of-the-art performance with ResNet50 on ImageNet with adversarial training.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://cdn-images-1.medium.com/max/930/1*PUgQQXrEWHNCk-y6V2o0Kw.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://cdn-images-1.medium.com/max/930/1*PUgQQXrEWHNCk-y6V2o0Kw.png 424w, https://cdn-images-1.medium.com/max/930/1*PUgQQXrEWHNCk-y6V2o0Kw.png 848w, https://cdn-images-1.medium.com/max/930/1*PUgQQXrEWHNCk-y6V2o0Kw.png 1272w, https://cdn-images-1.medium.com/max/930/1*PUgQQXrEWHNCk-y6V2o0Kw.png 1456w" sizes="100vw"><img src="https://cdn-images-1.medium.com/max/930/1*PUgQQXrEWHNCk-y6V2o0Kw.png" data-attrs="{&quot;src&quot;:&quot;https://cdn-images-1.medium.com/max/930/1*PUgQQXrEWHNCk-y6V2o0Kw.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://cdn-images-1.medium.com/max/930/1*PUgQQXrEWHNCk-y6V2o0Kw.png 424w, https://cdn-images-1.medium.com/max/930/1*PUgQQXrEWHNCk-y6V2o0Kw.png 848w, https://cdn-images-1.medium.com/max/930/1*PUgQQXrEWHNCk-y6V2o0Kw.png 1272w, https://cdn-images-1.medium.com/max/930/1*PUgQQXrEWHNCk-y6V2o0Kw.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">HYDRA, the proposed method, outperforms the previous best method Adv-LWM. Shown as {accuracy/robust accuracy}. Image reproduced from<a href="https://nips.cc/virtual/2020/public/poster_e3a72c791a69f87b05ea7742e04430ed.html"> the&nbsp;source</a>.</figcaption></figure></div><h4>Conclusion and Future&nbsp;work</h4><p>Neural Network Pruning research is evolving to be more scientific and rigorous. One of the reasons is, undoubtedly, the interaction between the wide adoption of deep learning methods in Computer Vision &amp; NLP and elsewhere and the increasing amount of memory, energy, and computational resources required for the state-of-the-art methods.</p><p>Going into 2021, we know how to prune a network efficiently without any data, how to strike a balance between Structured and Unstructured Pruning methods while retaining the benefits of both worlds, how to find lottery tickets (pruned networks) in diverse domains and how to best retrain them, and how to prune a model even when adversarial robustness is&nbsp;needed.</p><p>And as with any field which is scientific and rigorous, Pruning research is poised to become more objective/authoritative (as opposed to having very little comparative analysis), with validation of new techniques over diverse datasets, architectures, and domains not just on, as seen in many cases, ImageNet dataset with a ResNet variant. This will inform us plenty about how to work with deep learning methods across different tasks even in low-resource environments.</p><p><em>Editor&#8217;s Note: <a href="https://heartbeat.comet.ml/">Heartbeat</a> is a contributor-driven online publication and community dedicated to providing premier educational resources for data science, machine learning, and deep learning practitioners. We&#8217;re committed to supporting and inspiring developers and engineers from all walks of&nbsp;life.</em></p><p><em>Editorially independent, Heartbeat is sponsored and published by <a href="http://comet.ml/?utm_campaign=heartbeat-statement&amp;utm_source=blog&amp;utm_medium=medium">Comet</a>, an MLOps platform that enables data scientists &amp; ML teams to track, compare, explain, &amp; optimize their experiments. We pay our contributors, and we don&#8217;t sell&nbsp;ads.</em></p><p><em>If you&#8217;d like to contribute, head on over to our<a href="https://heartbeat.fritz.ai/call-for-contributors-october-2018-update-fee7f5b80f3e"> call for contributors</a>. You can also sign up to receive our weekly newsletters (<a href="https://www.deeplearningweekly.com/">Deep Learning Weekly</a> and the <a href="https://info.comet.ml/newsletter-signup/">Comet Newsletter</a>), join us on<a href="https://join.slack.com/t/fritz-ai-community/shared_invite/enQtNTY5NDM2MTQwMTgwLWU4ZDEwNTAxYWE2YjIxZDllMTcxMWE4MGFhNDk5Y2QwNTcxYzEyNWZmZWEwMzE4NTFkOWY2NTM0OGQwYjM5Y2U"> </a><a href="https://join.slack.com/t/cometml/shared_invite/zt-49v4zxxz-qHcTeyrMEzqZc5lQb9hgvw">Slack</a>, and follow Comet on <a href="https://twitter.com/Cometml">Twitter</a> and <a href="https://www.linkedin.com/company/comet-ml/">LinkedIn</a> for resources, events, and much more that will help you build better ML models,&nbsp;faster.</em></p><div><hr></div><p><a href="https://heartbeat.comet.ml/neural-network-pruning-research-review-2020-bc21a77f0295">Neural Network Pruning Research Review 2020</a> was originally published in <a href="https://heartbeat.comet.ml">Heartbeat</a> on Medium, where people are continuing the conversation by highlighting and responding to this story.</p>]]></content:encoded></item><item><title><![CDATA[NeurIPS 2020 Papers: Takeaways of a Deep Learning Engineer (Part 2 of 3)— Computer Vision]]></title><description><![CDATA[NeurIPS 2020 Papers: Takeaways of a Deep Learning Engineer&#8212; Computer Vision]]></description><link>https://www.aipapertrails.com/p/neurips-2020-papers-takeaways-of-a-deep-learning-engineer-part-2-of-3-computer-vision-ef5ea1abe525</link><guid isPermaLink="false">https://www.aipapertrails.com/p/neurips-2020-papers-takeaways-of-a-deep-learning-engineer-part-2-of-3-computer-vision-ef5ea1abe525</guid><dc:creator><![CDATA[Prakash Kagitha]]></dc:creator><pubDate>Sat, 05 Dec 2020 15:59:41 GMT</pubDate><enclosure url="https://cdn-images-1.medium.com/max/1024/1*KVWKMsVviBNpPQQp1RhCow.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<h3>NeurIPS 2020 Papers: Takeaways of a Deep Learning Engineer&#8212;&#8202;Computer&nbsp;Vision</h3><h4>Techniques and insights for applied deep learning (computer vision) from papers published at NeurIPS&nbsp;2020</h4><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://cdn-images-1.medium.com/max/1024/1*KVWKMsVviBNpPQQp1RhCow.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://cdn-images-1.medium.com/max/1024/1*KVWKMsVviBNpPQQp1RhCow.png 424w, https://cdn-images-1.medium.com/max/1024/1*KVWKMsVviBNpPQQp1RhCow.png 848w, https://cdn-images-1.medium.com/max/1024/1*KVWKMsVviBNpPQQp1RhCow.png 1272w, https://cdn-images-1.medium.com/max/1024/1*KVWKMsVviBNpPQQp1RhCow.png 1456w" sizes="100vw"><img src="https://cdn-images-1.medium.com/max/1024/1*KVWKMsVviBNpPQQp1RhCow.png" data-attrs="{&quot;src&quot;:&quot;https://cdn-images-1.medium.com/max/1024/1*KVWKMsVviBNpPQQp1RhCow.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://cdn-images-1.medium.com/max/1024/1*KVWKMsVviBNpPQQp1RhCow.png 424w, https://cdn-images-1.medium.com/max/1024/1*KVWKMsVviBNpPQQp1RhCow.png 848w, https://cdn-images-1.medium.com/max/1024/1*KVWKMsVviBNpPQQp1RhCow.png 1272w, https://cdn-images-1.medium.com/max/1024/1*KVWKMsVviBNpPQQp1RhCow.png 1456w" sizes="100vw" fetchpriority="high"></picture><div></div></div></a><figcaption class="image-caption">Image by&nbsp;Author</figcaption></figure></div><p>As mentioned in part 1&#8212; the most important thing:)&#8202;&#8212;&#8202;I went through all the titles of NeurIPS 2020 papers (more than 1900!) and read abstracts of 175 papers, and <strong>extracted DL engineer relevant insights from the following papers</strong>.</p><p>This is part 2. See the part 1&nbsp;below.</p><p><a href="https://towardsdatascience.com/neurips-2020-papers-a-deep-learning-engineers-takeaway-4f3066523151">NeurIPS 2020 Papers: A Deep Learning Engineer&#8217;s Takeaway</a></p><h4><a href="https://neurips.cc/virtual/2020/public/poster_27e9661e033a73a6ad8cefcde965c54d.html">Rethinking Pre-training and Self-training</a></h4><p>Using other datasets to better solve the target dataset is ubiquitous in deep learning practice. It could be supervised pre-training (Classification; <strong>ImageNet pre-trained</strong>) or self-supervised pre-training (SimCLR on unlabeled data) or <strong>self-training</strong>.</p><p>(Self-training is a process where an intermediate model (teacher model), which is trained on target dataset, is used to create &#8216;labels&#8217; (thus called pseudo labels) for another dataset and then the final model (student model) is trained with both target dataset and the pseudo labeled dataset.)</p><p>Building on the previous work, the current work shows that the usefulness of ImageNet pre-training (starting with pre-trained weights rather than random) or self-supervised pre-training <strong>decreases with the size of the target dataset and the strength of the data augmentation</strong>. ImageNet pretraining didn&#8217;t help, rather hurt in some cases, the model when training on COCO dataset for object detection.</p><p>But, self-training helped in both low-data and high-data regime and with both strong and weak data augmentation strategies.<strong> It helped when pre-training didn&#8217;t help and showed improvement on it when it&nbsp;did</strong>.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://cdn-images-1.medium.com/max/1024/1*nNTVprQe981LvX9hOsCMww.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://cdn-images-1.medium.com/max/1024/1*nNTVprQe981LvX9hOsCMww.png 424w, https://cdn-images-1.medium.com/max/1024/1*nNTVprQe981LvX9hOsCMww.png 848w, https://cdn-images-1.medium.com/max/1024/1*nNTVprQe981LvX9hOsCMww.png 1272w, https://cdn-images-1.medium.com/max/1024/1*nNTVprQe981LvX9hOsCMww.png 1456w" sizes="100vw"><img src="https://cdn-images-1.medium.com/max/1024/1*nNTVprQe981LvX9hOsCMww.png" data-attrs="{&quot;src&quot;:&quot;https://cdn-images-1.medium.com/max/1024/1*nNTVprQe981LvX9hOsCMww.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://cdn-images-1.medium.com/max/1024/1*nNTVprQe981LvX9hOsCMww.png 424w, https://cdn-images-1.medium.com/max/1024/1*nNTVprQe981LvX9hOsCMww.png 848w, https://cdn-images-1.medium.com/max/1024/1*nNTVprQe981LvX9hOsCMww.png 1272w, https://cdn-images-1.medium.com/max/1024/1*nNTVprQe981LvX9hOsCMww.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">ImageNet pre-training vs Self-training when the strength of the data augmentation is changed. Image from the pdf of the <a href="https://neurips.cc/virtual/2020/public/poster_27e9661e033a73a6ad8cefcde965c54d.html">current&nbsp;paper</a>.</figcaption></figure></div><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://cdn-images-1.medium.com/max/965/1*hP7wTS2tKkJr5yGgqoo2Gg.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://cdn-images-1.medium.com/max/965/1*hP7wTS2tKkJr5yGgqoo2Gg.png 424w, https://cdn-images-1.medium.com/max/965/1*hP7wTS2tKkJr5yGgqoo2Gg.png 848w, https://cdn-images-1.medium.com/max/965/1*hP7wTS2tKkJr5yGgqoo2Gg.png 1272w, https://cdn-images-1.medium.com/max/965/1*hP7wTS2tKkJr5yGgqoo2Gg.png 1456w" sizes="100vw"><img src="https://cdn-images-1.medium.com/max/965/1*hP7wTS2tKkJr5yGgqoo2Gg.png" data-attrs="{&quot;src&quot;:&quot;https://cdn-images-1.medium.com/max/965/1*hP7wTS2tKkJr5yGgqoo2Gg.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://cdn-images-1.medium.com/max/965/1*hP7wTS2tKkJr5yGgqoo2Gg.png 424w, https://cdn-images-1.medium.com/max/965/1*hP7wTS2tKkJr5yGgqoo2Gg.png 848w, https://cdn-images-1.medium.com/max/965/1*hP7wTS2tKkJr5yGgqoo2Gg.png 1272w, https://cdn-images-1.medium.com/max/965/1*hP7wTS2tKkJr5yGgqoo2Gg.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">Descriptions for Labels in the images above and below. Image from the pdf of the <a href="https://neurips.cc/virtual/2020/public/poster_27e9661e033a73a6ad8cefcde965c54d.html">current&nbsp;paper</a>.</figcaption></figure></div><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://cdn-images-1.medium.com/max/1024/1*aprjRi-a3xAH0KmJm2R4LQ.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://cdn-images-1.medium.com/max/1024/1*aprjRi-a3xAH0KmJm2R4LQ.png 424w, https://cdn-images-1.medium.com/max/1024/1*aprjRi-a3xAH0KmJm2R4LQ.png 848w, https://cdn-images-1.medium.com/max/1024/1*aprjRi-a3xAH0KmJm2R4LQ.png 1272w, https://cdn-images-1.medium.com/max/1024/1*aprjRi-a3xAH0KmJm2R4LQ.png 1456w" sizes="100vw"><img src="https://cdn-images-1.medium.com/max/1024/1*aprjRi-a3xAH0KmJm2R4LQ.png" data-attrs="{&quot;src&quot;:&quot;https://cdn-images-1.medium.com/max/1024/1*aprjRi-a3xAH0KmJm2R4LQ.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://cdn-images-1.medium.com/max/1024/1*aprjRi-a3xAH0KmJm2R4LQ.png 424w, https://cdn-images-1.medium.com/max/1024/1*aprjRi-a3xAH0KmJm2R4LQ.png 848w, https://cdn-images-1.medium.com/max/1024/1*aprjRi-a3xAH0KmJm2R4LQ.png 1272w, https://cdn-images-1.medium.com/max/1024/1*aprjRi-a3xAH0KmJm2R4LQ.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">Random Init vs ImageNet pre-training vs Self-training. Self-training helps when pre-training doesn&#8217;t help and improves on it when it does help. Image from the pdf of the <a href="https://neurips.cc/virtual/2020/public/poster_27e9661e033a73a6ad8cefcde965c54d.html">current&nbsp;paper</a>.</figcaption></figure></div><blockquote><p><strong>Takeaway:</strong> When you want to leverage other datasets in training a model on a target dataset, use self-training rather than ImageNet pretraining. But keep in mind that self-training takes more resources than just initializing your model with ImageNet pre-trained weights.</p></blockquote><h4><a href="https://neurips.cc/virtual/2020/public/poster_9d684c589d67031a627ad33d59db65e5.html">RelationNet++: Bridging Visual Representations for Object Detection via Transformer Decoder</a></h4><p>Different object detection models employ<strong> different intermediate representations</strong> from which the bounding box predictions are&nbsp;made.</p><p>For example, RetinaNet uses a <strong>bounding box (anchors)</strong> representational format, where it creates feature maps for each bounding box instance created by anchor boxes at each position of the feature grid. If a feature grid is of H x W, takes RetinaNet takes 9 anchor boxes (<strong>pre-specified aspect ratios</strong>) for each position of the feature grid giving us 9 x H x W bounding box instances to do IOU thresholding, predicting the classes and sub-pixel offsets, and do NMS on top among other things to get the final set of bounding boxes for an&nbsp;image.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://cdn-images-1.medium.com/max/1024/1*V-0cF8HTC-2jw-vBXS6WeA.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://cdn-images-1.medium.com/max/1024/1*V-0cF8HTC-2jw-vBXS6WeA.png 424w, https://cdn-images-1.medium.com/max/1024/1*V-0cF8HTC-2jw-vBXS6WeA.png 848w, https://cdn-images-1.medium.com/max/1024/1*V-0cF8HTC-2jw-vBXS6WeA.png 1272w, https://cdn-images-1.medium.com/max/1024/1*V-0cF8HTC-2jw-vBXS6WeA.png 1456w" sizes="100vw"><img src="https://cdn-images-1.medium.com/max/1024/1*V-0cF8HTC-2jw-vBXS6WeA.png" data-attrs="{&quot;src&quot;:&quot;https://cdn-images-1.medium.com/max/1024/1*V-0cF8HTC-2jw-vBXS6WeA.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://cdn-images-1.medium.com/max/1024/1*V-0cF8HTC-2jw-vBXS6WeA.png 424w, https://cdn-images-1.medium.com/max/1024/1*V-0cF8HTC-2jw-vBXS6WeA.png 848w, https://cdn-images-1.medium.com/max/1024/1*V-0cF8HTC-2jw-vBXS6WeA.png 1272w, https://cdn-images-1.medium.com/max/1024/1*V-0cF8HTC-2jw-vBXS6WeA.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">Different intermediate representations for different models. Image from the pdf of the<a href="https://neurips.cc/virtual/2020/public/poster_9d684c589d67031a627ad33d59db65e5.html"> current&nbsp;paper</a>.</figcaption></figure></div><p>FCOS and CenterNet use<strong> a center point</strong> as representation formats and estimates bounding boxes by predicting x and y dimensional offsets from the center point. And it has all the other processing steps very similar in objective with RetinaNet or any other object detection models.</p><p>CornerNet instead uses <strong>corner points</strong> as representation format (top left and bottom right) and creates a bounding box with those corner&nbsp;points.</p><p>Different representations are prevalent in object detection because each representation is good at some specific thing compared to all others. Bounding box representation is better aligned with annotation formats of datasets and is <strong>better at classification</strong>. Center point representation is <strong>better for detecting small objects</strong>. Corner point representation is <strong>better at localization</strong>.</p><p>This current work aims to combine the strengths of all these different representations. For a particular object detection model, <strong>they improve the features of its primary representation</strong>, bounding box for RetinaNet, by also taking into account features from other <strong>auxiliary representations</strong>, here, they are center points and corner&nbsp;points.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://cdn-images-1.medium.com/max/951/1*L7Cccu2xeWboFzNJEikG_Q.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://cdn-images-1.medium.com/max/951/1*L7Cccu2xeWboFzNJEikG_Q.png 424w, https://cdn-images-1.medium.com/max/951/1*L7Cccu2xeWboFzNJEikG_Q.png 848w, https://cdn-images-1.medium.com/max/951/1*L7Cccu2xeWboFzNJEikG_Q.png 1272w, https://cdn-images-1.medium.com/max/951/1*L7Cccu2xeWboFzNJEikG_Q.png 1456w" sizes="100vw"><img src="https://cdn-images-1.medium.com/max/951/1*L7Cccu2xeWboFzNJEikG_Q.png" data-attrs="{&quot;src&quot;:&quot;https://cdn-images-1.medium.com/max/951/1*L7Cccu2xeWboFzNJEikG_Q.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://cdn-images-1.medium.com/max/951/1*L7Cccu2xeWboFzNJEikG_Q.png 424w, https://cdn-images-1.medium.com/max/951/1*L7Cccu2xeWboFzNJEikG_Q.png 848w, https://cdn-images-1.medium.com/max/951/1*L7Cccu2xeWboFzNJEikG_Q.png 1272w, https://cdn-images-1.medium.com/max/951/1*L7Cccu2xeWboFzNJEikG_Q.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">Illustration of working of BVR combining bounding box representations, center representations, and corner points representations. Image from the pdf of the <a href="https://neurips.cc/virtual/2020/public/poster_9d684c589d67031a627ad33d59db65e5.html">current&nbsp;paper</a>.</figcaption></figure></div><p>The author proposed a <strong>Transformer</strong> model. When given a feature vector of primary representation for a location on a feature grid (query) it calculates <strong>attention weights with feature vectors of auxiliary representations</strong> at relevant locations and returns a weighted average of these auxiliary representations.</p><p>The model, called Bridging Visual Representations (BVR),<strong> will use both</strong> the feature vector for primary representation and the weighted average of feature vectors from auxiliary representations to do classification and localization thus combining the strengths and expressive power of different representational choices.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://cdn-images-1.medium.com/max/1024/1*jM5VjHA_z5ekeKQvzf0_LA.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://cdn-images-1.medium.com/max/1024/1*jM5VjHA_z5ekeKQvzf0_LA.png 424w, https://cdn-images-1.medium.com/max/1024/1*jM5VjHA_z5ekeKQvzf0_LA.png 848w, https://cdn-images-1.medium.com/max/1024/1*jM5VjHA_z5ekeKQvzf0_LA.png 1272w, https://cdn-images-1.medium.com/max/1024/1*jM5VjHA_z5ekeKQvzf0_LA.png 1456w" sizes="100vw"><img src="https://cdn-images-1.medium.com/max/1024/1*jM5VjHA_z5ekeKQvzf0_LA.png" data-attrs="{&quot;src&quot;:&quot;https://cdn-images-1.medium.com/max/1024/1*jM5VjHA_z5ekeKQvzf0_LA.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://cdn-images-1.medium.com/max/1024/1*jM5VjHA_z5ekeKQvzf0_LA.png 424w, https://cdn-images-1.medium.com/max/1024/1*jM5VjHA_z5ekeKQvzf0_LA.png 848w, https://cdn-images-1.medium.com/max/1024/1*jM5VjHA_z5ekeKQvzf0_LA.png 1272w, https://cdn-images-1.medium.com/max/1024/1*jM5VjHA_z5ekeKQvzf0_LA.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">RelationNet++ outperforms every other method. Image from the pdf of the <a href="https://neurips.cc/virtual/2020/public/poster_9d684c589d67031a627ad33d59db65e5.html">current&nbsp;paper</a>.</figcaption></figure></div><blockquote><p><strong>Takeaway:</strong> This is the state-of-the-art model and it makes sense. Any approach which combines the strengths of multiple solutions non-trivially would be valuable for a long time. Use this method when you train your next object detection model. (Too many good things for object detection!)</p></blockquote><h4><a href="https://neurips.cc/virtual/2020/public/poster_98dce83da57b0395e163467c9dae521b.html">Quantifying Learnability and Describability of Visual Concepts Emerging in Representation Learning</a></h4><p>Without a downstream task, it is hard to quantitatively evaluate image representations, i.e. the clusters formed with image representations for their semantic coherence and natural language describability.</p><p>This work formulates these tasks, learnability and describability of the clusters, as a forced-prediction problem and evaluates humans as predictors avoiding the issue of subjectivity which is a major problem with existing approaches. (Even though clusters are coherent, sometimes they can&#8217;t be described and even though they are describable different person might use different words and phrases).</p><p>After seeing a few samples of a cluster, a human should able to discriminate images of that cluster among images of other clusters. This means that clusters are separated in a human-interpretable way. The extent to which a human can do this is the <strong>metric for learnability</strong>.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://cdn-images-1.medium.com/max/1024/1*Hj0eJm7jWa_eiAB2Uvnmgg.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://cdn-images-1.medium.com/max/1024/1*Hj0eJm7jWa_eiAB2Uvnmgg.png 424w, https://cdn-images-1.medium.com/max/1024/1*Hj0eJm7jWa_eiAB2Uvnmgg.png 848w, https://cdn-images-1.medium.com/max/1024/1*Hj0eJm7jWa_eiAB2Uvnmgg.png 1272w, https://cdn-images-1.medium.com/max/1024/1*Hj0eJm7jWa_eiAB2Uvnmgg.png 1456w" sizes="100vw"><img src="https://cdn-images-1.medium.com/max/1024/1*Hj0eJm7jWa_eiAB2Uvnmgg.png" data-attrs="{&quot;src&quot;:&quot;https://cdn-images-1.medium.com/max/1024/1*Hj0eJm7jWa_eiAB2Uvnmgg.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://cdn-images-1.medium.com/max/1024/1*Hj0eJm7jWa_eiAB2Uvnmgg.png 424w, https://cdn-images-1.medium.com/max/1024/1*Hj0eJm7jWa_eiAB2Uvnmgg.png 848w, https://cdn-images-1.medium.com/max/1024/1*Hj0eJm7jWa_eiAB2Uvnmgg.png 1272w, https://cdn-images-1.medium.com/max/1024/1*Hj0eJm7jWa_eiAB2Uvnmgg.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">Image from the pdf of the <a href="https://neurips.cc/virtual/2020/public/poster_98dce83da57b0395e163467c9dae521b.html">current&nbsp;paper</a>.</figcaption></figure></div><p>After seeing the description of a cluster, a human should able to discriminate images of that cluster among images of other clusters. This means the given cluster is describable. (Description is sampled randomly from a manually populated set of descriptions for that cluster). The extent to which a human can do this is the <strong>metric for describability</strong>.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://cdn-images-1.medium.com/max/1024/1*fTexd5_Xw0W3lSWpR_iT9w.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://cdn-images-1.medium.com/max/1024/1*fTexd5_Xw0W3lSWpR_iT9w.png 424w, https://cdn-images-1.medium.com/max/1024/1*fTexd5_Xw0W3lSWpR_iT9w.png 848w, https://cdn-images-1.medium.com/max/1024/1*fTexd5_Xw0W3lSWpR_iT9w.png 1272w, https://cdn-images-1.medium.com/max/1024/1*fTexd5_Xw0W3lSWpR_iT9w.png 1456w" sizes="100vw"><img src="https://cdn-images-1.medium.com/max/1024/1*fTexd5_Xw0W3lSWpR_iT9w.png" data-attrs="{&quot;src&quot;:&quot;https://cdn-images-1.medium.com/max/1024/1*fTexd5_Xw0W3lSWpR_iT9w.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://cdn-images-1.medium.com/max/1024/1*fTexd5_Xw0W3lSWpR_iT9w.png 424w, https://cdn-images-1.medium.com/max/1024/1*fTexd5_Xw0W3lSWpR_iT9w.png 848w, https://cdn-images-1.medium.com/max/1024/1*fTexd5_Xw0W3lSWpR_iT9w.png 1272w, https://cdn-images-1.medium.com/max/1024/1*fTexd5_Xw0W3lSWpR_iT9w.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">Clusters and their descriptions from self-supervised model SeLa. Image from the pdf of the <a href="https://neurips.cc/virtual/2020/public/poster_98dce83da57b0395e163467c9dae521b.html">current&nbsp;paper</a>.</figcaption></figure></div><p>Authors also created a model to get automated descriptions for a cluster so that it could replace the human in the above describability metric.</p><blockquote><p><strong>Takeaway:</strong> If you have clusters of images with no labels, the extent to which you could discriminate other images as the same class or not, after seeing the images of a particular cluster, is a good metric to see whether your clusters are separated. The same goes for describability.</p></blockquote><h4><a href="https://neurips.cc/virtual/2020/public/poster_b2eeb7362ef83deff5c7813a67e14f0a.html">A Ranking-based, Balanced Loss Function Unifying Classification and Localisation in Object Detection</a></h4><p>There are a lot of outstanding problems to deal with in object detection. Prominent among them dealt with this work&nbsp;are:</p><ul><li><p><strong>Class imbalance problem</strong> between foreground/background (positive/negative) bounding boxes. Focal loss in RetinaNet helps but not&nbsp;enough.</p></li><li><p>Difficulty in <strong>tuning hyperparameters in the loss function </strong>(faster-RCNN has 9 of them to&nbsp;tune).</p></li><li><p>Discrepancy created by having <strong>separate localization and classification heads</strong>, for eg. classification loss is not dependent on the IOU or localization of that&nbsp;object.</p></li></ul><p>And this is how they deal with&nbsp;it:</p><ul><li><p>The ranking based loss function for classification <strong>is more stable and learns without overfitting</strong> when compared to Cross-Entropy, or weighted cross entropy variant like a focal&nbsp;loss.</p></li></ul><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://cdn-images-1.medium.com/max/961/1*IXHuyvDXU1N0m_yoLoHUhg.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://cdn-images-1.medium.com/max/961/1*IXHuyvDXU1N0m_yoLoHUhg.png 424w, https://cdn-images-1.medium.com/max/961/1*IXHuyvDXU1N0m_yoLoHUhg.png 848w, https://cdn-images-1.medium.com/max/961/1*IXHuyvDXU1N0m_yoLoHUhg.png 1272w, https://cdn-images-1.medium.com/max/961/1*IXHuyvDXU1N0m_yoLoHUhg.png 1456w" sizes="100vw"><img src="https://cdn-images-1.medium.com/max/961/1*IXHuyvDXU1N0m_yoLoHUhg.png" data-attrs="{&quot;src&quot;:&quot;https://cdn-images-1.medium.com/max/961/1*IXHuyvDXU1N0m_yoLoHUhg.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://cdn-images-1.medium.com/max/961/1*IXHuyvDXU1N0m_yoLoHUhg.png 424w, https://cdn-images-1.medium.com/max/961/1*IXHuyvDXU1N0m_yoLoHUhg.png 848w, https://cdn-images-1.medium.com/max/961/1*IXHuyvDXU1N0m_yoLoHUhg.png 1272w, https://cdn-images-1.medium.com/max/961/1*IXHuyvDXU1N0m_yoLoHUhg.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">Comparison of the number of hyperparameters in each loss function. Image from the pdf of the <a href="https://neurips.cc/virtual/2020/public/poster_b2eeb7362ef83deff5c7813a67e14f0a.html">current&nbsp;paper</a>.</figcaption></figure></div><ul><li><p>The proposed loss function has <strong>one hyperparameter</strong> and even that parameter <strong>doesn&#8217;t need tuning</strong>. (Results in the paper are without any tuning, it still outperformed baselines).</p></li></ul><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://cdn-images-1.medium.com/max/1024/1*LA2mreJG79FgM10-8sT85w.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://cdn-images-1.medium.com/max/1024/1*LA2mreJG79FgM10-8sT85w.png 424w, https://cdn-images-1.medium.com/max/1024/1*LA2mreJG79FgM10-8sT85w.png 848w, https://cdn-images-1.medium.com/max/1024/1*LA2mreJG79FgM10-8sT85w.png 1272w, https://cdn-images-1.medium.com/max/1024/1*LA2mreJG79FgM10-8sT85w.png 1456w" sizes="100vw"><img src="https://cdn-images-1.medium.com/max/1024/1*LA2mreJG79FgM10-8sT85w.png" data-attrs="{&quot;src&quot;:&quot;https://cdn-images-1.medium.com/max/1024/1*LA2mreJG79FgM10-8sT85w.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://cdn-images-1.medium.com/max/1024/1*LA2mreJG79FgM10-8sT85w.png 424w, https://cdn-images-1.medium.com/max/1024/1*LA2mreJG79FgM10-8sT85w.png 848w, https://cdn-images-1.medium.com/max/1024/1*LA2mreJG79FgM10-8sT85w.png 1272w, https://cdn-images-1.medium.com/max/1024/1*LA2mreJG79FgM10-8sT85w.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">aLRP loss (the current paper) is stable and not overfitting compared to Cross-Entropy Loss and Focal Loss. Image from the pdf of the <a href="https://neurips.cc/virtual/2020/public/poster_b2eeb7362ef83deff5c7813a67e14f0a.html">current&nbsp;paper</a>.</figcaption></figure></div><ul><li><p>As the proposed ranking-based loss function is <strong>not-differentiable</strong>, the authors provided equations for gradients of the loss function with respect to the parameters in the localization head and in the classification head. Here, <strong>the gradient update on parameters of the classification head is affected by the result of both classification head and localization head as well (and vice versa)</strong> so that an instance with less IOU with ground truth is penalized even though the ground truth class label is predicted confidently by the classification head. This makes the classification head work well where the localization head works well and vice versa, <strong>which gives more capacity to the model</strong> to get better at both precision and localization.</p></li></ul><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://cdn-images-1.medium.com/max/939/1*hnVuyxvyKrpnmEN_74wP8Q.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://cdn-images-1.medium.com/max/939/1*hnVuyxvyKrpnmEN_74wP8Q.png 424w, https://cdn-images-1.medium.com/max/939/1*hnVuyxvyKrpnmEN_74wP8Q.png 848w, https://cdn-images-1.medium.com/max/939/1*hnVuyxvyKrpnmEN_74wP8Q.png 1272w, https://cdn-images-1.medium.com/max/939/1*hnVuyxvyKrpnmEN_74wP8Q.png 1456w" sizes="100vw"><img src="https://cdn-images-1.medium.com/max/939/1*hnVuyxvyKrpnmEN_74wP8Q.png" data-attrs="{&quot;src&quot;:&quot;https://cdn-images-1.medium.com/max/939/1*hnVuyxvyKrpnmEN_74wP8Q.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://cdn-images-1.medium.com/max/939/1*hnVuyxvyKrpnmEN_74wP8Q.png 424w, https://cdn-images-1.medium.com/max/939/1*hnVuyxvyKrpnmEN_74wP8Q.png 848w, https://cdn-images-1.medium.com/max/939/1*hnVuyxvyKrpnmEN_74wP8Q.png 1272w, https://cdn-images-1.medium.com/max/939/1*hnVuyxvyKrpnmEN_74wP8Q.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">Gradients wrt parameter in classification head contain outputs of localization and vice versa. Image from the pdf of the <a href="https://neurips.cc/virtual/2020/public/poster_b2eeb7362ef83deff5c7813a67e14f0a.html">current&nbsp;paper</a>.</figcaption></figure></div><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://cdn-images-1.medium.com/max/617/1*lY7YAb7ncXr9LjcEbfyK8g.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://cdn-images-1.medium.com/max/617/1*lY7YAb7ncXr9LjcEbfyK8g.png 424w, https://cdn-images-1.medium.com/max/617/1*lY7YAb7ncXr9LjcEbfyK8g.png 848w, https://cdn-images-1.medium.com/max/617/1*lY7YAb7ncXr9LjcEbfyK8g.png 1272w, https://cdn-images-1.medium.com/max/617/1*lY7YAb7ncXr9LjcEbfyK8g.png 1456w" sizes="100vw"><img src="https://cdn-images-1.medium.com/max/617/1*lY7YAb7ncXr9LjcEbfyK8g.png" data-attrs="{&quot;src&quot;:&quot;https://cdn-images-1.medium.com/max/617/1*lY7YAb7ncXr9LjcEbfyK8g.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://cdn-images-1.medium.com/max/617/1*lY7YAb7ncXr9LjcEbfyK8g.png 424w, https://cdn-images-1.medium.com/max/617/1*lY7YAb7ncXr9LjcEbfyK8g.png 848w, https://cdn-images-1.medium.com/max/617/1*lY7YAb7ncXr9LjcEbfyK8g.png 1272w, https://cdn-images-1.medium.com/max/617/1*lY7YAb7ncXr9LjcEbfyK8g.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">aLRP Loss beats other losses for Faster R-CNN. Image from the pdf of the <a href="https://neurips.cc/virtual/2020/public/poster_b2eeb7362ef83deff5c7813a67e14f0a.html">current&nbsp;paper</a>.</figcaption></figure></div><blockquote><p><strong>Takeaway:</strong> Stability when training and having fewer hyper-parameters to tune is much desired in practive. I can remember a lot scenarios where results are not reproducable. This type of work would be more valuable for a deep learning engineer and I recommend one using it when training your next object detection model.</p></blockquote><h4><a href="https://neurips.cc/virtual/2020/public/poster_b5d17ed2b502da15aa727af0d51508d6.html">Disentangling Human Error from the Ground Truth in Segmentation of Medical&nbsp;Images</a></h4><p>Labeling in the medical image domain is cost-intensive and have a large inter-observer variability. A method that <strong>combines annotations from different annotators while modeling an annotator across images</strong> so that we can train with only a few annotations per image is desirable. This is that&nbsp;method.</p><p>Given an image with 3 ground truth masks labeled by three different annotators A1, A2, and A3, this work, which also models biases of each annotator, tries to <strong>predict three different versions of segmentation masks</strong> one for each annotator and tries to <strong>backpropagate the loss between these 3 predicted masks and 3 ground truth&nbsp;masks</strong>.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://cdn-images-1.medium.com/max/963/1*Rw-LKg8qDI4jSYv1Y-2Sug.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://cdn-images-1.medium.com/max/963/1*Rw-LKg8qDI4jSYv1Y-2Sug.png 424w, https://cdn-images-1.medium.com/max/963/1*Rw-LKg8qDI4jSYv1Y-2Sug.png 848w, https://cdn-images-1.medium.com/max/963/1*Rw-LKg8qDI4jSYv1Y-2Sug.png 1272w, https://cdn-images-1.medium.com/max/963/1*Rw-LKg8qDI4jSYv1Y-2Sug.png 1456w" sizes="100vw"><img src="https://cdn-images-1.medium.com/max/963/1*Rw-LKg8qDI4jSYv1Y-2Sug.png" data-attrs="{&quot;src&quot;:&quot;https://cdn-images-1.medium.com/max/963/1*Rw-LKg8qDI4jSYv1Y-2Sug.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://cdn-images-1.medium.com/max/963/1*Rw-LKg8qDI4jSYv1Y-2Sug.png 424w, https://cdn-images-1.medium.com/max/963/1*Rw-LKg8qDI4jSYv1Y-2Sug.png 848w, https://cdn-images-1.medium.com/max/963/1*Rw-LKg8qDI4jSYv1Y-2Sug.png 1272w, https://cdn-images-1.medium.com/max/963/1*Rw-LKg8qDI4jSYv1Y-2Sug.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">Illustration showing the prediction of Estimated true label distribution segmentation mask and then from that three annotator-specific masks and finally backpropagation through the three branches. Image from the pdf of the <a href="https://neurips.cc/virtual/2020/public/poster_b5d17ed2b502da15aa727af0d51508d6.html">current&nbsp;paper</a>.</figcaption></figure></div><p>As these annotator-specific segmentation masks are created with distortion (confusion matrix for each annotator) <strong>from the estimated true label which is predicted first</strong>, we would take the segmentation mask of the estimated true label as the prediction from the model during inference.</p><blockquote><p><strong>Takeaway:</strong> If your application has more inter-observer variability and you have the bandwidth to get multiple annotations per image, this seems to be the go-to right now&nbsp;to&nbsp;get&nbsp;one&nbsp;ground&nbsp;truth out of&nbsp;many.</p></blockquote><h4><a href="https://neurips.cc/virtual/2020/public/poster_bacadc62d6e67d7897cef027fa2d416c.html">Variational Amodal Object Completion</a></h4><p>Predicting segmentation maps for a complete object when it is occluded is called<strong> Amodal Object Completion</strong>.</p><p>This work presents Amodel-VAE, which <strong>encodes the partial mask into a latent vector and predicts a complete mask decoding that latent vector</strong>. This work <strong>doesn&#8217;t require full-object segmentation annotations</strong> for training making it desirable as previous works needed complete segmentation masks annotated.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://cdn-images-1.medium.com/max/942/1*UUK4BreUAi0mOuyaUz0yCQ.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://cdn-images-1.medium.com/max/942/1*UUK4BreUAi0mOuyaUz0yCQ.png 424w, https://cdn-images-1.medium.com/max/942/1*UUK4BreUAi0mOuyaUz0yCQ.png 848w, https://cdn-images-1.medium.com/max/942/1*UUK4BreUAi0mOuyaUz0yCQ.png 1272w, https://cdn-images-1.medium.com/max/942/1*UUK4BreUAi0mOuyaUz0yCQ.png 1456w" sizes="100vw"><img src="https://cdn-images-1.medium.com/max/942/1*UUK4BreUAi0mOuyaUz0yCQ.png" data-attrs="{&quot;src&quot;:&quot;https://cdn-images-1.medium.com/max/942/1*UUK4BreUAi0mOuyaUz0yCQ.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://cdn-images-1.medium.com/max/942/1*UUK4BreUAi0mOuyaUz0yCQ.png 424w, https://cdn-images-1.medium.com/max/942/1*UUK4BreUAi0mOuyaUz0yCQ.png 848w, https://cdn-images-1.medium.com/max/942/1*UUK4BreUAi0mOuyaUz0yCQ.png 1272w, https://cdn-images-1.medium.com/max/942/1*UUK4BreUAi0mOuyaUz0yCQ.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">Image from the pdf of the <a href="https://neurips.cc/virtual/2020/public/poster_bacadc62d6e67d7897cef027fa2d416c.html">current&nbsp;paper</a>.</figcaption></figure></div><p>To train without complete masks, they carefully train Amodel-VAE in three&nbsp;stages.</p><ul><li><p>At stage I, a decoder P(y_complete/z) is pre-trained <strong>with only masks that are complete</strong> thus learning a mapping from latent vector space to the space of complete&nbsp;masks.</p></li><li><p>At stage II, <strong>occluded partial masks are synthetically generated</strong> from a complete mask by randomly overlaying other objects (foreground) on it so that we will have a mapping between partial masks and complete masks. A <strong>VAE is trained with a pre-trained and frozen decoder</strong> to learn an encoder P(z/y_partial).</p></li><li><p>Finally, at stage III, encoder P(z/y_partial) is fine-tuned so that it could <strong>encode more complex occlusions</strong> which occur in the real-world dataset while the loss is propagated from the visible part of the object expecting the decoder to predict P(y_vis/z). (Decoder is not trained in this step. And the encoder is trading off <strong>some capability to produce latent vectors that predict full masks</strong> to its capability of encoding real-world/complex occlusions)</p></li></ul><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://cdn-images-1.medium.com/max/980/1*UUYX5qV1LFJthsXvZQz87w.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://cdn-images-1.medium.com/max/980/1*UUYX5qV1LFJthsXvZQz87w.png 424w, https://cdn-images-1.medium.com/max/980/1*UUYX5qV1LFJthsXvZQz87w.png 848w, https://cdn-images-1.medium.com/max/980/1*UUYX5qV1LFJthsXvZQz87w.png 1272w, https://cdn-images-1.medium.com/max/980/1*UUYX5qV1LFJthsXvZQz87w.png 1456w" sizes="100vw"><img src="https://cdn-images-1.medium.com/max/980/1*UUYX5qV1LFJthsXvZQz87w.png" data-attrs="{&quot;src&quot;:&quot;https://cdn-images-1.medium.com/max/980/1*UUYX5qV1LFJthsXvZQz87w.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://cdn-images-1.medium.com/max/980/1*UUYX5qV1LFJthsXvZQz87w.png 424w, https://cdn-images-1.medium.com/max/980/1*UUYX5qV1LFJthsXvZQz87w.png 848w, https://cdn-images-1.medium.com/max/980/1*UUYX5qV1LFJthsXvZQz87w.png 1272w, https://cdn-images-1.medium.com/max/980/1*UUYX5qV1LFJthsXvZQz87w.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">Different tasks made possible by Amodal-VAE. Image from the pdf of the<a href="https://neurips.cc/virtual/2020/public/poster_bacadc62d6e67d7897cef027fa2d416c.html"> current&nbsp;paper</a>.</figcaption></figure></div><blockquote><p><strong>Takeaway:</strong> Practically, knowing the complete locations of objects in occlusion would help to track multiple people and decrease Id-swaps that we see even in SOTA tracking models. It should be interesting if you want to smart photoshop as well. More importantly, this is kind of a problem where use cases are limited only by our creativity.</p></blockquote><h4><a href="https://neurips.cc/virtual/2020/public/poster_d85b63ef0ccb114d0a3bb7b7d808028f.html">RandAugment: Practical Automated Data Augmentation with a Reduced Search&nbsp;Space</a></h4><p>Automated data augmentation needs to find <strong>the probability of each transformation and the magnitude</strong> <strong>to be used</strong> for each of these transformations.</p><p>With large possible values for probabilities and magnitudes for each of the transformations, <strong>search space becomes intractable</strong>. Recent method AutoAugment used <strong>RL to find an optimal sequence of transformations and their magnitudes</strong>. More recent variants of AutoAugment tried to make use of more efficient learning algorithms to find the optimal sequence of transformations <strong>efficiently</strong>.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://cdn-images-1.medium.com/max/1024/1*Ab_mqxTRhbcto2JOh_RR2w.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://cdn-images-1.medium.com/max/1024/1*Ab_mqxTRhbcto2JOh_RR2w.png 424w, https://cdn-images-1.medium.com/max/1024/1*Ab_mqxTRhbcto2JOh_RR2w.png 848w, https://cdn-images-1.medium.com/max/1024/1*Ab_mqxTRhbcto2JOh_RR2w.png 1272w, https://cdn-images-1.medium.com/max/1024/1*Ab_mqxTRhbcto2JOh_RR2w.png 1456w" sizes="100vw"><img src="https://cdn-images-1.medium.com/max/1024/1*Ab_mqxTRhbcto2JOh_RR2w.png" data-attrs="{&quot;src&quot;:&quot;https://cdn-images-1.medium.com/max/1024/1*Ab_mqxTRhbcto2JOh_RR2w.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://cdn-images-1.medium.com/max/1024/1*Ab_mqxTRhbcto2JOh_RR2w.png 424w, https://cdn-images-1.medium.com/max/1024/1*Ab_mqxTRhbcto2JOh_RR2w.png 848w, https://cdn-images-1.medium.com/max/1024/1*Ab_mqxTRhbcto2JOh_RR2w.png 1272w, https://cdn-images-1.medium.com/max/1024/1*Ab_mqxTRhbcto2JOh_RR2w.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">Image from the pdf of the <a href="https://neurips.cc/virtual/2020/public/poster_d85b63ef0ccb114d0a3bb7b7d808028f.html">current&nbsp;paper</a>.</figcaption></figure></div><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://cdn-images-1.medium.com/max/1024/1*zFIlD7yqYxgf5nysBc4kRA.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://cdn-images-1.medium.com/max/1024/1*zFIlD7yqYxgf5nysBc4kRA.png 424w, https://cdn-images-1.medium.com/max/1024/1*zFIlD7yqYxgf5nysBc4kRA.png 848w, https://cdn-images-1.medium.com/max/1024/1*zFIlD7yqYxgf5nysBc4kRA.png 1272w, https://cdn-images-1.medium.com/max/1024/1*zFIlD7yqYxgf5nysBc4kRA.png 1456w" sizes="100vw"><img src="https://cdn-images-1.medium.com/max/1024/1*zFIlD7yqYxgf5nysBc4kRA.png" data-attrs="{&quot;src&quot;:&quot;https://cdn-images-1.medium.com/max/1024/1*zFIlD7yqYxgf5nysBc4kRA.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://cdn-images-1.medium.com/max/1024/1*zFIlD7yqYxgf5nysBc4kRA.png 424w, https://cdn-images-1.medium.com/max/1024/1*zFIlD7yqYxgf5nysBc4kRA.png 848w, https://cdn-images-1.medium.com/max/1024/1*zFIlD7yqYxgf5nysBc4kRA.png 1272w, https://cdn-images-1.medium.com/max/1024/1*zFIlD7yqYxgf5nysBc4kRA.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">Image from the pdf of the <a href="https://neurips.cc/virtual/2020/public/poster_d85b63ef0ccb114d0a3bb7b7d808028f.html">current&nbsp;paper</a>.</figcaption></figure></div><p>Nonetheless, the number of iterations of training a model with a set of transformations to find the optimal probability and magnitude values for transformations is still intractable in practice if we are doing it on large-scale models and large-scale datasets. So, <strong>proxy tasks are set up</strong>, with small models and less data among other tweaks, representative of the target task. Optimal probabilities and magnitudes are <strong>found on proxy tasks and are used for the target&nbsp;task</strong>.</p><p>But that these proxy tasks are not actually representative of the complete target tasks. This work showed that the &#8220;<strong>optimal magnitude of augmentation depends on the size of the model and the training&nbsp;set.</strong>&#8221;</p><p>Now, to make this optimal policy search feasible, this current work proposed RandAugment which is just a grid search on two parameters with <strong>~30 orders of magnitude smaller search space</strong>. This is, for sure, one of the few<strong> simple-but-powerful and back-to-basics</strong> kinds of work you could&nbsp;find.</p><p>First, RandAugment <strong>picks transformations with uniform probability.</strong> Because they observed that optimal policies from AutoAugment are making <strong>the dataset visually diverse</strong> rather than selecting a preferred set of particular transformations (different probabilities for different transformations).</p><p>Second, RandAugment has the same magnitude for all the transformations. Because they observed that optimal policies from an AutoAugment variant had <strong>similar magnitudes for all the transformations</strong>.</p><p>After these adjustments, automated data augmentation became a simple hyperparameter tuning task which could be done with a grid search and the whole algorithm might be <strong>written comfortably in 3&nbsp;lines</strong>.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://cdn-images-1.medium.com/max/680/1*yWjxCIolZHh5GDx2ZfYJQA.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://cdn-images-1.medium.com/max/680/1*yWjxCIolZHh5GDx2ZfYJQA.png 424w, https://cdn-images-1.medium.com/max/680/1*yWjxCIolZHh5GDx2ZfYJQA.png 848w, https://cdn-images-1.medium.com/max/680/1*yWjxCIolZHh5GDx2ZfYJQA.png 1272w, https://cdn-images-1.medium.com/max/680/1*yWjxCIolZHh5GDx2ZfYJQA.png 1456w" sizes="100vw"><img src="https://cdn-images-1.medium.com/max/680/1*yWjxCIolZHh5GDx2ZfYJQA.png" data-attrs="{&quot;src&quot;:&quot;https://cdn-images-1.medium.com/max/680/1*yWjxCIolZHh5GDx2ZfYJQA.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://cdn-images-1.medium.com/max/680/1*yWjxCIolZHh5GDx2ZfYJQA.png 424w, https://cdn-images-1.medium.com/max/680/1*yWjxCIolZHh5GDx2ZfYJQA.png 848w, https://cdn-images-1.medium.com/max/680/1*yWjxCIolZHh5GDx2ZfYJQA.png 1272w, https://cdn-images-1.medium.com/max/680/1*yWjxCIolZHh5GDx2ZfYJQA.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">3-line code for RandAugment. Image from the pdf of the <a href="https://neurips.cc/virtual/2020/public/poster_d85b63ef0ccb114d0a3bb7b7d808028f.html">current&nbsp;paper</a>.</figcaption></figure></div><blockquote><p><strong>Takeaway</strong>: Automated data augmentation evolved to a point that it is feasible to use in our &#8216;everyday&#8217; models. If you have resources to do hyperparameter tuning, tune these two parameters (N and M for number of transformations and their global magnitude) as well and get state-of-the-art results.</p></blockquote><h4>Learning Loss for Test-Time Augmentation</h4><p><a href="https://neurips.cc/virtual/2020/public/poster_2ba596643cbbbc20318224181fa46b28.html">https://neurips.cc/virtual/2020/public/poster_2ba596643cbbbc20318224181fa46b28.html</a></p><p>Let&#8217;s assume you want to test your model on a rotated image and images in your training set are<strong> never rotated or rotation data augmentation is not used</strong> while training. The best possible thing we could do is to do the rotation now at test time to make the images not rotated. And with 10 commonly used and naturally occurring transformations this <strong>could happen without you&nbsp;knowing</strong>.</p><p>So, what is the solution? While training, have a <strong>separate network that predicts the loss of a model for each of the transformations</strong> if applied to the&nbsp;image.</p><p>Using this model, apply only the transformations which give lower loss values at test&nbsp;time.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://cdn-images-1.medium.com/max/1024/1*X3ADz8vWePJAOtLlh-Djgg.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://cdn-images-1.medium.com/max/1024/1*X3ADz8vWePJAOtLlh-Djgg.png 424w, https://cdn-images-1.medium.com/max/1024/1*X3ADz8vWePJAOtLlh-Djgg.png 848w, https://cdn-images-1.medium.com/max/1024/1*X3ADz8vWePJAOtLlh-Djgg.png 1272w, https://cdn-images-1.medium.com/max/1024/1*X3ADz8vWePJAOtLlh-Djgg.png 1456w" sizes="100vw"><img src="https://cdn-images-1.medium.com/max/1024/1*X3ADz8vWePJAOtLlh-Djgg.png" data-attrs="{&quot;src&quot;:&quot;https://cdn-images-1.medium.com/max/1024/1*X3ADz8vWePJAOtLlh-Djgg.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://cdn-images-1.medium.com/max/1024/1*X3ADz8vWePJAOtLlh-Djgg.png 424w, https://cdn-images-1.medium.com/max/1024/1*X3ADz8vWePJAOtLlh-Djgg.png 848w, https://cdn-images-1.medium.com/max/1024/1*X3ADz8vWePJAOtLlh-Djgg.png 1272w, https://cdn-images-1.medium.com/max/1024/1*X3ADz8vWePJAOtLlh-Djgg.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">Proposed test-time augmentation (b) using a loss prediction model while inference. Image from the pdf of the <a href="https://neurips.cc/virtual/2020/public/poster_2ba596643cbbbc20318224181fa46b28.html">current&nbsp;paper</a>.</figcaption></figure></div><blockquote><p><strong>Takeaway:</strong> Didn&#8217;t train your model with necessary data augmentations? Want the best possible results on the test set? Use the above test-time Augmentation.</p></blockquote><div><hr></div><p><a href="https://medium.com/data-science/neurips-2020-papers-takeaways-of-a-deep-learning-engineer-part-2-of-3-computer-vision-ef5ea1abe525">NeurIPS 2020 Papers: Takeaways of a Deep Learning Engineer (Part 2 of 3)&#8212; Computer Vision</a> was originally published in <a href="https://medium.com/data-science">TDS Archive</a> on Medium, where people are continuing the conversation by highlighting and responding to this story.</p>]]></content:encoded></item><item><title><![CDATA[NeurIPS 2020 Papers: A Deep Learning Engineer’s Takeaway]]></title><description><![CDATA[NeurIPS 2020 Papers: Takeaways for a Deep Learning Engineer]]></description><link>https://www.aipapertrails.com/p/neurips-2020-papers-a-deep-learning-engineers-takeaway-4f3066523151</link><guid isPermaLink="false">https://www.aipapertrails.com/p/neurips-2020-papers-a-deep-learning-engineers-takeaway-4f3066523151</guid><dc:creator><![CDATA[Prakash Kagitha]]></dc:creator><pubDate>Fri, 27 Nov 2020 19:50:21 GMT</pubDate><enclosure url="https://cdn-images-1.medium.com/max/1024/1*P6eCrob_2P1cF8GJDo7law.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<h3>NeurIPS 2020 Papers: Takeaways for a Deep Learning&nbsp;Engineer</h3><h4>Techniques and insights for applied deep learning from papers published at NeurIPS&nbsp;2020</h4><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://cdn-images-1.medium.com/max/1024/1*P6eCrob_2P1cF8GJDo7law.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://cdn-images-1.medium.com/max/1024/1*P6eCrob_2P1cF8GJDo7law.png 424w, https://cdn-images-1.medium.com/max/1024/1*P6eCrob_2P1cF8GJDo7law.png 848w, https://cdn-images-1.medium.com/max/1024/1*P6eCrob_2P1cF8GJDo7law.png 1272w, https://cdn-images-1.medium.com/max/1024/1*P6eCrob_2P1cF8GJDo7law.png 1456w" sizes="100vw"><img src="https://cdn-images-1.medium.com/max/1024/1*P6eCrob_2P1cF8GJDo7law.png" data-attrs="{&quot;src&quot;:&quot;https://cdn-images-1.medium.com/max/1024/1*P6eCrob_2P1cF8GJDo7law.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://cdn-images-1.medium.com/max/1024/1*P6eCrob_2P1cF8GJDo7law.png 424w, https://cdn-images-1.medium.com/max/1024/1*P6eCrob_2P1cF8GJDo7law.png 848w, https://cdn-images-1.medium.com/max/1024/1*P6eCrob_2P1cF8GJDo7law.png 1272w, https://cdn-images-1.medium.com/max/1024/1*P6eCrob_2P1cF8GJDo7law.png 1456w" sizes="100vw" fetchpriority="high"></picture><div></div></div></a><figcaption class="image-caption">Image by&nbsp;Author</figcaption></figure></div><p>Advances in Deep Learning research are of great utility for a Deep Learning engineer working on real-world problems as most of the Deep Learning research is empirical with validation of new techniques and theories done on datasets that <strong>closely resemble real-world datasets/tasks</strong> (ImageNet pre-trained weights are still useful!).</p><p>But, churning a vast amount of research to acquire techniques, insights, and perspectives that are relevant to a DL engineer is time-consuming, stressful, and <strong>not the least overwhelming</strong>.</p><p>For what so ever reason, I am crazy (I mean, really crazy! See Exhibit A <a href="https://towardsdatascience.com/reading-abstracts-from-nips-neurips-2018-here-is-what-i-learned-16352d83416f">here</a> and <a href="https://twitter.com/prakashkagitha/status/1198503868977385472?s=20">here</a>) about Deep Learning research and also have to justify a Deep Learning engineer&#8217;s role to earn my living. So, this is a great place to be in to cater to these needs of DL engineer relevant research churning.</p><p>Therefore, I went through all the titles of NeurIPS 2020 papers (more than 1900!) and read abstracts of 175 papers, and <strong>extracted DL engineer relevant insights from the following papers</strong>.</p><p>Now, sit back and&nbsp;enjoy.</p><p>This is part 1. See the other parts&nbsp;below.</p><p><a href="https://towardsdatascience.com/neurips-2020-papers-takeaways-of-a-deep-learning-engineer-part-2-of-3-computer-vision-ef5ea1abe525">NeurIPS 2020 Papers: Takeaways of a Deep Learning Engineer (Part 2 of 3)&#8212; Computer Vision</a></p><h4><a href="https://neurips.cc/virtual/2020/public/poster_a1140a3d0df1c81e24ae954d935e8926.html">Accelerating Training of Transformer-Based Language Models with Progressive Layer&nbsp;Dropping</a></h4><p>2.5x faster pre-training with <strong>Switchable Transformers(ST)</strong> compared to standard Transformers.</p><p>Equipped by Switchable Gates (G in the fig. below), some of the layers are skipped randomly according to sampled 0 or 1 from a Bernoulli distribution which is 25% time-efficient per&nbsp;sample.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://cdn-images-1.medium.com/max/988/1*l1mzl6SYOIoedGknBKSZSA.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://cdn-images-1.medium.com/max/988/1*l1mzl6SYOIoedGknBKSZSA.png 424w, https://cdn-images-1.medium.com/max/988/1*l1mzl6SYOIoedGknBKSZSA.png 848w, https://cdn-images-1.medium.com/max/988/1*l1mzl6SYOIoedGknBKSZSA.png 1272w, https://cdn-images-1.medium.com/max/988/1*l1mzl6SYOIoedGknBKSZSA.png 1456w" sizes="100vw"><img src="https://cdn-images-1.medium.com/max/988/1*l1mzl6SYOIoedGknBKSZSA.png" data-attrs="{&quot;src&quot;:&quot;https://cdn-images-1.medium.com/max/988/1*l1mzl6SYOIoedGknBKSZSA.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://cdn-images-1.medium.com/max/988/1*l1mzl6SYOIoedGknBKSZSA.png 424w, https://cdn-images-1.medium.com/max/988/1*l1mzl6SYOIoedGknBKSZSA.png 848w, https://cdn-images-1.medium.com/max/988/1*l1mzl6SYOIoedGknBKSZSA.png 1272w, https://cdn-images-1.medium.com/max/988/1*l1mzl6SYOIoedGknBKSZSA.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">(a) Standard Transformer (b) Reordering to make it PreLN (c) Switchable Gates (G) to decide whether to include a layer or not. (The image is reproduced from the pdf of the <a href="https://neurips.cc/virtual/2020/public/poster_a1140a3d0df1c81e24ae954d935e8926.html">current&nbsp;paper</a>.)</figcaption></figure></div><p>And remarkably, it is shown that it reaches the same validation error as the baselines with 53% fewer training&nbsp;samples.</p><p>Combining both time and sample efficiency, <strong>pre-training is 2.5x faster</strong> with comparable and sometimes better performance on downstream tasks.</p><blockquote><p><strong>Takeaway:</strong> When you want to pretrain or finetune a transformer, try out Switchable Transformers for faster training along with low inference times.</p></blockquote><h4><a href="https://neurips.cc/virtual/2020/public/poster_8493eeaccb772c0878f99d60a0bd2bb3.html">Coresets for Robust Training of Neural Networks against Noisy&nbsp;Labels</a></h4><p>It is shown before that the Jacobian of neural network weights(W) and clean data (X), after some training, <strong>would approximate to a low-rank matrix</strong>, with a few large singular values and a lot of very small singular&nbsp;values.</p><p>Also, learning which generalizes (i.e from clean data) is in a low-dimensional space called <strong>Information space (I)</strong> and learning that doesn't generalize (i.e. from noisy labels, mostly memorization) is in a high-dimensional space called <strong>Nuisance space&nbsp;(N)</strong>.</p><p>The current work introduces a technique that <strong>creates sets of mostly clean data (Coresets) to train a model with</strong> and show a significant increase in performance on noisy datasets i.e. 7% increase on mini Webvision with 50% noisy labels compared to the state-of-the-art.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://cdn-images-1.medium.com/max/1024/1*Q8RAROrANTkHpSyHTl6WKw.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://cdn-images-1.medium.com/max/1024/1*Q8RAROrANTkHpSyHTl6WKw.png 424w, https://cdn-images-1.medium.com/max/1024/1*Q8RAROrANTkHpSyHTl6WKw.png 848w, https://cdn-images-1.medium.com/max/1024/1*Q8RAROrANTkHpSyHTl6WKw.png 1272w, https://cdn-images-1.medium.com/max/1024/1*Q8RAROrANTkHpSyHTl6WKw.png 1456w" sizes="100vw"><img src="https://cdn-images-1.medium.com/max/1024/1*Q8RAROrANTkHpSyHTl6WKw.png" data-attrs="{&quot;src&quot;:&quot;https://cdn-images-1.medium.com/max/1024/1*Q8RAROrANTkHpSyHTl6WKw.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://cdn-images-1.medium.com/max/1024/1*Q8RAROrANTkHpSyHTl6WKw.png 424w, https://cdn-images-1.medium.com/max/1024/1*Q8RAROrANTkHpSyHTl6WKw.png 848w, https://cdn-images-1.medium.com/max/1024/1*Q8RAROrANTkHpSyHTl6WKw.png 1272w, https://cdn-images-1.medium.com/max/1024/1*Q8RAROrANTkHpSyHTl6WKw.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">The method introduced in this work, CRUST, performs significantly better than the state-of-the-art. (The image is reproduced from the pdf of the <a href="https://neurips.cc/virtual/2020/public/poster_8493eeaccb772c0878f99d60a0bd2bb3.html">current&nbsp;paper.</a>)</figcaption></figure></div><blockquote><p><strong>Takeaway:</strong> When you suspect the dataset you collected has noisy/mislabeled data points, use CRUST to train the model only on the clean data and improve performance and robustness.</p></blockquote><h4><a href="https://neurips.cc/virtual/2020/public/poster_b6af2c9703f203a2794be03d443af2e3.html">The Lottery Ticket Hypothesis for Pre-trained BERT&nbsp;Networks</a></h4><p>There exists <strong>a sub-network that exhibits performance comparable to the original complete network</strong> while the training process is the same. These sub-networks are called lottery tickets and are defined by masks that tell which weight is zeroed out in the original&nbsp;network.</p><p>Current work adopted <strong>Iterative Magnitude pruning (IMP)</strong> which trains a subnetwork for some time and prunes k% weights which are of less magnitude. This process is repeated multiple times until the sparsity reaches the target sparsity. Important thing is that after every iteration of training the model starts again with the initial parameters rather than weights updated till then, which is called rewinding.</p><p>Here, the pre-trained weights of the BERT are the initialization we start IMP with. And the lottery ticket which is a subnetwork of the pre-trained BERT also contains <strong>the same pre-trained weights</strong> with some of them zeroed&nbsp;out.</p><p>This work showed that the lottery ticket hypothesis holds for pre-trained BERT models as well. And <strong>found subnetworks at 40% to 90% sparsity</strong> for a range of downstream tasks.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://cdn-images-1.medium.com/max/1024/1*xjykLz_3fqFoKrybWxmeiQ.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://cdn-images-1.medium.com/max/1024/1*xjykLz_3fqFoKrybWxmeiQ.png 424w, https://cdn-images-1.medium.com/max/1024/1*xjykLz_3fqFoKrybWxmeiQ.png 848w, https://cdn-images-1.medium.com/max/1024/1*xjykLz_3fqFoKrybWxmeiQ.png 1272w, https://cdn-images-1.medium.com/max/1024/1*xjykLz_3fqFoKrybWxmeiQ.png 1456w" sizes="100vw"><img src="https://cdn-images-1.medium.com/max/1024/1*xjykLz_3fqFoKrybWxmeiQ.png" data-attrs="{&quot;src&quot;:&quot;https://cdn-images-1.medium.com/max/1024/1*xjykLz_3fqFoKrybWxmeiQ.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://cdn-images-1.medium.com/max/1024/1*xjykLz_3fqFoKrybWxmeiQ.png 424w, https://cdn-images-1.medium.com/max/1024/1*xjykLz_3fqFoKrybWxmeiQ.png 848w, https://cdn-images-1.medium.com/max/1024/1*xjykLz_3fqFoKrybWxmeiQ.png 1272w, https://cdn-images-1.medium.com/max/1024/1*xjykLz_3fqFoKrybWxmeiQ.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">The last row corresponds to the approach introduced in this paper. Even though it is 40%-90% sparse, performance is comparable to a Full BERT base. (The image is reproduced from the pdf of the <a href="https://neurips.cc/virtual/2020/public/poster_b6af2c9703f203a2794be03d443af2e3.html">current&nbsp;paper.</a>)</figcaption></figure></div><p>Also, the authors found a pre-trained BERT ticket with 70% sparsity which can transfer to many downstream tasks and perform at least as good as or better than a 70% sparse ticket found for that particular downstream task.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://cdn-images-1.medium.com/max/1024/1*WgFg_EquX7_tekSIMmw5Vg.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://cdn-images-1.medium.com/max/1024/1*WgFg_EquX7_tekSIMmw5Vg.png 424w, https://cdn-images-1.medium.com/max/1024/1*WgFg_EquX7_tekSIMmw5Vg.png 848w, https://cdn-images-1.medium.com/max/1024/1*WgFg_EquX7_tekSIMmw5Vg.png 1272w, https://cdn-images-1.medium.com/max/1024/1*WgFg_EquX7_tekSIMmw5Vg.png 1456w" sizes="100vw"><img src="https://cdn-images-1.medium.com/max/1024/1*WgFg_EquX7_tekSIMmw5Vg.png" data-attrs="{&quot;src&quot;:&quot;https://cdn-images-1.medium.com/max/1024/1*WgFg_EquX7_tekSIMmw5Vg.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://cdn-images-1.medium.com/max/1024/1*WgFg_EquX7_tekSIMmw5Vg.png 424w, https://cdn-images-1.medium.com/max/1024/1*WgFg_EquX7_tekSIMmw5Vg.png 848w, https://cdn-images-1.medium.com/max/1024/1*WgFg_EquX7_tekSIMmw5Vg.png 1272w, https://cdn-images-1.medium.com/max/1024/1*WgFg_EquX7_tekSIMmw5Vg.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">Last but one row, (IMP) MLM (70%), shows that there is a general 70% sparse BERT which would generalize to all the downstream task being at least as good as 70% sparse tickets of that particular task. (The image is reproduced from the pdf of the <a href="https://neurips.cc/virtual/2020/public/poster_b6af2c9703f203a2794be03d443af2e3.html">current&nbsp;paper.</a>)</figcaption></figure></div><blockquote><p><strong>Takeaway:</strong> A Deep Learning engineer working on NLP has to finetune pre-trained BERT on a downstream task very often. Instead of from a full-size BERT, start fine-tuning with the 70% sparse lottery ticket found on MLM downstream task (last but one row) to train faster and decrease inference times and memory bandwidth without losing out on performance. It&#8217;s a no-brainer!</p></blockquote><h4><a href="https://neurips.cc/virtual/2020/public/poster_c3a690be93aa602ee2dc0ccab5b7b67e.html">MPNet: Masked and Permuted Pre-training for Language Understanding</a></h4><p>MPNet is a hybrid of Masked Language Modeling(MLM) and auto-regressive Permuted Language Modeling(PLM) <strong>adopting the strengths and avoiding their limitations</strong> from each of its constituents.</p><p>Masked language modeling, as in BERT-style models, mask out ~15% of the data and try to predict those masked tokens. As the dependency between the masked tokens is not modeled it leads to pretrain-finetune discrepancy which is termed as <strong>Output Dependency</strong>.</p><p>On the other side, auto-regressive permuted language modeling, as in XLNet, doesn&#8217;t have entire information about the input sentence i.e when predicting say 5th element in the 8-element sequence the model doesn&#8217;t know that there are 8 elements in the sequence, thus lead to pretrain-finetune discrepancy (as the model see entire input sentence/paragraph in the downstream tasks) which is termed as <strong>Input Consistency</strong>.</p><p>MPNet combines both of them. XLNet-like architecture is modified <strong>by adding additional masks up to the end of the sentence</strong> so that prediction at any position would attend to N number of tokens where N is the length of the sequence, with some of them being&nbsp;masks.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://cdn-images-1.medium.com/max/1024/1*4TGY4m6O3xaMD7LD6ZTi2w.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://cdn-images-1.medium.com/max/1024/1*4TGY4m6O3xaMD7LD6ZTi2w.png 424w, https://cdn-images-1.medium.com/max/1024/1*4TGY4m6O3xaMD7LD6ZTi2w.png 848w, https://cdn-images-1.medium.com/max/1024/1*4TGY4m6O3xaMD7LD6ZTi2w.png 1272w, https://cdn-images-1.medium.com/max/1024/1*4TGY4m6O3xaMD7LD6ZTi2w.png 1456w" sizes="100vw"><img src="https://cdn-images-1.medium.com/max/1024/1*4TGY4m6O3xaMD7LD6ZTi2w.png" data-attrs="{&quot;src&quot;:&quot;https://cdn-images-1.medium.com/max/1024/1*4TGY4m6O3xaMD7LD6ZTi2w.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://cdn-images-1.medium.com/max/1024/1*4TGY4m6O3xaMD7LD6ZTi2w.png 424w, https://cdn-images-1.medium.com/max/1024/1*4TGY4m6O3xaMD7LD6ZTi2w.png 848w, https://cdn-images-1.medium.com/max/1024/1*4TGY4m6O3xaMD7LD6ZTi2w.png 1272w, https://cdn-images-1.medium.com/max/1024/1*4TGY4m6O3xaMD7LD6ZTi2w.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">Illustrative example showing how MPNet combines MLM and PLM. (The image is reproduced from the pdf of the <a href="https://neurips.cc/virtual/2020/public/poster_c3a690be93aa602ee2dc0ccab5b7b67e.html">current&nbsp;paper.</a>)</figcaption></figure></div><p>They use two-stream self-attention which is introduced in XLNet to enable auto-regressive type prediction, <strong>at one go</strong>, where at any position content should be masked for prediction at that step but should be visible for the predictions at later&nbsp;steps.</p><p>&#8220;MPNet outperforms MLM and PLM by a large margin and achieves better results on tasks including GLUE, SQUAD compared with previous state-of-the-art pre-trained methods (e.g., BERT, XLNet, RoBERTa).&#8221;</p><blockquote><p><strong>Takeaway:</strong> If you ever wanted to pretrain a language model on your domain-specific data or with extra data than the state-of-the-art, use MPNet which is shown to have the best of both MLP and PLM&nbsp;worlds.</p></blockquote><h4><a href="https://neurips.cc/virtual/2020/public/poster_c6102b3727b2a7d8b1bb6981147081ef.html">Identifying Mislabeled Data using the Area Under the Margin&nbsp;Ranking</a></h4><p>Mislabeled data is common in large-scale datasets as they are crowdsourced or scraped from the internet which is noise&nbsp;prone.</p><p>This work formulates a simple intuitive idea. Let's say there are 100 dog images but 20 of them are labeled as &#8216;bird&#8217;. And similarly, 100 bird images but labeled 20 of them are labeled as&nbsp;&#8216;dog&#8217;.</p><p>After some training, for an image of a dog wrongly labeled as &#8216;bird&#8217;, the model gives a considerable probability for label &#8216;dog&#8217; because of<strong> generalization from 80 correctly labeled images</strong>. The model also gives a considerable probability for the label &#8216;bird&#8217; as well because of<strong> memorizing those 20 wrongly labeled&nbsp;images</strong>.</p><p>Now, the difference between the probability of &#8216;dog&#8217; and the probability of &#8216;bird&#8217; is called <strong>Area Under the Margin (AUM)</strong>. This work recommends that if AUM is below some pre-defined threshold we should treat it as a wrongly labeled data sample and remove it from training.</p><p>If we can&#8217;t able to settle on one threshold value, we can <strong>populate wrongly labeled data intentionally</strong> and see what is AUM for those examples. This would be our threshold.</p><p>&#8220;On the WebVision50 classification task, this method removes 17% of training data, yielding a 1.6% (absolute) drop in test error. On CIFAR100 removing 13% of the data leads to a 1.2% drop in&nbsp;error.&#8221;</p><blockquote><p><strong>Takeaway:</strong> When creating a dataset, noisy/mislabeled data samples are mostly unavoidable. Then, use the AUM method to find the mislabeled data samples and remove them from the final training&nbsp;dataset.</p></blockquote><h4><a href="https://neurips.cc/virtual/2020/public/poster_e025b6279c1b88d3ec0eca6fcb6e6280.html">Rethinking the Value of Labels for Improving Class-Imbalanced Learning</a></h4><p>Do we need labels when existing labels are class imbalanced (some classes have more labeled examples than others) and we have a lot of unlabeled data?</p><p>Positive. Yes, we need labels. <strong>Self-train on the unlabeled data and you would be golden</strong>. (Self-training is a process where an intermediate model, which is trained on human-labeled data, is used to create &#8216;labels&#8217; (thus, pseudo labels) and then the final model is trained on both human-labeled and intermediate model labeled&nbsp;data).</p><p>Negative. We may do away with the labels. One can <strong>use self-supervised pretraining on all the data available to learn meaningful representations</strong> and then learn the actual classification task. It is shown that this approach improves performance.</p><blockquote><p><strong>Takeaway:</strong> If you have class-imbalanced labels and more unlabeled data, do self-training or self-supervised pretraining. (It is shown that self-training beats self-supervised learning on CIFAR-10-LT though).</p></blockquote><h4>Big Bird: Transformers for Longer Sequences</h4><p><a href="https://neurips.cc/virtual/2020/public/poster_c8512d142a2d849725f31a9a7a361ab9.html">https://neurips.cc/virtual/2020/public/poster_c8512d142a2d849725f31a9a7a361ab9.html</a></p><p>Self-attention in standard Transformers is of quadratic complexity (in memory and computation) wrt sequence length. So, training longer sequences is not feasible.</p><p>Enter Big Bird. It uses sparse attention where a particular position only attends to <strong>a few randomly selected tokens and some neighboring tokens</strong>.</p><p>That&#8217;s not what makes it work though. Big Bird has multiple CLS tokens that attend to the entire sequence. And a token in any position attends to these CLS tokens which give them relevant context, dependencies, and who knows what else self-attention layers&nbsp;learn.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://cdn-images-1.medium.com/max/1024/1*HtLzvFNiEuSKuYj5A23w7A.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://cdn-images-1.medium.com/max/1024/1*HtLzvFNiEuSKuYj5A23w7A.png 424w, https://cdn-images-1.medium.com/max/1024/1*HtLzvFNiEuSKuYj5A23w7A.png 848w, https://cdn-images-1.medium.com/max/1024/1*HtLzvFNiEuSKuYj5A23w7A.png 1272w, https://cdn-images-1.medium.com/max/1024/1*HtLzvFNiEuSKuYj5A23w7A.png 1456w" sizes="100vw"><img src="https://cdn-images-1.medium.com/max/1024/1*HtLzvFNiEuSKuYj5A23w7A.png" data-attrs="{&quot;src&quot;:&quot;https://cdn-images-1.medium.com/max/1024/1*HtLzvFNiEuSKuYj5A23w7A.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://cdn-images-1.medium.com/max/1024/1*HtLzvFNiEuSKuYj5A23w7A.png 424w, https://cdn-images-1.medium.com/max/1024/1*HtLzvFNiEuSKuYj5A23w7A.png 848w, https://cdn-images-1.medium.com/max/1024/1*HtLzvFNiEuSKuYj5A23w7A.png 1272w, https://cdn-images-1.medium.com/max/1024/1*HtLzvFNiEuSKuYj5A23w7A.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">Different types of attention in sparse attention (a) Random attention (b) Window neighborhood attention (c) Global attention on added CLS tokens. (The image is reproduced from the pdf of the <a href="https://neurips.cc/virtual/2020/public/poster_c8512d142a2d849725f31a9a7a361ab9.html">current&nbsp;paper</a>.)</figcaption></figure></div><p>&#8220;Big Bird&#8217;s sparse attention can handle<strong> sequences of length up to 8x of what was previously possible using similar hardware</strong>. As a consequence of the capability to handle longer context, Big Bird drastically improves performance on various NLP tasks such as question answering, summarization, and novel applications to genomics&nbsp;data.&#8221;</p><blockquote><p><strong>Takeaway:</strong> If you are working with longer sentences or sequences like in summarization or applications of genomic data, use Big Bird for feasible training and respectable inference times. Even with smaller sentences, use Big Bird. I will take linear-complexity self-attention rather than quadratic any&nbsp;day!</p></blockquote><h4>Improving Auto-Augment via Augmentation-Wise Weight&nbsp;Sharing</h4><p><a href="https://neurips.cc/virtual/2020/public/poster_dc49dfebb0b00fd44aeff5c60cc1f825.html">https://neurips.cc/virtual/2020/public/poster_dc49dfebb0b00fd44aeff5c60cc1f825.html</a></p><p>Choosing a sequence of transformations and their magnitude for data augmentation for a particular task is domain-specific and time-consuming.</p><p>Auto-Augment is a technique to <strong>learn an optimal sequence of transformations</strong> where the reward is the validation loss negated. Usually, RL is used to learn this policy. One iteration in learning this optimal policy involves training a model completely and thus is a very expensive process.</p><p>So, the current work tries to make this process more efficient. It is based on the insight shown before that while training with a sequence of transformations the effect of the transformations is only prominent at the later stage of training.</p><p>In this current work, for each iteration to evaluate a particular policy (sequence of transformations), <strong>most of the training is done with a shared policy</strong>, and only the last part of the training is done with the current policy to be evaluated. This is called Augmentation-Wise Weight&nbsp;Sharing.</p><p>As the training with the shared policy is <strong>done only once for all the iterations </strong>this method is efficient in learning an optimal&nbsp;policy.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://cdn-images-1.medium.com/max/1024/1*KQL7FxK_-NBGh5mhGucCRg.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://cdn-images-1.medium.com/max/1024/1*KQL7FxK_-NBGh5mhGucCRg.png 424w, https://cdn-images-1.medium.com/max/1024/1*KQL7FxK_-NBGh5mhGucCRg.png 848w, https://cdn-images-1.medium.com/max/1024/1*KQL7FxK_-NBGh5mhGucCRg.png 1272w, https://cdn-images-1.medium.com/max/1024/1*KQL7FxK_-NBGh5mhGucCRg.png 1456w" sizes="100vw"><img src="https://cdn-images-1.medium.com/max/1024/1*KQL7FxK_-NBGh5mhGucCRg.png" data-attrs="{&quot;src&quot;:&quot;https://cdn-images-1.medium.com/max/1024/1*KQL7FxK_-NBGh5mhGucCRg.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://cdn-images-1.medium.com/max/1024/1*KQL7FxK_-NBGh5mhGucCRg.png 424w, https://cdn-images-1.medium.com/max/1024/1*KQL7FxK_-NBGh5mhGucCRg.png 848w, https://cdn-images-1.medium.com/max/1024/1*KQL7FxK_-NBGh5mhGucCRg.png 1272w, https://cdn-images-1.medium.com/max/1024/1*KQL7FxK_-NBGh5mhGucCRg.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">Two stages of training the model when evaluating the given policy. (The image is reproduced from the pdf of the <a href="https://neurips.cc/virtual/2020/public/poster_dc49dfebb0b00fd44aeff5c60cc1f825.html">current&nbsp;paper</a>.)</figcaption></figure></div><p>&#8220;On CIFAR-10, this method achieves a top-1 error rate of 1.24%, which is currently the best performing single model without extra training data. On ImageNet, this method gets a top-1 error rate of 20.36% for ResNet-50, which leads to a 3.34% absolute error rate reduction over the baseline augmentation.&#8221;</p><blockquote><p><strong>Takeaway:</strong> When you have resources to use an optimal sequence of data augmentations to increase the performance of a model, use this method to train the RL agent which learns the optimal policy which is more efficient also making Auto-Augmentation feasible for large datasets.</p></blockquote><h4>Fast Transformers with Clustered Attention</h4><p><a href="https://neurips.cc/virtual/2020/public/poster_f6a8dd1c954c8506aadc764cc32b895e.html">https://neurips.cc/virtual/2020/public/poster_f6a8dd1c954c8506aadc764cc32b895e.html</a></p><p>Like Big Bird above, Fast Transformers approximates the standard self-attention to make it linear from quadratic dependency.</p><p>To do this, instead of calculating attention all-to-all (O(sequence_length*sequence_length)), queries are clustered and the <strong>attention values are calculated only for the centroids</strong>. And all the queries in a particular cluster would get the same attention values. This makes the overall computation of self-attention linear wrt sequence length. O(num_clusters*sequence_length).</p><p>To improve this approximation by handling the case where there could be some keys which have a large dot product with the centroid query but not with some of the cluster member queries, authors take top-k keys which the centroid query most attended to and <strong>calculate the exact key-value attention values for all the queries in the cluster with those top-k keys</strong>. This increases computation and memory but still is better than all-to-all.</p><p>&#8220;This paper shows that Fast Transformers can approximate arbitrarily complex attention distributions with a minimal number of clusters by approximating a pre-trained BERT model on GLUE and SQuAD benchmarks with only 25 clusters and no loss in performance.&#8221;</p><blockquote><p><strong>Takeaway:</strong> This is not as elegant as Big Bird we saw above but one has to try every option to bring the quadratic complexity of self-attention to&nbsp;linear.</p></blockquote><h4>Limits to Depth Efficiencies of Self-Attention</h4><p><a href="https://neurips.cc/virtual/2020/public/poster_ff4dfdf5904e920ce52b48c1cef97829.html">https://neurips.cc/virtual/2020/public/poster_ff4dfdf5904e920ce52b48c1cef97829.html</a></p><p>To scale transformers, it is empirically shown that increasing the width (dimension of internal representation) is as efficient as increasing the depth (number of self-attention layers).</p><p>Contrarily and more concretely, this work establishes that we can scale the transformers up to the <strong>&#8216;depth threshold&#8217; which is the base 3 logarithm of the width</strong>. If the depth is below this depth threshold increasing depth is more efficient than increasing the width. This is termed depth efficiency.</p><p>And if the depth is higher than this depth threshold increasing depth will hurt compared to increasing the width. This is termed as depth inefficiency.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://cdn-images-1.medium.com/max/1024/1*qBy18dKw_QQHhHDBRJGECg.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://cdn-images-1.medium.com/max/1024/1*qBy18dKw_QQHhHDBRJGECg.png 424w, https://cdn-images-1.medium.com/max/1024/1*qBy18dKw_QQHhHDBRJGECg.png 848w, https://cdn-images-1.medium.com/max/1024/1*qBy18dKw_QQHhHDBRJGECg.png 1272w, https://cdn-images-1.medium.com/max/1024/1*qBy18dKw_QQHhHDBRJGECg.png 1456w" sizes="100vw"><img src="https://cdn-images-1.medium.com/max/1024/1*qBy18dKw_QQHhHDBRJGECg.png" data-attrs="{&quot;src&quot;:&quot;https://cdn-images-1.medium.com/max/1024/1*qBy18dKw_QQHhHDBRJGECg.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://cdn-images-1.medium.com/max/1024/1*qBy18dKw_QQHhHDBRJGECg.png 424w, https://cdn-images-1.medium.com/max/1024/1*qBy18dKw_QQHhHDBRJGECg.png 848w, https://cdn-images-1.medium.com/max/1024/1*qBy18dKw_QQHhHDBRJGECg.png 1272w, https://cdn-images-1.medium.com/max/1024/1*qBy18dKw_QQHhHDBRJGECg.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">The number of parameters is directly proportional to the width of the network when layers are constant. The figure shows when the depth is more useful when it has sufficient width, i.e in depth efficiency phase. (The image is reproduced from the pdf of the <a href="https://neurips.cc/virtual/2020/public/poster_ff4dfdf5904e920ce52b48c1cef97829.html">current&nbsp;paper</a>.)</figcaption></figure></div><p>&#8220;By identifying network width as a limiting factor, our analysis indicates that solutions for dramatically increasing the width can facilitate the next leap in self-attention expressivity.&#8221;</p><blockquote><p><strong>Takeaway:</strong> When you want to scale the Transformer architecture for the next big language model, keep in mind that if the width is not large enough increasing depth doesn&#8217;t help. Depth should always be less than the &#8216;depth threshold&#8217; which is the base-3 logarithm of width. So, increase the width before increasing the depth to scale your transformers to almost insane&nbsp;depth.</p></blockquote><div><hr></div><p><a href="https://medium.com/data-science/neurips-2020-papers-a-deep-learning-engineers-takeaway-4f3066523151">NeurIPS 2020 Papers: A Deep Learning Engineer&#8217;s Takeaway</a> was originally published in <a href="https://medium.com/data-science">TDS Archive</a> on Medium, where people are continuing the conversation by highlighting and responding to this story.</p>]]></content:encoded></item><item><title><![CDATA[ICLR 2020 papers: On Compositional/Systematic generalization (including ours)]]></title><description><![CDATA[An introduction to systematic generalization and a summary of 4 papers from ICLR 2020]]></description><link>https://www.aipapertrails.com/p/iclr-2020-papers-on-compositional-systematic-generalization-including-ours-54a45badfcae</link><guid isPermaLink="false">https://www.aipapertrails.com/p/iclr-2020-papers-on-compositional-systematic-generalization-including-ours-54a45badfcae</guid><dc:creator><![CDATA[Prakash Kagitha]]></dc:creator><pubDate>Sun, 26 Apr 2020 05:56:46 GMT</pubDate><enclosure url="https://cdn-images-1.medium.com/max/1024/0*HHerFTDHFQ9ryWD5" length="0" type="image/jpeg"/><content:encoded><![CDATA[<h4>An introduction to systematic generalization and a summary of 4 papers from ICLR&nbsp;2020</h4><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://cdn-images-1.medium.com/max/1024/0*HHerFTDHFQ9ryWD5" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://cdn-images-1.medium.com/max/1024/0*HHerFTDHFQ9ryWD5 424w, https://cdn-images-1.medium.com/max/1024/0*HHerFTDHFQ9ryWD5 848w, https://cdn-images-1.medium.com/max/1024/0*HHerFTDHFQ9ryWD5 1272w, https://cdn-images-1.medium.com/max/1024/0*HHerFTDHFQ9ryWD5 1456w" sizes="100vw"><img src="https://cdn-images-1.medium.com/max/1024/0*HHerFTDHFQ9ryWD5" data-attrs="{&quot;src&quot;:&quot;https://cdn-images-1.medium.com/max/1024/0*HHerFTDHFQ9ryWD5&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://cdn-images-1.medium.com/max/1024/0*HHerFTDHFQ9ryWD5 424w, https://cdn-images-1.medium.com/max/1024/0*HHerFTDHFQ9ryWD5 848w, https://cdn-images-1.medium.com/max/1024/0*HHerFTDHFQ9ryWD5 1272w, https://cdn-images-1.medium.com/max/1024/0*HHerFTDHFQ9ryWD5 1456w" sizes="100vw" fetchpriority="high"></picture><div></div></div></a><figcaption class="image-caption">Photo by <a href="https://unsplash.com/@mvdheuvel?utm_source=medium&amp;utm_medium=referral">Maarten van den Heuvel</a> on&nbsp;<a href="https://unsplash.com?utm_source=medium&amp;utm_medium=referral">Unsplash</a></figcaption></figure></div><h4>Systematicity as an argument against deep&nbsp;learning</h4><p>In 1988, Jerry Fodor leveled a concern against connectionist models (deep learning) explaining/modeling human language understanding and cognition that they are not systematic. Meant to say that some of the data points (like sentences) are systematically similar to some of the other data points and humans can understand all of these data points given that they understand one data&nbsp;point.</p><p>For example, if we understand a sentence &#8216;John loves Mary&#8217;, we would also understand &#8216;Mary loves John&#8217; or for that matter, any sentence of the pattern &#8216;NP Vt NP&#8217; because, the underlying knowledge (concepts?) in understanding all these sentences is one and the same i.e understanding syntax &#8216;NP Vt&nbsp;NP&#8217;.</p><p>As per Fodor, the same behavior should be expected of the model explaining language understanding that it should understand all systematically similar data points if it could understand (learns)&nbsp;one.</p><p>As systematicity in human cognition is very strong, this argument against connectionist models has been very prominent, historically, stimulating debate and a great exchange of scientific arguments.</p><h4>Modernization of systematicity by Lake et.al.&nbsp;2017</h4><p><a href="https://arxiv.org/abs/1711.00350">Lake et. al. 2017</a> instantiated this expectation as an aspect of generalization (systematic generalization) in sequence to sequence(Seq2Seq) learning models with the SCAN dataset. The model is tested on new combinations of already learning concepts as a requirement for the model to be systematic.</p><p>For example, the model trained on the input-output pairs (walk, WALK), (jump, JUMP), and (walk left, LTURN WALK) is tested with a pair (jump left, LTURN JUMP). The standard seq2seq models based on GRU, LSTM and their variants with attention failed catastrophically on the SCAN dataset of which the above examples are&nbsp;from.</p><p>After this, different datasets are proposed to test deep learning models on Visual Question Answering (VQA). Most recently, <a href="https://arxiv.org/abs/1912.05783">Bahadanu et.al. 2019 </a>created CLOSURE, a variant of the CLEVR dataset, which tests the model&#8217;s performance on the questions which contain familiar parts in a more complex context. They showed that existing VQA models don&#8217;t perform well on this dataset and proposed a variant of Neural Module Networks to improve the performance.</p><h4>ICLR 2020 and systematic generalization</h4><p>It is interesting that systematic generalization is explored in different tasks and domains as there are four papers related to systematic generalization published at ICLR 2020 with a new way of solving systematic generalization, a new way of creating train-test splits to test systematic generalization, an investigation of the drivers of systematic generalization in an RL agent and finally, our work showing that the standard LSTM+attn models also exhibit systematic generalization.</p><p><strong><a href="https://iclr.cc/virtual/poster_SklGryBtwr.html">Environmental drivers of systematicity and generalization in a situated&nbsp;agent</a></strong></p><p>In this work, an RL agent is trained to follow commands like &#8216;<em>find Obj&#8217; or &#8216;lift Obj&#8217; </em>in a 2D/3D environment with different objects.<em> </em>After training &#8216;<em>find Obj</em>&#8217; with all the objects in the environment and training &#8216;<em>lift Obj</em>&#8217; with a subset of objects it has to generalize to the commands containing &#8216;<em>lift&#8217;</em> with new objects (&#8216;<em>lift NewObj</em>&#8217;).</p><p>This is analogous to the SCAN dataset where the model is tested on new combinations of already learning concepts. Here, the concept of a particular object is learned from &#8216;<em>find</em>&#8217; command and the concept of &#8216;<em>lift&#8217;</em> is learned being trained with lift combined with the objects in the trainset.</p><p>The following are the drivers of systematicity found in this investigation which would give a lot of insight into systematic generalization of deep learning models. Check out the paper for more information.</p><blockquote><p>&#8220;[a] the number of object/word experiences in the training&nbsp;set;</p></blockquote><blockquote><p>[b] the visual invariances afforded by the agent&#8217;s perspective, or frame of reference; and</p></blockquote><blockquote><p>[c] the variety of visual input inherent in the perceptual aspect of the agent&#8217;s perception.&#8221;</p></blockquote><p>Point [a] is the starting point for our investigation which is also published at ICLR 2020 (workshop&#8202;&#8212;&#8202;Bridging AI and Cognitive Science), the last paper we discuss in this paper. (see&nbsp;below)</p><p><strong><a href="https://iclr.cc/virtual/poster_SygcCnNKwr.html">Measuring Compositional Generalization: A Comprehensive Method on Realistic Data</a></strong></p><p>As we saw earlier, the test set for measuring systematic/compositional generalization in the SCAN dataset contains all the combinations of a primitive <em>jump</em> with modifiers, the combinations that never occurred in the trainset. The model needs to understand the concept of <em>jump</em> from input-output pair (jump, JUMP) and the concept of different modifiers from their combinations with other primitives.</p><p>Although, this type of train/test split strategy may not be feasible and even not efficient for measuring compositional generalization in some cases. This is the premise and the problem this work is addressing by formulating a method that automatically creates train/test splits with lower atom divergence and higher compound divergence. See a paragraph from the paper&nbsp;below.</p><blockquote><p>&#8220;We use the term compositionality experiment to mean a particular way of splitting the data into train and test sets with the goal of measuring compositional generalization. Based on the notions of atoms and compounds described above, we say that an ideal compositionality experiment should adhere to the following two principles</p></blockquote><blockquote><p>Similar atom distribution: All atoms present in the test set are also present in the train set, and the distribution of atoms in the train set is as similar as possible to their distribution in the test&nbsp;set.</p></blockquote><blockquote><p>Different compound distribution: The distribution of compounds in the train set is as different as possible from the distribution in the test&nbsp;set.&#8221;</p></blockquote><p>With this method they proposed, they created multiple splits from the SCAN dataset and showed that they are better at evaluating compositional generalization than the original train/test splits.</p><p>Also, they created a dataset for sematic parsing called Compositional Freebase Questions (CFQ) which is the largest dataset for studying compositional generalization in natural language known. I recommend to check out their very nice blog post about this work in google ai&nbsp;blog.</p><p><a href="https://ai.googleblog.com/2020/03/measuring-compositional-generalization.html">Measuring Compositional Generalization</a></p><p><strong><a href="https://iclr.cc/virtual/poster_SylVNerFvr.html">Permutation Equivariant Models for Compositional Generalization in&nbsp;Language</a></strong></p><p>This is a new way of thinking about systematic/compositional generalization with the hypothesis that it could be seen as a form of group-invariance. They designed this group-invariance property right into the Seq2Seq model and solved different tasks in the SCAN&nbsp;dataset.</p><p>The following is the hypothesis they made and showed good evidence of by solving the SCAN dataset with the group-invariance property.</p><blockquote><p>&#8220;Models achieving the compositional generalization required in certain SCAN tasks are equivariant with respect to permutation group operations in the input and output languages.&#8221;</p></blockquote><p><strong><a href="https://baicsworkshop.github.io/program/baics_11.html">Systematic generalization emerges in Seq2Seq models with variability in data</a> (Our&nbsp;Paper)</strong></p><p>The first paper we discussed showed that an RL agent systematically generalizes to a new object if a command like &#8216;lift&#8217; is trained with more objects. i.e. generalizes to &#8216;pick NewObj&#8217; after trained on &#8216;pick Obj1&#8217;, &#8216;pick Obj2&#8217;&#8230; &#8216;pick ObjN&#8217;, where N is good enough number of&nbsp;objects.</p><p>Turns out, just an LSTM+attn learns 6 modifiers in SCAN (out of 8 modifiers &amp; 2 conjunctions) with an increased number of distinct primitives in the dataset and generalizes to commands with new primitives never trained with any modifiers like (&#8216;jump twice&#8217;, &#8216;JUMP JUMP&#8217;) which is never shown&nbsp;before.</p><p>Helping us better understand the characteristics of systematic generalization in deep learning models, we found another behavior that is highly correlated with systematic generalization, instance independent representations of modifiers. I.e. if we subtract &#8216;walk&#8217; from the command &#8216;walk twice&#8217; and add &#8216;jump&#8217;, the model would give the output of &#8216;jump twice&#8217;. Vectors added/subtracted are from the encoder's final hidden state. This behavior and systematic generalization are highly correlated with a Pearson coefficient of&nbsp;0.99.</p><blockquote><p>The more number of distinct primitives a modifier is operated on in the trainset gives rise to systematic generalization and also makes the model represent a modifier independent of any instance of variable it operated on. Instance Independent representations for modifiers. And it is highly correlated with systematic generalization.</p></blockquote><p>We also showed that, with 300 distinct primitives in the dataset, models trained on primitive variables (like &#8216;jump twice&#8217; ) generalized to compound variables &#8216;{jump twice} twice&#8217;, though in a limited way. The approaches like Syntactic attention and Meta-seq2seq learning which solved the SCAN dataset didn&#8217;t show this behavior, more interestingly, even when trained with 300 different primitives.</p><h4>Systematic generalization and the&nbsp;future</h4><p>We are still in the initial stage of understanding the systematicity of human cognition and exploring systematic generalization in deep learning models by evaluating for it and finding the inductive biases that enable&nbsp;it.</p><p>For now, it is clear that systematicity would be a concern in many domains and tasks like language understanding, abstract and analogical reasoning, semantic scene analysis if we aim to build models that interact with the world efficiently the way humans&nbsp;do.</p><p>Finally, check out the ICLR 2020 workshop where our work is published which is a great place to explore all the amazing working at the intersection of cognitive science and&nbsp;AI.</p><p><a href="https://baicsworkshop.github.io/index.html">Bridging AI and Cognitive Science (BAICS)</a></p><p>Feel free to discuss any work related to the systematicity of human cognition and systematic generalization. Follow me here for more updates and recent work at the intersection of cognitive science and deep learning.</p>]]></content:encoded></item><item><title><![CDATA[Computer Vision and Computer Vision applied]]></title><description><![CDATA[Computer Vision is a scientific endeavor that aims to automate the human visual system, not necessarily aims to imitate it but emulate all the ability of the human visual system and beyond.]]></description><link>https://www.aipapertrails.com/p/computer-vision-and-computer-vision-applied-2ef2cc0f7116</link><guid isPermaLink="false">https://www.aipapertrails.com/p/computer-vision-and-computer-vision-applied-2ef2cc0f7116</guid><dc:creator><![CDATA[Prakash Kagitha]]></dc:creator><pubDate>Thu, 24 Oct 2019 09:08:17 GMT</pubDate><enclosure url="https://cdn-images-1.medium.com/max/1024/0*77QZ-aSQ-5FIB39G" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://cdn-images-1.medium.com/max/1024/0*77QZ-aSQ-5FIB39G" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://cdn-images-1.medium.com/max/1024/0*77QZ-aSQ-5FIB39G 424w, https://cdn-images-1.medium.com/max/1024/0*77QZ-aSQ-5FIB39G 848w, https://cdn-images-1.medium.com/max/1024/0*77QZ-aSQ-5FIB39G 1272w, https://cdn-images-1.medium.com/max/1024/0*77QZ-aSQ-5FIB39G 1456w" sizes="100vw"><img src="https://cdn-images-1.medium.com/max/1024/0*77QZ-aSQ-5FIB39G" data-attrs="{&quot;src&quot;:&quot;https://cdn-images-1.medium.com/max/1024/0*77QZ-aSQ-5FIB39G&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://cdn-images-1.medium.com/max/1024/0*77QZ-aSQ-5FIB39G 424w, https://cdn-images-1.medium.com/max/1024/0*77QZ-aSQ-5FIB39G 848w, https://cdn-images-1.medium.com/max/1024/0*77QZ-aSQ-5FIB39G 1272w, https://cdn-images-1.medium.com/max/1024/0*77QZ-aSQ-5FIB39G 1456w" sizes="100vw" fetchpriority="high"></picture><div></div></div></a><figcaption class="image-caption">Photo by <a href="https://unsplash.com/@rohanmakhecha?utm_source=medium&amp;utm_medium=referral">Rohan Makhecha</a> on&nbsp;<a href="https://unsplash.com?utm_source=medium&amp;utm_medium=referral">Unsplash</a></figcaption></figure></div><p>Computer Vision is a scientific endeavor that aims to automate the human visual system, not necessarily aims to imitate it but emulate all the ability of the human visual system and beyond. So, as is the case with any field which has the potential to change the course of humanity, Computer vision has a lot of history made up of the likes of obsessive human effort, hard problems, victories, failures, and immense&nbsp;hope.</p><p>In spite of the complexity of visual scenes in the world around us, Computer vision is now capable of detecting practically any type of object including people, vehicles, and so on. Not only that it can detect objects it can recognize the characteristics of those objects like identity and gestures in the case of humans, deformations/defects in the case of objects, intrusions into boundaries, and much more. This is just a small part of an ocean that is applied Computer&nbsp;Vision.</p><p>Computer vision is the visual system for computers to play very strategic games like Go and Starcraft better than humans [1]. It can pave a way for self-driving cars to take automatic actions that don&#8217;t so that they won&#8217;t run into other cars or people. It seems that Computer vision can do anything as long as we have data and computation. Models learning to classify discriminatively and models that actually learn the whole distribution that generated the dataset both have shown significant progress in solving many real-world problems.</p><p>As a matter of fact, we at <a href="http://akaiketech.com">Akaike Technologies</a>, believe that it can solve a lot of problems as long as we could formulate them. As a deep learning services company, we deployed various applications containing object detection, image segmentation, and machine inspection among others with computer vision. Alongside that, at the frontiers of computer vision, we know what are the current limitations and are ambitious to take the field further towards the collective vision of the community.</p><p>Now, we wish to delve into how computer vision evolved to be so successful. How started as an application in itself became a great scientific endeavor. And also what are its limitations and prospects, inclining towards its application to real-world problems around&nbsp;us.</p><h3>Initial aspirations for Computer Vision (1950s-2012)</h3><p>Computer vision emerged not so later than the field which aimed to make intelligent machines, Artificial Intelligence. In the 1950s and 1960s, there were small computer vision projects in the labs of AI pioneers like Marvin Minsky and Seymour Papert among others. We could see the optimism in 1956 when Seymour Papert wrote a proposal for imparting visual intelligence to machines describing it as a research project our a single summer [2]. Of course, it turned that it has a lot more to&nbsp;it.</p><p>There was great work done in David Marr's lab, who dealt with the problem from a neuroscience perspective [3]. Very prominent people like Marvin Minsky and Patrick Winston among others also worked on visual systems for robots and were betting on the idea that even visual intelligence could be symbolized and then be reasoned about as a top-down approach of visual intelligence [4], which, despite its potential, didn&#8217;t deliver any considerable development in Computer vision considering how much we have been able to do now. Alternatively, the bottom-up approach which is loosely modeled from processing units in the brain, neurons, composed with end-to-end learning mechanisms dominated the field as we discuss in the next&nbsp;section.</p><p>From the 1980s to even 2010, the efficient way to do computer vision, more specifically object detection, scene understanding, or classifying images on different schemas, is to develop mechanisms to extract some broad range of relevant features with expert domain knowledge and apply traditional statistics/machine learning on top of those feature characteristics [5]. These models, even though worked for some cases, needed a lot of domain-specific knowledge and manual engineering, and were so brittle outside their very specific&nbsp;problem.</p><h3>The success of CNNs for Computer&nbsp;Vision</h3><p>In 2012, the paradigm shift happened popularly when Convolutional Neural Networks(CNNs), a variant of neural networks with characteristics suitable for performing very well on images, reduced the error rate on a widely studied object recognition dataset called ImageNet by 50% [6]. From around that point, all or most of the computer vision turned to convolutional neural networks. This is more dramatic than we could imagine. Yan Lecun, the inventor of convolutional neural networks, was told, even though his approach(CNNs) achieved state of the art, that it doesn&#8217;t worth a publication in CVPR (a very respectable conference of Computer Vision) because it didn&#8217;t tell us anything about the visual system. This is in 2010. Now, it&#8217;s hard to find a CVPR paper that doesn&#8217;t contain Convolutional neural networks.</p><p>This very successful approach didn&#8217;t emerge overnight but we couldn&#8217;t say the same for the realization of emergence at all. The notion of creating intelligent machines by connecting brain like units, neurons, in groups is very old. It used to be called Connectionism [7]. The notion of convolutions which turned out efficient for computer vision is also not new. It is often credited to Fukushima coming up with or to formalize the idea related to convolutions in a network of neurons to make the model shift-invariant [8]. But a significant structure to the idea of convolution neural networks came from the work of Yan Lecun, credited as the inventor CNNs. This is in&nbsp;1974.</p><p>As Prof. Lecun&#8217;s paper got rejected even in 2010, imagine the ideology in the 1970s. Many people devoted their life to Neural networks (and CNNs) but they don&#8217;t know any practical ways to train them so that they could perform on par with manually engineered systems or systems based on classical machine learning. Adding to that, there was a symbolic approach to creating intelligent machines with very prominent supporters, as we talked about in the last section, which slowed the progress of connectionism.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://cdn-images-1.medium.com/max/600/1*zh8tEde1Qur9LWYxmGYMPg.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://cdn-images-1.medium.com/max/600/1*zh8tEde1Qur9LWYxmGYMPg.png 424w, https://cdn-images-1.medium.com/max/600/1*zh8tEde1Qur9LWYxmGYMPg.png 848w, https://cdn-images-1.medium.com/max/600/1*zh8tEde1Qur9LWYxmGYMPg.png 1272w, https://cdn-images-1.medium.com/max/600/1*zh8tEde1Qur9LWYxmGYMPg.png 1456w" sizes="100vw"><img src="https://cdn-images-1.medium.com/max/600/1*zh8tEde1Qur9LWYxmGYMPg.png" data-attrs="{&quot;src&quot;:&quot;https://cdn-images-1.medium.com/max/600/1*zh8tEde1Qur9LWYxmGYMPg.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://cdn-images-1.medium.com/max/600/1*zh8tEde1Qur9LWYxmGYMPg.png 424w, https://cdn-images-1.medium.com/max/600/1*zh8tEde1Qur9LWYxmGYMPg.png 848w, https://cdn-images-1.medium.com/max/600/1*zh8tEde1Qur9LWYxmGYMPg.png 1272w, https://cdn-images-1.medium.com/max/600/1*zh8tEde1Qur9LWYxmGYMPg.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a></figure></div><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://cdn-images-1.medium.com/max/572/1*OtW4yIHw7p2srKAv9sFIKA.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://cdn-images-1.medium.com/max/572/1*OtW4yIHw7p2srKAv9sFIKA.png 424w, https://cdn-images-1.medium.com/max/572/1*OtW4yIHw7p2srKAv9sFIKA.png 848w, https://cdn-images-1.medium.com/max/572/1*OtW4yIHw7p2srKAv9sFIKA.png 1272w, https://cdn-images-1.medium.com/max/572/1*OtW4yIHw7p2srKAv9sFIKA.png 1456w" sizes="100vw"><img src="https://cdn-images-1.medium.com/max/572/1*OtW4yIHw7p2srKAv9sFIKA.png" data-attrs="{&quot;src&quot;:&quot;https://cdn-images-1.medium.com/max/572/1*OtW4yIHw7p2srKAv9sFIKA.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://cdn-images-1.medium.com/max/572/1*OtW4yIHw7p2srKAv9sFIKA.png 424w, https://cdn-images-1.medium.com/max/572/1*OtW4yIHw7p2srKAv9sFIKA.png 848w, https://cdn-images-1.medium.com/max/572/1*OtW4yIHw7p2srKAv9sFIKA.png 1272w, https://cdn-images-1.medium.com/max/572/1*OtW4yIHw7p2srKAv9sFIKA.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">The upper part shows how a kernel is applied to an image that turns it into a feature map of compressed size. The lower part of the image describes the sequence of those operations (hence the name deep) to classify which number an image contains. This is a Deep Convolutional Neural Network which conceptually very similar to networks we use today. This work is from&nbsp;1990.</figcaption></figure></div><p>With the effort of several researchers over several decades, the efficacy of connectionism(CNNs) became evident in smaller specific problems. Yan Lecun, in 1990, used convolutional neural networks to recognize digits to automatically understand the zipcode on checks at AT&amp;T [9]. Among many people in the community, Yoshua Bengio and Geoffrey Hinton contributed a significant amount to learning algorithms behind these networks, insights about their inabilities and techniques to overcome them [10][11], to eventually cause the rise of Neural Networks, effectively, to be an answer for everything.</p><p>In 2012, two of the students from Geoffrey Hinton&#8217;s lab applied these techniques on an object detection dataset called Imagenet [6]. This is the most dramatic moment that started the Deep Learning revolution we see around us. The availability of data and the economic feasibility of computation are big factors for the whole field of deep learning to be practical and disruptive.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://cdn-images-1.medium.com/max/831/1*00uIns6ZpfzToAwy-4-zzg.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://cdn-images-1.medium.com/max/831/1*00uIns6ZpfzToAwy-4-zzg.png 424w, https://cdn-images-1.medium.com/max/831/1*00uIns6ZpfzToAwy-4-zzg.png 848w, https://cdn-images-1.medium.com/max/831/1*00uIns6ZpfzToAwy-4-zzg.png 1272w, https://cdn-images-1.medium.com/max/831/1*00uIns6ZpfzToAwy-4-zzg.png 1456w" sizes="100vw"><img src="https://cdn-images-1.medium.com/max/831/1*00uIns6ZpfzToAwy-4-zzg.png" data-attrs="{&quot;src&quot;:&quot;https://cdn-images-1.medium.com/max/831/1*00uIns6ZpfzToAwy-4-zzg.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://cdn-images-1.medium.com/max/831/1*00uIns6ZpfzToAwy-4-zzg.png 424w, https://cdn-images-1.medium.com/max/831/1*00uIns6ZpfzToAwy-4-zzg.png 848w, https://cdn-images-1.medium.com/max/831/1*00uIns6ZpfzToAwy-4-zzg.png 1272w, https://cdn-images-1.medium.com/max/831/1*00uIns6ZpfzToAwy-4-zzg.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">The 2012 paper that caused a 50% decrease in the error rate on the Imagenet dataset. (Left) Some of the examples from the evaluation set. (Right) Rows of nearest neighbors of last hidden layer encodings. The leftmost image is the quey and others are top N results&nbsp;[6].</figcaption></figure></div><p>Astonishingly, the reason why Yan Lecun&#8217;s CVPR paper got rejected in 2010 is the reason for the success of computer vision. Specifically, we can credit the success to one fundamental aspect of the deep learning revolution, i.e models learning everything end-to-end without any of the hand-engineering that was necessary for computer vision till now. With CNNs, models will automatically learn the features either discriminatively, when classifying, or generatively when actually have to generate natural images. We just have to define a task, prepare some data, and the model will learn everything that&#8217;s needed. This is the formula that brought down the revolution of learning machines and presumably a change in the course of humanity.</p><p>As shown below, the model learns to detect simple features and then effectively combines them to detect more complex features in a subsequent layer. In the end, if you have another layer to predict the probabilities of each image being in a particular class, we have a model that can classify images. The same paradigm applies even when you are doing a very complicated classification of whether an image of the object has a defect or not. Also, these models can be extended to mark the boundaries of different objects present in the image to masking objects and then tracking them. <strong>Effectively, everything comes down to automatically learning the features specific to the task end to&nbsp;end.</strong></p><p>Below is the visualization of the filters that learned to detect specific things in each of the layers of Deep Convolutional Neural&nbsp;Network:</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://cdn-images-1.medium.com/max/1024/1*bFsBBcFVXk9cxtfAxa7A-g.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://cdn-images-1.medium.com/max/1024/1*bFsBBcFVXk9cxtfAxa7A-g.png 424w, https://cdn-images-1.medium.com/max/1024/1*bFsBBcFVXk9cxtfAxa7A-g.png 848w, https://cdn-images-1.medium.com/max/1024/1*bFsBBcFVXk9cxtfAxa7A-g.png 1272w, https://cdn-images-1.medium.com/max/1024/1*bFsBBcFVXk9cxtfAxa7A-g.png 1456w" sizes="100vw"><img src="https://cdn-images-1.medium.com/max/1024/1*bFsBBcFVXk9cxtfAxa7A-g.png" data-attrs="{&quot;src&quot;:&quot;https://cdn-images-1.medium.com/max/1024/1*bFsBBcFVXk9cxtfAxa7A-g.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://cdn-images-1.medium.com/max/1024/1*bFsBBcFVXk9cxtfAxa7A-g.png 424w, https://cdn-images-1.medium.com/max/1024/1*bFsBBcFVXk9cxtfAxa7A-g.png 848w, https://cdn-images-1.medium.com/max/1024/1*bFsBBcFVXk9cxtfAxa7A-g.png 1272w, https://cdn-images-1.medium.com/max/1024/1*bFsBBcFVXk9cxtfAxa7A-g.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">Layer 1 learned to detect plane surfaces, edges. Layer 2 weights learned to detect some contours&nbsp;[12].</figcaption></figure></div><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://cdn-images-1.medium.com/max/1024/1*ZbT8Kk8MHr1qOuL-eu6Dyg.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://cdn-images-1.medium.com/max/1024/1*ZbT8Kk8MHr1qOuL-eu6Dyg.png 424w, https://cdn-images-1.medium.com/max/1024/1*ZbT8Kk8MHr1qOuL-eu6Dyg.png 848w, https://cdn-images-1.medium.com/max/1024/1*ZbT8Kk8MHr1qOuL-eu6Dyg.png 1272w, https://cdn-images-1.medium.com/max/1024/1*ZbT8Kk8MHr1qOuL-eu6Dyg.png 1456w" sizes="100vw"><img src="https://cdn-images-1.medium.com/max/1024/1*ZbT8Kk8MHr1qOuL-eu6Dyg.png" data-attrs="{&quot;src&quot;:&quot;https://cdn-images-1.medium.com/max/1024/1*ZbT8Kk8MHr1qOuL-eu6Dyg.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://cdn-images-1.medium.com/max/1024/1*ZbT8Kk8MHr1qOuL-eu6Dyg.png 424w, https://cdn-images-1.medium.com/max/1024/1*ZbT8Kk8MHr1qOuL-eu6Dyg.png 848w, https://cdn-images-1.medium.com/max/1024/1*ZbT8Kk8MHr1qOuL-eu6Dyg.png 1272w, https://cdn-images-1.medium.com/max/1024/1*ZbT8Kk8MHr1qOuL-eu6Dyg.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">Layer 3 learned to detect a combination of contours, sometimes meaningful parts of objects&nbsp;[12].</figcaption></figure></div><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://cdn-images-1.medium.com/max/1024/1*XlMWFuofKzCznL-iMwRbfw.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://cdn-images-1.medium.com/max/1024/1*XlMWFuofKzCznL-iMwRbfw.png 424w, https://cdn-images-1.medium.com/max/1024/1*XlMWFuofKzCznL-iMwRbfw.png 848w, https://cdn-images-1.medium.com/max/1024/1*XlMWFuofKzCznL-iMwRbfw.png 1272w, https://cdn-images-1.medium.com/max/1024/1*XlMWFuofKzCznL-iMwRbfw.png 1456w" sizes="100vw"><img src="https://cdn-images-1.medium.com/max/1024/1*XlMWFuofKzCznL-iMwRbfw.png" data-attrs="{&quot;src&quot;:&quot;https://cdn-images-1.medium.com/max/1024/1*XlMWFuofKzCznL-iMwRbfw.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://cdn-images-1.medium.com/max/1024/1*XlMWFuofKzCznL-iMwRbfw.png 424w, https://cdn-images-1.medium.com/max/1024/1*XlMWFuofKzCznL-iMwRbfw.png 848w, https://cdn-images-1.medium.com/max/1024/1*XlMWFuofKzCznL-iMwRbfw.png 1272w, https://cdn-images-1.medium.com/max/1024/1*XlMWFuofKzCznL-iMwRbfw.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">Layer 4 and 5 learned to detect parts or the whole object of some class&nbsp;[12].</figcaption></figure></div><p>Computer Vision community as a whole bet that this is the way to deal with everything. The formula is to define a task, get some data, and apply Neural Networks(CNNs). It will get the job done. This is what all the data science/machine learning teams in the world do. If you want to detect cracks or defects on the windmills or segment different types of lands from just aerial images, this is the way. If you aim to give personalized health insights or do personalized marketing for Electronic health data, this is the way. If you have to show personalized ads in an optimal way, this is the way. If you have to detect people, vehicles, street signs even km away, this is the way. If you have to detect intrusion into your property or a supermarket, this is the way. If you want to build intelligent systems to assist radiologists, this is the way. In fact, at <a href="http://akaiketech.com">Akaike Technologies</a>, we used computer vision to deal with all of these problems and brought a great impact on efficiency and&nbsp;revenue.</p><p>These are the ways companies are using computer vision. This will give a sense of how broad computer vision is. <a href="https://www.aes.com/">AES</a> is using drones and computer vision to make inspecting energy assets safer and more efficient. <a href="http://lgcns.co.kr/">LG CNS</a> is using Computer Vision to accurately detect defects in various products on the assembly line. <a href="https://shop.nordstrom.com/">Nordstrom</a> is using Computer Vision-based product search to enable shoppers to easily find products simply by taking a photo. <a href="https://www.unilever.com/">Unilever</a> is using Computer Vision to gain new insights on consumer behavior and improve ad campaign effectiveness. <a href="https://www.idexx.com/">IDEXX</a> is using Computer Vision to automatically organize medical imagery and improve the productivity of their radiologists. All of these systems are based on Convolutional Neural Networks. Most businesses could bring a great change in adopting computer&nbsp;vision.</p><p>But don&#8217;t get deceived, it takes a lot of skill to formulate the problem as one that aligns with the strength of current computer vision models, to design a model specific to a problem and to train it, to iterate on it till it is production-ready, and to scale the solutions to impact billions of people. The skill is a sound combination of research and implementation in both adopting current literature into production and at the same time contributing to this scientific endeavor. At <a href="http://akaiketech.com">Akaike Technologies</a>, we have proved to be excellent in solving the hardest of problems. You can see some of our accomplishments below at the&nbsp;end.</p><h3>The prospects of Computer&nbsp;Vision</h3><p>Being computer vision as successful as it is now, we are not anywhere near to the ultimate aim of this research endeavor, i.e to automate the visual intelligence we possess. There is quite a lot of discussion in this space of how we can make sure that current computer vision techniques would acquire comprehensive scene understanding. Apart from the statements where this term is used more liberally, comprehensive scene understanding is in some sense the ultimate fulfillment of computer&nbsp;vision.</p><p>To be intentionally and sensibly critical about the models we have today, simply put, we can say that it doesn&#8217;t have a clue of how the world works and how humans behave/feel. It could detect people in a scene and identify them, but it doesn&#8217;t know if someone is a customer or intruder based on their behavior which a human would recognize in most cases within moments. Ideally, it is intuitive in the sense that if we have data for each thing we can learn everything. Theoretically, this is true but nobody can find data of what happens if people do one thing rather than other (just as a single instance lets say, teasing about throwing something at you as opposed to actually throwing), and nobody can bring enough data of visual behavior that the model can understand how the person is feeling, his aims/goal/motivation.</p><p>Sometimes, it even feels practical to get the data and make machines learn the things described above, but it is just because we singled out a simple instance of so many great skills humans possess. The ability to understand the concept with minimal supervision, to combine different concepts to understand more complex concepts, to have the knowledge of intuitive physics of the world, intuitive psychology of humans, the ability to learn, and so on. Aiming to learn these things in the ways we model data now feels, and rightly argued by so many people,&nbsp;stupid.</p><p>This is evident for the computer vision(deep learning) community at large as well. And the community is striving towards models that can learn from data without many human labels sometimes even unsupervised, models that learn to learn different tasks at the same time, models that could interactively acquire data and then reason about it to make decisions. Also, to solve computer vision completely, things that feel distinct from it, like decision making, attention, and interactive learning are nevertheless paramount in building systems that compete with very efficiently evolved human visual intelligence.</p><p>Along with this, there are a lot of investigations from cognitive science, neuroscience, and philosophy that are flowing into computer vision (deep learning) conferences to equip current models with the abilities that surfaced when thinking about computer vision critically [13], [14]. We, at <a href="http://akaiketech.com">Akaike Technologies</a>, see imparting these abilities into computer vision is one of the important directions in the next five to ten years and actively participate and contribute along with the community.</p><p>Computer vision in itself emerged as an application of learning systems, but now it grows into a significant &#8216;scientific&#8217; endeavor to model visual intelligence. Now there is a different sense to computer vision and computer vision applied. We are happy that the field separated different goals and their approach and moving forward complementing each&nbsp;other.</p><p>Comprehensive scene understanding, at large, can impart potentially huge value to applied computer vision as well. Next-generation intruder detection, disaster management, and behavior profile analytics systems are just some of the low hanging fruits for the community, along with improvement in about everything we do&nbsp;now.</p><p>We believe that to excel at applied computer vision(or applied deep learning), beyond current successes, and to be sustainable in five to ten years, one has to exploit current technology with a skill to discern its strengths and weaknesses and be on the horizon of advancements by being critical and asking questions nobody else dares to&nbsp;ask.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://cdn-images-1.medium.com/max/556/1*3jAez1aNEKhsJ94pzz5vmA.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://cdn-images-1.medium.com/max/556/1*3jAez1aNEKhsJ94pzz5vmA.png 424w, https://cdn-images-1.medium.com/max/556/1*3jAez1aNEKhsJ94pzz5vmA.png 848w, https://cdn-images-1.medium.com/max/556/1*3jAez1aNEKhsJ94pzz5vmA.png 1272w, https://cdn-images-1.medium.com/max/556/1*3jAez1aNEKhsJ94pzz5vmA.png 1456w" sizes="100vw"><img src="https://cdn-images-1.medium.com/max/556/1*3jAez1aNEKhsJ94pzz5vmA.png" data-attrs="{&quot;src&quot;:&quot;https://cdn-images-1.medium.com/max/556/1*3jAez1aNEKhsJ94pzz5vmA.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://cdn-images-1.medium.com/max/556/1*3jAez1aNEKhsJ94pzz5vmA.png 424w, https://cdn-images-1.medium.com/max/556/1*3jAez1aNEKhsJ94pzz5vmA.png 848w, https://cdn-images-1.medium.com/max/556/1*3jAez1aNEKhsJ94pzz5vmA.png 1272w, https://cdn-images-1.medium.com/max/556/1*3jAez1aNEKhsJ94pzz5vmA.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a></figure></div><h4>Akaike Technologies</h4><p>We are a group of rapidly growing deep learning experts with two decades of experience in a wide range of domains. We have a proven track record of solving hard problems with computer vision, NLP, and deep learning. We strive to always be at the edge of deep learning, exploiting, and exploring.</p><h4>References</h4><p>[1] Silver, D., Huang, A., Maddison, C.J., Guez, A., Sifre, L., Van Den Driessche, G., Schrittwieser, J., Antonoglou, I., Panneershelvam, V., Lanctot, M. and Dieleman, S., 2016. Mastering the game of Go with deep neural networks and tree search. <em>nature</em>, <em>529</em>(7587), p.484.</p><p>[2] <a href="https://en.wikipedia.org/wiki/Seymour_Papert">Papert, Seymour</a> (1966&#8211;07&#8211;01). &#8220;The Summer Vision Project&#8221;. <em>MIT AI Memos (1959&#8211;2004)</em>. <a href="https://en.wikipedia.org/wiki/Handle_System">HDL</a>:<a href="https://hdl.handle.net/1721.1%2F6125">1721.1/6125</a>.</p><p>[3] Marr, D., 1982. Vision: A computational investigation into the human representation and processing of visual information.</p><p>[4] Minsky, M., 1990. Logical vs. Analogical or Symbolic vs. Connectionist or Neat vs. Scruffy Artificial Intelligence at MIT. Expanding Frontiers, Patrick H. Winston&nbsp;(Ed.).</p><p>[5] Jiang, X., 2009, August. Feature extraction for image recognition and computer vision. In <em>2009 2nd IEEE International Conference on Computer Science and Information Technology</em> (pp. 1&#8211;15).&nbsp;IEEE.</p><p>[6] Krizhevsky, A., Sutskever, I. and Hinton, G.E., 2012. Imagenet classification with deep convolutional neural networks. In <em>Advances in neural information processing systems</em> (pp. 1097&#8211;1105).</p><p>[7] McClelland, J.L., Rumelhart, D.E., and PDP Research Group, 1986. Parallel distributed processing. <em>Explorations in the Microstructure of Cognition</em>, <em>2</em>, pp.216&#8211;271.</p><p>[8] Fukushima, K., 1980. Neocognitron: A self-organizing neural network model for a mechanism of pattern recognition unaffected by shift in position. <em>Biological Cybernetics</em>, <em>36</em>(4), pp.193&#8211;202.</p><p>[9] Le Cun, Y., Matan, O., Boser, B., Denker, J.S., Henderson, D., Howard, R.E., Hubbard, W., Jackel, L.D. and Baird, H.S., 1990, June. Handwritten zip code recognition with multilayer networks. In <em>Proc. 10th International Conference on Pattern Recognition</em> (Vol. 2, pp.&nbsp;35&#8211;40).</p><p>[10] Glorot, X. and Bengio, Y., 2010, March. Understanding the difficulty of training deep feedforward neural networks. In <em>Proceedings of the thirteenth international conference on artificial intelligence and statistics</em> (pp. 249&#8211;256).</p><p>[11] Rumelhart, D.E., Hinton, G.E., and Williams, R.J., 1985. <em>Learning internal representations by error propagation</em> (No. ICS-8506). California Univ San Diego La Jolla Inst for Cognitive Science.</p><p>[12] Zeiler, M.D., and Fergus, R., 2014, September. Visualizing and understanding convolutional networks. In <em>European conference on computer vision</em> (pp. 818&#8211;833). Springer, Cham.</p><p>[13] Hassabis, D., Kumaran, D., Summerfield, C., and Botvinick, M., 2017. Neuroscience-inspired artificial intelligence. <em>Neuron</em>, <em>95</em>(2), pp.245&#8211;258.</p><p>[14] Lake, B.M., Ullman, T.D., Tenenbaum, J.B. and Gershman, S.J., 2017. Building machines that learn and think like people. <em>Behavioral and brain sciences</em>, <em>40</em>.</p><div><hr></div><p><a href="https://medium.com/akaike-technologies/computer-vision-and-computer-vision-applied-2ef2cc0f7116">Computer Vision and Computer Vision applied</a> was originally published in <a href="https://medium.com/akaike-technologies">Akaike Technologies</a> on Medium, where people are continuing the conversation by highlighting and responding to this story.</p>]]></content:encoded></item><item><title><![CDATA[NIPS/NeurIPS 2018: Best* of the First Two Poster Sessions]]></title><description><![CDATA[NeurIPS is a great conference attracting the state of the art in almost every aspect of machine learning research.]]></description><link>https://www.aipapertrails.com/p/neurips-2018-reading-list-from-tue-poster-sessions-a-b-fce561e56be8</link><guid isPermaLink="false">https://www.aipapertrails.com/p/neurips-2018-reading-list-from-tue-poster-sessions-a-b-fce561e56be8</guid><dc:creator><![CDATA[Prakash Kagitha]]></dc:creator><pubDate>Tue, 04 Dec 2018 13:15:09 GMT</pubDate><enclosure url="https://cdn-images-1.medium.com/max/1024/0*MhcsDC_2hD_xwUrX" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://cdn-images-1.medium.com/max/1024/0*MhcsDC_2hD_xwUrX" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://cdn-images-1.medium.com/max/1024/0*MhcsDC_2hD_xwUrX 424w, https://cdn-images-1.medium.com/max/1024/0*MhcsDC_2hD_xwUrX 848w, https://cdn-images-1.medium.com/max/1024/0*MhcsDC_2hD_xwUrX 1272w, https://cdn-images-1.medium.com/max/1024/0*MhcsDC_2hD_xwUrX 1456w" sizes="100vw"><img src="https://cdn-images-1.medium.com/max/1024/0*MhcsDC_2hD_xwUrX" data-attrs="{&quot;src&quot;:&quot;https://cdn-images-1.medium.com/max/1024/0*MhcsDC_2hD_xwUrX&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://cdn-images-1.medium.com/max/1024/0*MhcsDC_2hD_xwUrX 424w, https://cdn-images-1.medium.com/max/1024/0*MhcsDC_2hD_xwUrX 848w, https://cdn-images-1.medium.com/max/1024/0*MhcsDC_2hD_xwUrX 1272w, https://cdn-images-1.medium.com/max/1024/0*MhcsDC_2hD_xwUrX 1456w" sizes="100vw" fetchpriority="high"></picture><div></div></div></a><figcaption class="image-caption">Photo by <a href="https://unsplash.com/@rohanmakhecha?utm_source=medium&amp;utm_medium=referral">Rohan Makhecha</a> on&nbsp;<a href="https://unsplash.com?utm_source=medium&amp;utm_medium=referral">Unsplash</a></figcaption></figure></div><p>NeurIPS is a great conference attracting the state of the art in almost every aspect of machine learning research.</p><p>There are a few things that a researcher in the field should, for sure, give attention to in a conference. In my perspective, those things would cluster research articles similar to these groups: <strong>1.</strong> <strong>Understanding, 2. Essentials, 3. Progress, 4. Big problems/future</strong></p><p>So, I grouped all the things that I find influential in to these categories. These are posters from the NeurIPS 2018 Tue Poster Sessions A&amp;B with their abstracts.</p><h3>1. Understanding</h3><h4><a href="http://papers.nips.cc/paper/7350-are-gans-created-equal-a-large-scale-study">Are GANs Created Equal? A Large-Scale Study</a></h4><p><em>Retrospective</em></p><p>Despite a very rich research activity leading to numerous interesting GAN algorithms, it is still very hard to assess which algorithm(s) perform better than others. We conduct a neutral, multi-faceted large-scale empirical study on state-of-the art models and evaluation measures. We find that most models can reach similar scores with enough hyperparameter optimization and random restarts. This suggests that improvements can arise from a higher computational budget and tuning more than fundamental algorithmic changes. To overcome some limitations of the current metrics, we also propose several data sets on which precision and recall can be computed. Our experimental results suggest that future GAN research should be based on more systematic and objective evaluation procedures. Finally, we did not find evidence that any of the tested algorithms consistently outperforms the non-saturating GAN introduced in \cite{goodfellow2014generative}.</p><h4><a href="http://papers.nips.cc/paper/8169-an-intriguing-failing-of-convolutional-neural-networks-and-the-coordconv-solution">An intriguing failing of convolutional neural networks and the CoordConv solution</a></h4><p><em>Interesting&nbsp;, About&nbsp;time</em></p><p>We have shown the curious inability of CNNs to model the coordinate transform task, shown a simple fix in the form of the CoordConv layer, and given results that suggest including these layers can boost performance in a wide range of applications. Using CoordConv in a GAN produced less mode collapse as the transform between high-level spatial latents and pixels becomes easier to learn. A Faster R-CNN detection model trained on MNIST detection showed 24% better IOU when using CoordConv, and in the Reinforcement Learning (RL) domain agents playing Atari games benefit significantly from the use of CoordConv layers.</p><h4><a href="http://papers.nips.cc/paper/7519-a-linear-speedup-analysis-of-distributed-deep-learning-with-sparse-and-quantized-communication">A Linear Speedup Analysis of Distributed Deep Learning with Sparse and Quantized Communication</a></h4><p><em>Efficiency</em></p><p>The large communication overhead has imposed a bottleneck on the performance of distributed Stochastic Gradient Descent (SGD) for training deep neural networks. Previous works have demonstrated the potential of using gradient sparsification and quantization to reduce the communication cost. However, there is still a lack of understanding about how sparse and quantized communication affects the convergence rate of the training algorithm. In this paper, we study the convergence rate of distributed SGD for non-convex optimization with two communication reducing strategies: sparse parameter averaging and gradient quantization. We show that <em>O</em>(1/&#8730;<em>MK</em>) convergence rate can be achieved if the sparsification and quantization hyperparameters are configured properly. We also propose a strategy called periodic quantized averaging (PQASGD) that further reduces the communication cost while preserving the <em>O</em>(1/&#8730;<em>MK</em>) convergence rate. Our evaluation validates our theoretical results and shows that our PQASGD can converge as fast as full-communication SGD with only 3%&#8722;5% communication data&nbsp;size.</p><h4><a href="http://papers.nips.cc/paper/7368-on-the-dimensionality-of-word-embedding">On the Dimensionality of Word Embedding</a></h4><p><em>Elegent</em></p><p>In this paper, we provide a theoretical understanding of word embedding and its dimensionality. Motivated by the unitary-invariance of word embedding, we propose the Pairwise Inner Product (PIP) loss, a novel metric on the dissimilarity between word embeddings. Using techniques from matrix perturbation theory, we reveal a fundamental bias-variance trade-off in dimensionality selection for word embeddings. This bias-variance trade-off sheds light on many empirical observations which were previously unexplained, for example the existence of an optimal dimensionality. Moreover, new insights and discoveries, like when and how word embeddings are robust to over-fitting, are revealed. By optimizing over the bias-variance trade-off of the PIP loss, we can explicitly answer the open question of dimensionality selection for word embedding.</p><h4><a href="http://papers.nips.cc/paper/7647-adversarial-examples-that-fool-both-computer-vision-and-time-limited-humans">Adversarial Examples that Fool both Computer Vision and Time-Limited Humans</a></h4><p><em>Fundamental</em></p><p>Machine learning models are vulnerable to adversarial examples: small changes to images can cause computer vision models to make mistakes such as identifying a school bus as an ostrich. However, it is still an open question whether humans are prone to similar mistakes. Here, we address this question by leveraging recent techniques that transfer adversarial examples from computer vision models with known parameters and architecture to other models with unknown parameters and architecture, and by matching the initial processing of the human visual system. We find that adversarial examples that strongly transfer across computer vision models influence the classifications made by time-limited human observers.</p><h4><a href="http://papers.nips.cc/paper/8089-dendritic-cortical-microcircuits-approximate-the-backpropagation-algorithm">Dendritic cortical microcircuits approximate the backpropagation algorithm</a></h4><p><em>Insight</em></p><p>Deep learning has seen remarkable developments over the last years, many of them inspired by neuroscience. However, the main learning mechanism behind these advances&#8202;&#8212;&#8202;error backpropagation&#8202;&#8212;&#8202;appears to be at odds with neurobiology. Here, we introduce a multilayer neuronal network model with simplified dendritic compartments in which error-driven synaptic plasticity adapts the network towards a global desired output. In contrast to previous work our model does not require separate phases and synaptic learning is driven by local dendritic prediction errors continuously in time. Such errors originate at apical dendrites and occur due to a mismatch between predictive input from lateral interneurons and activity from actual top-down feedback. Through the use of simple dendritic compartments and different cell-types our model can represent both error and normal activity within a pyramidal neuron. We demonstrate the learning capabilities of the model in regression and classification tasks, and show analytically that it approximates the error backpropagation algorithm. Moreover, our framework is consistent with recent observations of learning between brain areas and the architecture of cortical microcircuits. Overall, we introduce a novel view of learning on dendritic cortical circuits and on how the brain may solve the long-standing synaptic credit assignment problem.</p><h4><a href="http://papers.nips.cc/paper/7999-on-neuronal-capacity">On Neuronal&nbsp;Capacity</a></h4><p><em>Novel formulation</em></p><p>We define the capacity of a learning machine to be the logarithm of the number (or volume) of the functions it can implement. We review known results, and derive new results, estimating the capacity of several neuronal models: linear and polynomial threshold gates, linear and polynomial threshold gates with constrained weights (binary weights, positive weights), and ReLU neurons. We also derive capacity estimates and bounds for fully recurrent networks and layered feedforward networks.</p><h4><a href="http://papers.nips.cc/paper/8277-bias-and-generalization-in-deep-generative-models-an-empirical-study">Bias and Generalization in Deep Generative Models: An Empirical Study</a></h4><p><em>True understanding</em></p><p>n high dimensional settings, density estimation algorithms rely crucially on their inductive bias. Despite recent empirical success, the inductive bias of deep generative models is not well understood. In this paper we propose a framework to systematically investigate bias and generalization in deep generative models of images by probing the learning algorithm with carefully designed training datasets. By measuring properties of the learned distribution, we are able to find interesting patterns of generalization. We verify that these patterns are consistent across datasets, common models and architectures.</p><h4><a href="http://papers.nips.cc/paper/7515-how-does-batch-normalization-help-optimization">How Does Batch Normalization Help Optimization?</a></h4><p><em>Perspective</em></p><p>Batch Normalization (BatchNorm) is a widely adopted technique that enables faster and more stable training of deep neural networks (DNNs). Despite its pervasiveness, the exact reasons for BatchNorm&#8217;s effectiveness are still poorly understood. The popular belief is that this effectiveness stems from controlling the change of the layers&#8217; input distributions during training to reduce the so-called &#8220;internal covariate shift&#8221;. In this work, we demonstrate that such distributional stability of layer inputs has little to do with the success of BatchNorm. Instead, we uncover a more fundamental impact of BatchNorm on the training process: it makes the optimization landscape significantly smoother. This smoothness induces a more predictive and stable behavior of the gradients, allowing for faster training.</p><h4><a href="http://papers.nips.cc/paper/7938-a-probabilistic-population-code-based-on-neural-samples">A probabilistic population code based on neural&nbsp;samples</a></h4><p><em>Neural Encoding</em></p><p>Sensory processing is often characterized as implementing probabilistic inference: networks of neurons compute posterior beliefs over unobserved causes given the sensory inputs. How these beliefs are computed and represented by neural responses is much-debated (Fiser et al. 2010, Pouget et al. 2013). A central debate concerns the question of whether neural responses represent samples of latent variables (Hoyer &amp; Hyvarinnen 2003) or parameters of their distributions (Ma et al. 2006) with efforts being made to distinguish between them (Grabska-Barwinska et al. 2013). A separate debate addresses the question of whether neural responses are proportionally related to the encoded probabilities (Barlow 1969), or proportional to the logarithm of those probabilities (Jazayeri &amp; Movshon 2006, Ma et al. 2006, Beck et al. 2012). Here, we show that these alternatives&#8202;&#8212;&#8202;contrary to common assumptions&#8202;&#8212;&#8202;are not mutually exclusive and that the very same system can be compatible with all of them. As a central analytical result, we show that modeling neural responses in area V1 as samples from a posterior distribution over latents in a linear Gaussian model of the image implies that those neural responses form a linear Probabilistic Population Code (PPC, Ma et al. 2006). In particular, the posterior distribution over some experimenter-defined variable like &#8220;orientation&#8221; is part of the exponential family with sufficient statistics that are linear in the neural sampling-based firing&nbsp;rates.</p><h3>2. Essentials</h3><h4><a href="http://papers.nips.cc/paper/7957-training-deep-models-faster-with-robust-approximate-importance-sampling">Training Deep Models Faster with Robust, Approximate Importance Sampling</a></h4><p><em>Solution</em></p><p>In theory, importance sampling speeds up stochastic gradient algorithms for supervised learning by prioritizing training examples. In practice, the cost of computing importances greatly limits the impact of importance sampling. We propose a robust, approximate importance sampling procedure (RAIS) for stochastic gradient de- scent. By approximating the ideal sampling distribution using robust optimization, RAIS provides much of the benefit of exact importance sampling with drastically reduced overhead. Empirically, we find RAIS-SGD and standard SGD follow similar learning curves, but RAIS moves faster through these paths, achieving speed-ups of at least 20% and sometimes much&nbsp;more.</p><h4><a href="http://papers.nips.cc/paper/7371-dropmax-adaptive-variational-softmax">DropMax: Adaptive Variational Softmax</a></h4><p><em>Elegant</em></p><p>We propose DropMax, a stochastic version of softmax classifier which at each iteration drops non-target classes according to dropout probabilities adaptively decided for each instance. Specifically, we overlay binary masking variables over class output probabilities, which are input-adaptively learned via variational inference. This stochastic regularization has an effect of building an ensemble classifier out of exponentially many classifiers with different decision boundaries. Moreover, the learning of dropout rates for non-target classes on each instance allows the classifier to focus more on classification against the most confusing classes. We validate our model on multiple public datasets for classification, on which it obtains significantly improved accuracy over the regular softmax classifier and other baselines. Further analysis of the learned dropout probabilities shows that our model indeed selects confusing classes more often when it performs classification.</p><h4><a href="http://papers.nips.cc/paper/7338-how-to-start-training-the-effect-of-initialization-and-architecture">How to Start Training: The Effect of Initialization and Architecture</a></h4><p><em>Utility</em></p><p>We identify and study two common failure modes for early training in deep ReLU nets. For each, we give a rigorous proof of when it occurs and how to avoid it, for fully connected, convolutional, and residual architectures. We show that the first failure mode, exploding or vanishing mean activation length, can be avoided by initializing weights from a symmetric distribution with variance 2/fan-in and, for ResNets, by correctly scaling the residual modules. We prove that the second failure mode, exponentially large variance of activation length, never occurs in residual nets once the first failure mode is avoided. In contrast, for fully connected nets, we prove that this failure mode can happen and is avoided by keeping constant the sum of the reciprocals of layer widths. We demonstrate empirically the effectiveness of our theoretical results in predicting when networks are able to start training. In particular, we note that many popular initializations fail our criteria, whereas correct initialization and architecture allows much deeper networks to be&nbsp;trained.</p><h4><a href="http://papers.nips.cc/paper/7481-regularizing-by-the-variance-of-the-activations-sample-variances">Regularizing by the Variance of the Activations&#8217; Sample-Variances</a></h4><p><em>Simple-not-trivial</em></p><p>Normalization techniques play an important role in supporting efficient and often more effective training of deep neural networks. While conventional methods explicitly normalize the activations, we suggest to add a loss term instead. This new loss term encourages the variance of the activations to be stable and not vary from one random mini-batch to the next. Finally, we are able to link the new regularization term to the batchnorm method, which provides it with a regularization perspective. Our experiments demonstrate an improvement in accuracy over the batchnorm technique for both CNNs and fully connected networks.</p><h4><a href="http://papers.nips.cc/paper/8242-mesh-tensorflow-deep-learning-for-supercomputers">Mesh-TensorFlow: Deep Learning for Supercomputers</a></h4><p><em>Solution</em></p><p>Batch-splitting (data-parallelism) is the dominant distributed Deep Neural Network (DNN) training strategy, due to its universal applicability and its amenability to Single-Program-Multiple-Data (SPMD) programming. However, batch-splitting suffers from problems including the inability to train very large models (due to memory constraints), high latency, and inefficiency at small batch sizes. All of these can be solved by more general distribution strategies (model-parallelism). Unfortunately, efficient model-parallel algorithms tend to be complicated to discover, describe, and to implement, particularly on large clusters. We introduce Mesh-TensorFlow, a language for specifying a general class of distributed tensor computations. Where data-parallelism can be viewed as splitting tensors and operations along the &#8220;batch&#8221; dimension, in Mesh-TensorFlow, the user can specify any tensor-dimensions to be split across any dimensions of a multi-dimensional mesh of processors. A Mesh-TensorFlow graph compiles into a SPMD program consisting of parallel operations coupled with collective communication primitives such as Allreduce. We use Mesh-TensorFlow to implement an efficient data-parallel, model-parallel version of the Transformer sequence-to-sequence model. Using TPU meshes of up to 512 cores, we train Transformer models with up to 5 billion parameters, surpassing SOTA results on WMT&#8217;14 English-to-French translation task and the one-billion-word Language modeling benchmark. Mesh-Tensorflow is available at <a href="https://github.com/tensorflow/mesh">https://github.com/tensorflow/mesh</a></p><h4><a href="http://papers.nips.cc/paper/7287-structure-aware-convolutional-neural-networks">Structure-Aware Convolutional Neural&nbsp;Networks</a></h4><p><em>Real world</em></p><p>Convolutional neural networks (CNNs) are inherently subject to invariable filters that can only aggregate local inputs with the same topological structures. It causes that CNNs are allowed to manage data with Euclidean or grid-like structures (e.g., images), not ones with non-Euclidean or graph structures (e.g., traffic networks). To broaden the reach of CNNs, we develop structure-aware convolution to eliminate the invariance, yielding a unified mechanism of dealing with both Euclidean and non-Euclidean structured data. Technically, filters in the structure-aware convolution are generalized to univariate functions, which are capable of aggregating local inputs with diverse topological structures. Since infinite parameters are required to determine a univariate function, we parameterize these filters with numbered learnable parameters in the context of the function approximation theory. By replacing the classical convolution in CNNs with the structure-aware convolution, Structure-Aware Convolutional Neural Networks (SACNNs) are readily established. Extensive experiments on eleven datasets strongly evidence that SACNNs outperform current models on various machine learning tasks, including image classification and clustering, text categorization, skeleton-based action recognition, molecular activity detection, and taxi flow prediction.</p><h4><a href="http://papers.nips.cc/paper/7875-visualizing-the-loss-landscape-of-neural-nets">Visualizing the Loss Landscape of Neural&nbsp;Nets</a></h4><p><em>Perspective</em></p><p>Neural network training relies on our ability to find &#8220;good&#8221; minimizers of highly non-convex loss functions. It is well known that certain network architecture designs (e.g., skip connections) produce loss functions that train easier, and well-chosen training parameters (batch size, learning rate, optimizer) produce minimizers that generalize better. However, the reasons for these differences, and their effect on the underlying loss landscape, is not well understood. In this paper, we explore the structure of neural loss functions, and the effect of loss landscapes on generalization, using a range of visualization methods. First, we introduce a simple &#8220;filter normalization&#8221; method that helps us visualize loss function curvature, and make meaningful side-by-side comparisons between loss functions. Then, using a variety of visualizations, we explore how network architecture affects the loss landscape, and how training parameters affect the shape of minimizers.</p><h4><a href="http://papers.nips.cc/paper/7327-training-dnns-with-hybrid-block-floating-point">Training DNNs with Hybrid Block Floating&nbsp;Point</a></h4><p><em>Craving to&nbsp;solve</em></p><p>The wide adoption of DNNs has given birth to unrelenting computing requirements, forcing datacenter operators to adopt domain-specific accelerators to train them. These accelerators typically employ densely packed full-precision floating-point arithmetic to maximize performance per area. Ongoing research efforts seek to further increase that performance density by replacing floating-point with fixed-point arithmetic. However, a significant roadblock for these attempts has been fixed point&#8217;s narrow dynamic range, which is insufficient for DNN training convergence. We identify block floating point (BFP) as a promising alternative representation since it exhibits wide dynamic range and enables the majority of DNN operations to be performed with fixed-point logic. Unfortunately, BFP alone introduces several limitations that preclude its direct applicability. In this work, we introduce HBFP, a hybrid BFP-FP approach, which performs all dot products in BFP and other operations in floating point. HBFP delivers the best of both worlds: the high accuracy of floating point at the superior hardware density of fixed point. For a wide variety of models, we show that HBFP matches floating point&#8217;s accuracy while enabling hardware implementations that deliver up to 8.5x higher throughput.</p><h4><a href="http://papers.nips.cc/paper/7408-frage-frequency-agnostic-word-representation">FRAGE: Frequency-Agnostic Word Representation</a></h4><p><em>Interesting</em></p><p>Continuous word representation (aka word embedding) is a basic building block in many neural network-based models used in natural language processing tasks. Although it is widely accepted that words with similar semantics should be close to each other in the embedding space, we find that word embeddings learned in several tasks are biased towards word frequency: the embeddings of high-frequency and low-frequency words lie in different subregions of the embedding space, and the embedding of a rare word and a popular word can be far from each other even if they are semantically similar. This makes learned word embeddings ineffective, especially for rare words, and consequently limits the performance of these neural network models. In order to mitigate the issue, in this paper, we propose a neat, simple yet effective adversarial training method to blur the boundary between the embeddings of high-frequency words and low-frequency words. We conducted comprehensive studies on ten datasets across four natural language processing tasks, including word similarity, language modeling, machine translation and text classification. Results show that we achieve higher performance than the baselines in all&nbsp;tasks.</p><h4><a href="http://papers.nips.cc/paper/7965-unsupervised-cross-modal-alignment-of-speech-and-text-embedding-spaces">Unsupervised Cross-Modal Alignment of Speech and Text Embedding Spaces</a></h4><p><em>At the&nbsp;core</em></p><p>Recent research has shown that word embedding spaces learned from text corpora of different languages can be aligned without any parallel data supervision. Inspired by the success in unsupervised cross-lingual word embeddings, in this paper we target learning a cross-modal alignment between the embedding spaces of speech and text learned from corpora of their respective modalities in an unsupervised fashion. The proposed framework learns the individual speech and text embedding spaces, and attempts to align the two spaces via adversarial training, followed by a refinement procedure. We show how our framework could be used to perform the tasks of spoken word classification and translation, and the experimental results on these two tasks demonstrate that the performance of our unsupervised alignment approach is comparable to its supervised counterpart. Our framework is especially useful for developing automatic speech recognition (ASR) and speech-to-text translation systems for low- or zero-resource languages, which have little parallel audio-text data for training modern supervised ASR and speech-to-text translation models, but account for the majority of the languages spoken across the&nbsp;world.</p><h4><a href="http://papers.nips.cc/paper/7886-compact-generalized-non-local-network">Compact Generalized Non-local Network</a></h4><p><em>Clean</em></p><p>The non-local module is designed for capturing long-range spatio-temporal dependencies in images and videos. Although having shown excellent performance, it lacks the mechanism to model the interactions between positions across channels, which are of vital importance in recognizing fine-grained objects and actions. To address this limitation, we generalize the non-local module and take the correlations between the positions of any two channels into account. This extension utilizes the compact representation for multiple kernel functions with Taylor expansion that makes the generalized non-local module in a fast and low-complexity computation flow. Moreover, we implement our generalized non-local method within channel groups to ease the optimization. Experimental results illustrate the clear-cut improvements and practical applicability of the generalized non-local module on both fine-grained object recognition and video classification. Code is available at: <a href="https://github.com/KaiyuYue/cgnl-network.pytorch">https://github.com/KaiyuYue/cgnl-network.pytorch</a>.</p><h3>3. Progress</h3><h4><a href="http://papers.nips.cc/paper/8226-generalizing-point-embeddings-using-the-wasserstein-space-of-elliptical-distributions">Generalizing Point Embeddings using the Wasserstein Space of Elliptical Distributions</a></h4><p><em>Stretching</em></p><p>A novel framework for embeddings which are are numerical flexible and which extend the point embeddings, elliptical embeddings in wessserstein space. Wasserstein elliptical embeddings are more intuitive and yield tools that are better behaved numerically than the alternative choice of Gaussian embeddings with the Kullback-Leibler divergence. The paper demonstrates the advantages of elliptical embeddings by using them for visualization, to compute embeddings of words, and to reflect entailment or hypernymy.</p><h4><a href="http://papers.nips.cc/paper/7356-fishnet-a-versatile-backbone-for-image-region-and-pixel-level-prediction">FishNet: A Versatile Backbone for Image, Region, and Pixel Level Prediction</a></h4><p><em>Fundamentals</em></p><p>The basic principles in designing convolutional neural network (CNN) structures for predicting objects on different levels, e.g., image-level, region-level, and pixel-level, are diverging. Generally, network structures designed specifically for image classification are directly used as default backbone structure for other tasks including detection and segmentation, but there is seldom backbone structure designed under the consideration of unifying the advantages of networks designed for pixel-level or region-level predicting tasks, which may require very deep features with high resolution. Towards this goal, we design a fish-like network, called FishNet. In FishNet, the information of all resolutions is preserved and refined for the final task. Besides, we observe that existing works still cannot \emph{directly} propagate the gradient information from deep layers to shallow layers. Our design can better handle this problem. Extensive experiments have been conducted to demonstrate the remarkable performance of the FishNet. In particular, on ImageNet-1k, the accuracy of FishNet is able to surpass the performance of DenseNet and ResNet with fewer parameters. FishNet was applied as one of the modules in the winning entry of the COCO Detection 2018 challenge. The code is available at <a href="https://github.com/kevin-ssy/FishNet">https://github.com/kevin-ssy/FishNet</a>.</p><h4><a href="http://papers.nips.cc/paper/8003-towards-robust-interpretability-with-self-explaining-neural-networks">Towards Robust Interpretability with Self-Explaining Neural&nbsp;Networks</a></h4><p><em>With out side&nbsp;effects</em></p><p>Most recent work on interpretability of complex machine learning models has focused on estimating a-posteriori explanations for previously trained models around specific predictions. Self-explaining models where interpretability plays a key role already during learning have received much less attention. We propose three desiderata for explanations in general&#8202;&#8212;&#8202;explicitness, faithfulness, and stability&#8202;&#8212;&#8202;and show that existing methods do not satisfy them. In response, we design self-explaining models in stages, progressively generalizing linear classifiers to complex yet architecturally explicit models. Faithfulness and stability are enforced via regularization specifically tailored to such models. Experimental results across various benchmark datasets show that our framework offers a promising direction for reconciling model complexity and interpretability.</p><h4><a href="http://papers.nips.cc/paper/7960-relational-recurrent-neural-networks">Relational recurrent neural&nbsp;networks</a></h4><p><em>Revolutionary</em></p><p>Memory-based neural networks model temporal data by leveraging an ability to remember information for long periods. It is unclear, however, whether they also have an ability to perform complex relational reasoning with the information they remember. Here, we first confirm our intuitions that standard memory architectures may struggle at tasks that heavily involve an understanding of the ways in which entities are connected&#8202;&#8212;&#8202;i.e., tasks involving relational reasoning. We then improve upon these deficits by using a new memory module&#8202;&#8212;&#8202;a Relational Memory Core (RMC)&#8202;&#8212;&#8202;which employs multi-head dot product attention to allow memories to interact. Finally, we test the RMC on a suite of tasks that may profit from more capable relational reasoning across sequential information, and show large gains in RL domains (BoxWorld &amp; Mini PacMan), program evaluation, and language modeling, achieving state-of-the-art results on the WikiText-103, Project Gutenberg, and GigaWord datasets.</p><h4><a href="http://papers.nips.cc/paper/7938-a-probabilistic-population-code-based-on-neural-samples">A probabilistic population code based on neural&nbsp;samples</a></h4><p><em>Seek the&nbsp;might</em></p><p>Sensory processing is often characterized as implementing probabilistic inference: networks of neurons compute posterior beliefs over unobserved causes given the sensory inputs. How these beliefs are computed and represented by neural responses is much-debated (Fiser et al. 2010, Pouget et al. 2013). A central debate concerns the question of whether neural responses represent samples of latent variables (Hoyer &amp; Hyvarinnen 2003) or parameters of their distributions (Ma et al. 2006) with efforts being made to distinguish between them (Grabska-Barwinska et al. 2013). A separate debate addresses the question of whether neural responses are proportionally related to the encoded probabilities (Barlow 1969), or proportional to the logarithm of those probabilities (Jazayeri &amp; Movshon 2006, Ma et al. 2006, Beck et al. 2012). Here, we show that these alternatives&#8202;&#8212;&#8202;contrary to common assumptions&#8202;&#8212;&#8202;are not mutually exclusive and that the very same system can be compatible with all of them. As a central analytical result, we show that modeling neural responses in area V1 as samples from a posterior distribution over latents in a linear Gaussian model of the image implies that those neural responses form a linear Probabilistic Population Code (PPC, Ma et al. 2006). In particular, the posterior distribution over some experimenter-defined variable like &#8220;orientation&#8221; is part of the exponential family with sufficient statistics that are linear in the neural sampling-based firing&nbsp;rates.</p><h4><a href="http://papers.nips.cc/paper/7381-neural-symbolic-vqa-disentangling-reasoning-from-vision-and-language-understanding">Neural-Symbolic VQA: Disentangling Reasoning from Vision and Language Understanding</a></h4><p><em>Encompassing everything</em></p><p>We marry two powerful ideas: deep representation learning for visual recognition and language understanding, and symbolic program execution for reasoning. Our neural-symbolic visual question answering (NS-VQA) system first recovers a structural scene representation from the image and a program trace from the question. It then executes the program on the scene representation to obtain an answer. Incorporating symbolic structure as prior knowledge offers three unique advantages. First, executing programs on a symbolic space is more robust to long program traces; our model can solve complex reasoning tasks better, achieving an accuracy of 99.8% on the CLEVR dataset. Second, the model is more data- and memory-efficient: it performs well after learning on a small number of training data; it can also encode an image into a compact representation, requiring less storage than existing methods for offline question answering. Third, symbolic program execution offers full transparency to the reasoning process; we are thus able to interpret and diagnose each execution step.</p><h4><a href="http://papers.nips.cc/paper/8206-neural-voice-cloning-with-a-few-samples">Neural Voice Cloning with a Few&nbsp;Samples</a></h4><p><em>Artificial need</em></p><p>Voice cloning is a highly desired feature for personalized speech interfaces. We introduce a neural voice cloning system that learns to synthesize a person&#8217;s voice from only a few audio samples. We study two approaches: speaker adaptation and speaker encoding. Speaker adaptation is based on fine-tuning a multi-speaker generative model. Speaker encoding is based on training a separate model to directly infer a new speaker embedding, which will be applied to a multi-speaker generative model. In terms of naturalness of the speech and similarity to the original speaker, both approaches can achieve good performance, even with a few cloning audios. While speaker adaptation can achieve slightly better naturalness and similarity, cloning time and required memory for the speaker encoding approach are significantly less, making it more favorable for low-resource deployment.</p><h4><a href="http://papers.nips.cc/paper/8027-neural-arithmetic-logic-units">Neural Arithmetic Logic&nbsp;Units</a></h4><p><em>Being normal</em></p><p>Neural networks can learn to represent and manipulate numerical information, but they seldom generalize well outside of the range of numerical values encountered during training. To encourage more systematic numerical extrapolation, we propose an architecture that represents numerical quantities as linear activations which are manipulated using primitive arithmetic operators, controlled by learned gates. We call this module a neural arithmetic logic unit (NALU), by analogy to the arithmetic logic unit in traditional processors. Experiments show that NALU-enhanced neural networks can learn to track time, perform arithmetic over images of numbers, translate numerical language into real-valued scalars, execute computer code, and count objects in images. In contrast to conventional architectures, we obtain substantially better generalization both inside and outside of the range of numerical values encountered during training, often extrapolating orders of magnitude beyond trained numerical ranges.</p><h3>4. Big Problems/Future</h3><h4><a href="http://papers.nips.cc/paper/7473-embedding-logical-queries-on-knowledge-graphs">Embedding Logical Queries on Knowledge Graphs</a></h4><p><em>Minority approach</em></p><p>Learning low-dimensional embeddings of knowledge graphs is a powerful approach used to predict unobserved or missing edges between entities. However, an open challenge in this area is developing techniques that can go beyond simple edge prediction and handle more complex logical queries, which might involve multiple unobserved edges, entities, and variables. For instance, given an incomplete biological knowledge graph, we might want to predict &#8220;em what drugs are likely to target proteins involved with both diseases X and Y?&#8221;&#8202;&#8212;&#8202;a query that requires reasoning about all possible proteins that might interact with diseases X and Y. Here we introduce a framework to efficiently make predictions about conjunctive logical queries&#8202;&#8212;&#8202;a flexible but tractable subset of first-order logic&#8202;&#8212;&#8202;on incomplete knowledge graphs. In our approach, we embed graph nodes in a low-dimensional space and represent logical operators as learned geometric operations (e.g., translation, rotation) in this embedding space. By performing logical operations within a low-dimensional embedding space, our approach achieves a time complexity that is linear in the number of query variables, compared to the exponential complexity required by a naive enumeration-based approach. We demonstrate the utility of this framework in two application studies on real-world datasets with millions of relations: predicting logical relationships in a network of drug-gene-disease interactions and in a graph-based representation of social interactions derived from a popular web&nbsp;forum.</p><h4><a href="http://papers.nips.cc/paper/7334-multi-task-learning-as-multi-objective-optimization">Multi-Task Learning as Multi-Objective Optimization</a></h4><p><em>The whole&nbsp;deal</em></p><p>In multi-task learning, multiple tasks are solved jointly, sharing inductive bias between them. Multi-task learning is inherently a multi-objective problem because different tasks may conflict, necessitating a trade-off. A common compromise is to optimize a proxy objective that minimizes a weighted linear combination of per-task losses. However, this workaround is only valid when the tasks do not compete, which is rarely the case. In this paper, we explicitly cast multi-task learning as multi-objective optimization, with the overall objective of finding a Pareto optimal solution. To this end, we use algorithms developed in the gradient-based multi-objective optimization literature. These algorithms are not directly applicable to large-scale learning problems since they scale poorly with the dimensionality of the gradients and the number of tasks. We therefore propose an upper bound for the multi-objective loss and show that it can be optimized efficiently. We further prove that optimizing this upper bound yields a Pareto optimal solution under realistic assumptions. We apply our method to a variety of multi-task deep learning problems including digit classification, scene understanding (joint semantic segmentation, instance segmentation, and depth estimation), and multi-label classification. Our method produces higher-performing models than recent multi-task learning formulations or per-task training.</p><h4><a href="http://papers.nips.cc/paper/8014-rendernet-a-deep-convolutional-network-for-differentiable-rendering-from-3d-shapes">RenderNet: A deep convolutional network for differentiable rendering from 3D&nbsp;shapes</a></h4><p><em>Imminent</em></p><p>Traditional computer graphics rendering pipelines are designed for procedurally generating 2D images from 3D shapes with high performance. The nondifferentiability due to discrete operations (such as visibility computation) makes it hard to explicitly correlate rendering parameters and the resulting image, posing a significant challenge for inverse rendering tasks. Recent work on differentiable rendering achieves differentiability either by designing surrogate gradients for non-differentiable operations or via an approximate but differentiable renderer. These methods, however, are still limited when it comes to handling occlusion, and restricted to particular rendering effects. We present RenderNet, a differentiable rendering convolutional network with a novel projection unit that can render 2D images from 3D shapes. Spatial occlusion and shading calculation are automatically encoded in the network. Our experiments show that RenderNet can successfully learn to implement different shaders, and can be used in inverse rendering tasks to estimate shape, pose, lighting and texture from a single&nbsp;image.</p><h4><a href="http://papers.nips.cc/paper/8163-e-snli-natural-language-inference-with-natural-language-explanations">e-SNLI: Natural Language Inference with Natural Language Explanations</a></h4><p><em>Big dreams</em></p><p>In order for machine learning to garner widespread public adoption, models must be able to provide interpretable and robust explanations for their decisions, as well as learn from human-provided explanations at train time. In this work, we extend the Stanford Natural Language Inference dataset with an additional layer of human-annotated natural language explanations of the entailment relations. We further implement models that incorporate these explanations into their training process and output them at test time. We show how our corpus of explanations, which we call e-SNLI, can be used for various goals, such as obtaining full sentence justifications of a model&#8217;s decisions, improving universal sentence representations and transferring to out-of-domain NLI datasets. Our dataset thus opens up a range of research directions for using natural language explanations, both for improving models and for asserting their&nbsp;trust.</p><h4><a href="http://papers.nips.cc/paper/7592-speaker-follower-models-for-vision-and-language-navigation">Speaker-Follower Models for Vision-and-Language Navigation</a></h4><p><em>Future</em></p><p>Navigation guided by natural language instructions presents a challenging reasoning problem for instruction followers. Natural language instructions typically identify only a few high-level decisions and landmarks rather than complete low-level motor behaviors; much of the missing information must be inferred based on perceptual context. In machine learning settings, this is doubly challenging: it is difficult to collect enough annotated data to enable learning of this reasoning process from scratch, and also difficult to implement the reasoning process using generic sequence models. Here we describe an approach to vision-and-language navigation that addresses both these issues with an embedded speaker model. We use this speaker model to (1) synthesize new instructions for data augmentation and to (2) implement pragmatic reasoning, which evaluates how well candidate action sequences explain an instruction. Both steps are supported by a panoramic action space that reflects the granularity of human-generated instructions. Experiments show that all three components of this approach&#8202;&#8212;&#8202;-speaker-driven data augmentation, pragmatic reasoning and panoramic action space&#8202;&#8212;&#8202;-dramatically improve the performance of a baseline instruction follower, more than doubling the success rate over the best existing approach on a standard benchmark.</p><h4><a href="http://papers.nips.cc/paper/7617-neural-code-comprehension-a-learnable-representation-of-code-semantics">Neural Code Comprehension: A Learnable Representation of Code Semantics</a></h4><p><em>The alchemy</em></p><p>With the recent success of embeddings in natural language processing, research has been conducted into applying similar methods to code analysis. Most works attempt to process the code directly or use a syntactic tree representation, treating it like sentences written in a natural language. However, none of the existing methods are sufficient to comprehend program semantics robustly, due to structural features such as function calls, branching, and interchangeable order of statements. In this paper, we propose a novel processing technique to learn code semantics, and apply it to a variety of program analysis tasks. In particular, we stipulate that a robust distributional hypothesis of code applies to both human- and machine-generated programs. Following this hypothesis, we define an embedding space, inst2vec, based on an Intermediate Representation (IR) of the code that is independent of the source programming language. We provide a novel definition of contextual flow for this IR, leveraging both the underlying data- and control-flow of the program. We then analyze the embeddings qualitatively using analogies and clustering, and evaluate the learned representation on three different high-level tasks. We show that even without fine-tuning, a single RNN architecture and fixed inst2vec embeddings outperform specialized approaches for performance prediction (compute device mapping, optimal thread coarsening); and algorithm classification from raw code (104 classes), where we set a new state-of-the-art.</p><h4><a href="http://papers.nips.cc/paper/8068-generalisation-of-structural-knowledge-in-the-hippocampal-entorhinal-system">Generalisation of structural knowledge in the hippocampal-entorhinal system</a></h4><p><em>Moonshot</em></p><p>A central problem to understanding intelligence is the concept of generalisation. This allows previously learnt structure to be exploited to solve tasks in novel situations differing in their particularities. We take inspiration from neuroscience, specifically the hippocampal-entorhinal system known to be important for generalisation. We propose that to generalise structural knowledge, the representations of the structure of the world, i.e. how entities in the world relate to each other, need to be separated from representations of the entities themselves. We show, under these principles, artificial neural networks embedded with hierarchy and fast Hebbian memory, can learn the statistics of memories and generalise structural knowledge. Spatial neuronal representations mirroring those found in the brain emerge, suggesting spatial cognition is an instance of more general organising principles. We further unify many entorhinal cell types as basis functions for constructing transition graphs, and show these representations effectively utilise memories. We experimentally support model assumptions, showing a preserved relationship between entorhinal grid and hippocampal place cells across environments.</p><h4><a href="http://papers.nips.cc/paper/7419-where-do-you-think-youre-going-inferring-beliefs-about-dynamics-from-behavior">Where Do You Think You&#8217;re Going?: Inferring Beliefs about Dynamics from&nbsp;Behavior</a></h4><p><em>Internals</em></p><p>Inferring intent from observed behavior has been studied extensively within the frameworks of Bayesian inverse planning and inverse reinforcement learning. These methods infer a goal or reward function that best explains the actions of the observed agent, typically a human demonstrator. Another agent can use this inferred intent to predict, imitate, or assist the human user. However, a central assumption in inverse reinforcement learning is that the demonstrator is close to optimal. While models of suboptimal behavior exist, they typically assume that suboptimal actions are the result of some type of random noise or a known cognitive bias, like temporal inconsistency. In this paper, we take an alternative approach, and model suboptimal behavior as the result of internal model misspecification: the reason that user actions might deviate from near-optimal actions is that the user has an incorrect set of beliefs about the rules&#8202;&#8212;&#8202;the dynamics&#8202;&#8212;&#8202;governing how actions affect the environment. Our insight is that while demonstrated actions may be suboptimal in the real world, they may actually be near-optimal with respect to the user&#8217;s internal model of the dynamics. By estimating these internal beliefs from observed behavior, we arrive at a new method for inferring intent. We demonstrate in simulation and in a user study with 12 participants that this approach enables us to more accurately model human intent, and can be used in a variety of applications, including offering assistance in a shared autonomy framework and inferring human preferences.</p><h4><a href="http://papers.nips.cc/paper/7775-task-driven-convolutional-recurrent-models-of-the-visual-system">Task-Driven Convolutional Recurrent Models of the Visual&nbsp;System</a></h4><p><em>Big problems/Future</em></p><p>Feed-forward convolutional neural networks (CNNs) are currently state-of-the-art for object classification tasks such as ImageNet. Further, they are quantitatively accurate models of temporally-averaged responses of neurons in the primate brain&#8217;s visual system. However, biological visual systems have two ubiquitous architectural features not shared with typical CNNs: local recurrence within cortical areas, and long-range feedback from downstream areas to upstream areas. Here we explored the role of recurrence in improving classification performance. We found that standard forms of recurrence (vanilla RNNs and LSTMs) do not perform well within deep CNNs on the ImageNet task. In contrast, novel cells that incorporated two structural features, bypassing and gating, were able to boost task accuracy substantially. We extended these design principles in an automated search over thousands of model architectures, which identified novel local recurrent cells and long-range feedback connections useful for object recognition. Moreover, these task-optimized ConvRNNs matched the dynamics of neural activity in the primate visual system better than feedforward networks, suggesting a role for the brain&#8217;s recurrent connections in performing difficult visual behaviors.</p><div><hr></div><p><a href="https://medium.com/data-science/neurips-2018-reading-list-from-tue-poster-sessions-a-b-fce561e56be8">NIPS/NeurIPS 2018: Best* of the First Two Poster Sessions</a> was originally published in <a href="https://medium.com/data-science">TDS Archive</a> on Medium, where people are continuing the conversation by highlighting and responding to this story.</p>]]></content:encoded></item></channel></rss>