{"id":3940,"date":"2022-12-01T10:27:45","date_gmt":"2022-12-01T01:27:45","guid":{"rendered":"https:\/\/www.phys.s.u-tokyo.ac.jp\/en\/?p=3940"},"modified":"2023-01-13T10:22:09","modified_gmt":"2023-01-13T01:22:09","slug":"ipi-seminar-900-1030-thursday-dec-8-2022","status":"publish","type":"post","link":"https:\/\/www.phys.s.u-tokyo.ac.jp\/en\/about\/3940\/","title":{"rendered":"[finished] ipi seminar 9:00-10:30, Thursday Dec. 8, 2022"},"content":{"rendered":"\n<h3>\u77e5\u306e\u7269\u7406\u5b66\u7814\u7a76\u30bb\u30f3\u30bf\u30fc \/ Institute for Physics of Intelligence (ipi)<\/h3>\n<p>\u3010Date\u3011<br>9:00-10:30, Thur. Dec. 8<\/p>\n<p>\u3010Speaker\u3011<br>Tomer Galanti\/Massachusetts Institute of Technology<\/p>\n<p>\u3010Title\u3011<br>\"On the Role of Neural Collapse in Transfer Learning\"<\/p>\n<p>\u3010Abstract\u3011<br>In a variety of machine learning applications, we have access to a limited amount of data from the task that we would like to solve, as labeled data is often scarce and\/ or expensive. In such cases, training directly on the available data is unlikely to produce a model that performs well on new, unseen test samples.<br>A prominent solution to this problem is to apply transfer learning. In transfer learning, we typically pre-train a foundation model on a given large-scale source task (e.g., ImageNet) and fine-tune it to fit the available data from the downstream task. Recent results show that representations learned by a single classifier over many classes can adapt to new classes with very few samples.<br>In this talk, we provide an explanation for this behavior based on the recently observed phenomenon of neural collapse. We demonstrate both theoretically and empirically that neural collapse generalizes to new samples from the training classes, and - more importantly - to new classes as well, allowing foundation models to provide feature maps that work well in transfer learning and, specifically, in the few-shot setting.<br>This work is based on the following publications:<br>1. T. Galanti, A. Gyorgy, M. Hutter. \"On the Role of Neural Collapse in Transfer Learning\", ICLR 2022.<br>2. T. Galanti, A. Gyorgy, M. Hutter. \"Improved Generalization Bounds for Transfer Learning via Neural Collapse\", ICML Workshop on Pre-Training: Perspectives, Pitfalls, and Paths Forward 2022.<br>3. C. Xu, S. Yang, T. Galanti, B. Wu, X. Yue, B. Zhai, W. Zhan, P. Vajda, K. Keutzer, M. Tomizuka. \"Image2Point: 3D Point-Cloud Understanding with 2D Image Pretrained Models\", ECCV 2022.<\/p>\n<h3><a href=\"https:\/\/www.phys.s.u-tokyo.ac.jp\/wp-content\/uploads\/2022\/12\/Slides-Tomer-Galanti-IPI-seminar.pdf\">PDF<\/a><\/h3>\n<p>&nbsp;<\/p>\n<div><span class=\"x_author-d-iz88z86z86za0dz67zz78zz78zz74zz68zjz80zz71z9iz90z8z86zdnagz83zz74zikpz70zz72zkz79zkfbz71z2z88zz74zm9gxc1z74zz79zk\">*To receive the Zoom invitation and monthly reminders,<span class=\"x_Apple-converted-space\">&nbsp;<\/span><\/span><span class=\"x_author-d-iz88z86z86za0dz67zz78zz78zz74zz68zjz80zz71z9iz90z8z86zdnagz83zz74zikpz70zz72zkz79zkfbz71z2z88zz74zm9gxc1z74zz79zk\"><b>please register via this google form<\/b><\/span><span class=\"x_author-d-iz88z86z86za0dz67zz78zz78zz74zz68zjz80zz71z9iz90z8z86zdnagz83zz74zikpz70zz72zkz79zkfbz71z2z88zz74zm9gxc1z74zz79zk\">:<span class=\"x_Apple-converted-space\">&nbsp;<\/span><\/span><span class=\"x_attrlink x_url x_author-d-iz88z86z86za0dz67zz78zz78zz74zz68zjz80zz71z9iz90z8z86zdnagz83zz74zikpz70zz72zkz79zkfbz71z2z88zz74zm9gxc1z74zz79zk x_url\"><a class=\"x_attrlink\" href=\"https:\/\/forms.gle\/dqxhpsZXLNYvbSB38\" target=\"_blank\" rel=\"noreferrer nofollow noopener\" data-auth=\"NotApplicable\"><u>https:\/\/forms.gle\/dqxhpsZXLNYvbSB38<\/u><\/a><\/span><\/div>\n<div><span class=\"x_author-d-iz88z86z86za0dz67zz78zz78zz74zz68zjz80zz71z9iz90z8z86zdnagz83zz74zikpz70zz72zkz79zkfbz71z2z88zz74zm9gxc1z74zz79zk\">Your e-mail addresses will be used for this purpose only, you can unsubscribe anytime, and we will not send more than three e-mails per month.<\/span><\/div>\n<div>&nbsp;<\/div>\n<div style=\"text-align: right\">Tilman HARTWIG, Ken NAKANISHI, Shinichiro AKIYAMA and Takashi TAKAHASHI<\/div>\n<div>&nbsp;<\/div>\n<div><a href=\"https:\/\/www.phys.s.u-tokyo.ac.jp\/en\/lp\/ipi\/\">\u21e6Top page<\/a><\/div>\n","protected":false},"excerpt":{"rendered":"\u77e5\u306e\u7269\u7406\u5b66\u7814\u7a76\u30bb\u30f3\u30bf\u30fc \/ Institute for Physics of Intelligence (ipi) \u3010Date\u30119:00-10:30, Thur. Dec. 8 \u3010Speaker\u3011Tomer Galant [&hellip;]","protected":false},"author":13,"featured_media":3941,"comment_status":"closed","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":[],"categories":[22],"tags":[],"_links":{"self":[{"href":"https:\/\/www.phys.s.u-tokyo.ac.jp\/en\/wp-json\/wp\/v2\/posts\/3940"}],"collection":[{"href":"https:\/\/www.phys.s.u-tokyo.ac.jp\/en\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.phys.s.u-tokyo.ac.jp\/en\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.phys.s.u-tokyo.ac.jp\/en\/wp-json\/wp\/v2\/users\/13"}],"replies":[{"embeddable":true,"href":"https:\/\/www.phys.s.u-tokyo.ac.jp\/en\/wp-json\/wp\/v2\/comments?post=3940"}],"version-history":[{"count":4,"href":"https:\/\/www.phys.s.u-tokyo.ac.jp\/en\/wp-json\/wp\/v2\/posts\/3940\/revisions"}],"predecessor-version":[{"id":4008,"href":"https:\/\/www.phys.s.u-tokyo.ac.jp\/en\/wp-json\/wp\/v2\/posts\/3940\/revisions\/4008"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/www.phys.s.u-tokyo.ac.jp\/en\/wp-json\/wp\/v2\/media\/3941"}],"wp:attachment":[{"href":"https:\/\/www.phys.s.u-tokyo.ac.jp\/en\/wp-json\/wp\/v2\/media?parent=3940"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.phys.s.u-tokyo.ac.jp\/en\/wp-json\/wp\/v2\/categories?post=3940"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.phys.s.u-tokyo.ac.jp\/en\/wp-json\/wp\/v2\/tags?post=3940"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}