this is a facebook post (https://www.facebook.com/cho.k.hyun/posts/10215959079883430) i wrote in korean as a response/addendum to another facebook post (https://www.facebook.com/jihoon.jeong/posts/10222328298068240) by Prof. Jihoon Jeong on the necessity of rapid, on-going tests in korea. i’m copy-pasting it here for the purpose of leaving a record. NYU에서는 이번 가을에 blended insturction을 했다. 각 수업은 규모 (학생 수 및 주당 강의 수), 특성 (대면 필수) 등을 고려하여 remote, in-person 또는 blended로 학기 시작 전 구분을 지었고, 나는 blended mode의 강의를 진행했다. blended mode 수업의 강의는 in-person 그리고 lab sessions은 평소의 2-3배로 갯수를 늘려서 in-person과 remote를 모두 가졌다. 모든 강의와 lab은 zoom을 통해 livestream했고 이를 통해
So, this happened. My alma mater chose me as the Alumnus of the Year! As a proud alumni of Aalto University, this is my immense honour to be selected as one. See the announcement in Suomi and English. Aalto University (in particular School of Science within) and Finland just keep on giving, and I feel like I continue to receive without giving anything back. I will have to think of some way for me to pay back all that I have received from them. Kiitos paljon! Of course, the whole event was virtual, and due to the time difference, I
i’ve just finished teaching <Deep Learning> this semester together with Yann and Alfredo. the course was in a “blended mode”, implying that lectures were given in person and live-streamed with a limited subset of students allowed to join each week and all the other students joining remotely via Zoom. this has resulted in more active online discussion among students, instructors and assistants over the course, and indeed there were quite a few interesting questions posted on the course page which was run on campuswire. i enjoyed answering those questions, because they made me think quite a bit about them myself.
[Initial posting on Nov 29 2020][Updated on Nov 30 2020] added a section about the scaling law w.r.t. the model size, per request from Felix Hill. [Updated on Dec 1 2020] added a paragraph referring to Dauphin & Bengio’s “Big Neural Networks Waste Capacity“. this is a short post on why i thought (or more like imagined) the scaling laws from <scaling laws for autoregressive generative modeling> by Heninghan et al. “[is] inevitable from using log loss (the reducible part of KL(p||q))” when “the log loss [was used] with a max entropy model“, which was my response to Tim Dettmers’s
Earlier this month (Nov 2020) at the Samsung AI Forum 2020 I was one of the five recipients of the inaugural Samsung AI Researcher of the Year Award by the Samsung Advanced Institute of Technology (SAIT). Samsung has been supporting my research ever since I was a postdoc at Mila in Montreal, and without their support I wouldn’t have been able to support all my PhD students (NSF, i’m looking at you!) Because of this prolonged support, I had been already grateful to Samsung even before this award, and I am even more thankful. It was also a humbling experience