Loading...

ÇмúÇà»ç

Korean AI Association

  >   ÇмúÇà»ç   >   ±¹³»Çмú´ëȸ

±¹³»Çмú´ëȸ

¿¬»ç ¹× ÃÊ·Ï (16ÀÏ)
 
Ȳ¿ø¼® ±³¼ö(¼­¿ï½Ã¸³´ëÇб³)
 
Title: Tutorial on Large Language Models
 
Abs:

Large language models (LLMs) have demonstrated remarkable performance across diverse domains. For instance, in the legal domain, GPT-4 successfully passes the uniform bar exam. However, the same model fails to pass the Chinese lawyer qualification test and shows suboptimal performance in various other legal AI tasks. Why LLMs display such divergent performance?

This tutorial lecture delves into the underlying technology of LLMs. Starting with a concise introduction to natural language processing, we will briefly review the basics of Transformer neural architecture and study what happens when the model and data scale. Afterward, we will examine the algorithms that empower LLMs to understand and follow instructions. Finally, we will wrap-up the lecture by covering the recent trends in LLMs.

Bio

(Çö) ¼­¿ï½Ã¸³´ë ÀΰøÁö´ÉÇаú Á¶±³¼ö (2023.8-)
(Çö) ¿¤¹Ú½º Research Scientist (2021.7-2023.7, 2023.10-)
(Àü) ³×À̹ö Ŭ·Î¹Ù Research Engineer (2018.3-2021.7)
(Àü) °íµî°úÇпø Research Fellow (2014.9 - 2018.3)
¼­¿ï´ë ¹°¸®Ãµ¹®ÇкΠ¹°¸®ÇÐ ¹Ú»ç(2010.3-2014.8)
¼­¿ï´ë ¹°¸®Ãµ¹®ÇкΠ¹°¸®ÇÐ ¼®»ç(2007.9-2010.2)
¼­¿ï´ë ¹°¸®ÇкΠÇлç (2003.3 - 2007.8)
 

 
±Ç¼¼Áß ¸®´õ(³×À̹öŬ¶ó¿ìµå)
 
Title: Unlocking Efficiency in LLMs: The Role of Model Compression​
 
Abs:
º» °­Á¿¡¼­´Â ±×°£ ´Ù¾çÇÏ°Ô ¿¬±¸µÇ¾î ¿Â Model Compression ±â¼úµé¿¡ ´ëÇØ ¼Ò°³ÇÏ°í, ÇØ´ç ±â¼úµéÀÌ Large Langauge Model ±â¹Ý AI ServiceµéÀÇ ¿î¿ëºñ¿ëÀ» ³·Ãß´Â °üÁ¡¿¡¼­ ¾î¶² Àǹ̰¡ ÀÖ´ÂÁö¸¦ ´Ù·é´Ù.
 
Bio
±Ç¼¼ÁßÀº 2009³â KAIST Àü»êÇаú¸¦ Á¹¾÷ÇÏ°í, KAIST Àü±â¹×ÀüÀÚ°øÇаú¿¡¼­ Modeling and SimulationÀ» ÁÖÁ¦·Î ¹Ú»ç ÇÐÀ§¸¦ ¹Þ¾Ò´Ù. »ï¼º¸®¼­Ä¡¿¡¼­ 3³â°£ On-device AI¸¦ ÁÖÁ¦·Î ¸ðµ¨ ¾ÐÃà/ÃÖÀûÈ­ ¾Ë°í¸®ÁòÀ» ¿¬±¸ÇßÀ¸¸ç, Áö±ÝÀº ³×À̹ö Ŭ¶ó¿ìµå HyperScale AI Á¶Á÷¿¡¼­ Model Optimization ÆÀÀ» ¸®µùÇÏ¿©, HyperCLOVA-XÀÇ °æ·®È­/ÃÖÀûÈ­ ¾÷¹«¸¦ ÁøÇàÇØ¿À°í ÀÖ´Ù. CVPR/ICML/ICLR/Neurips/EMNLP µî ¿©·¯ µö·¯´× ÇÐȸ »Ó¸¸ ¾Æ´Ï¶ó DAC/MLSys/SC µîÀÇ ½Ã½ºÅÛ/Çϵå¿þ¾î ÇÐȸ¿¡µµ ´Ù¾çÇÑ ³í¹®À» äÅýÃÄ×À¸¸ç, ÃÖ±Ù¿¡´Â »ï¼ºÀüÀÚ-³×À̹ö AI¹ÝµµÃ¼ °³¹ßÇù¾÷ ¾÷¹«¸¦ ¸Ã¾Æ, »õ·Î¿î AI ¹ÝµµÃ¼¸¦ ÅëÇØ Àúºñ¿ë/°íÈ¿À²ÀÇ AI ¼­ºñ½º ¼­ºù ½Ã½ºÅÛÀ» ½ÇÇöÇÏ°íÀÚ ³ë·ÂÇÏ°í ÀÖ´Ù.
 

 
 
¹ÚÂù¿µ ±³¼ö(KAIST)
 
Title: Learning on Graph and its Application to Biomedical Data
 
Abs:
In recent times, in the field of chemistry and materials science, Graph Neural Networks (GNNs) have been extensively utilized for modeling the properties and behavior of materials and molecules. GNNs are effective in learning from data with graph structures, making them widely applicable to molecular and crystal structures, which can be easily represented as graphs. GNNs that incorporate both spatial and non-spatial information have shown promising results in predicting properties such as band gaps, formation energies, and reaction energies. In this lecture, I will introduce the fundamental concepts of GNNs and explore examples of their application in predicting chemical reactions.
 
Bio
2020 - Present :  Assistant Professor, KAIST
2019 - 2020 : Postdoctoral Research Fellow, University of Illinois at Urbana-Champaign
2019 : Ph.D, POSTECH 
 

 
¿ÁÁ¤½½ ±³¼ö(POSTECH) 
 
Title: Reinforcement learning: basics and applications 
 
Abs:
Reinforcement Learning (RL) has been emerged as a core paradigm for a wide spectrum of applications. This is perhaps because RL framework can be applied for optimizing various objectives, regardless of their differentiability, including a simple scalar and even human preference (e.g., fine-tuning via RL for chatGPT). This talk will provide the fundamentals of RL, elucidating its core concepts and algorithms. Beginning with an overview of the basic principles underlying RL, including Markov Decision Processes (MDPs) and the exploration-exploitation dilemma, key algorithms such as Q-learning, Deep Q-Networks (DQN), and Policy Gradient methods will be introduced. Then, this talk will be concluded with several real-world applications such as game playing, complex visual recognition system, and language models.
 
Bio
Jungseul Ok is an associate professor in the Department of Computer Science and Engineering and Graduate School of Artificial Intelligence, and a member of Machine Learning Lab at POSTECH. He completed Ph.D program in  School of Electrical Engineering at Korea Advanced Institute of Science and Technology (KAIST), South Korea, under the supervision of Prof. Yung Yi and Prof. Jinwoo Shin. After graduation, he worked with Prof. Alexandre Proutiere and Prof. Sewoong Oh as a postdoctoral researcher, respectively, in School of Electrical Engineering at KTH, Stockholm, Sweden, and Paul G. Allen School of Computer Science & Engineering, University of Washington, WA, US.