National Taiwan Univ

Dara mining and machine learning

Chih-Jen Lin
https://www.csie.ntu.edu.tw/~cjlin/

Research Field

Smart Computing (Information)

Introduction

Chih-Jen Lin is currently a distinguished professor at the Department of Computer Science, National Taiwan University, and an affiliated professor at the Department of Machine Learning, MBZUAI. He obtained his B.S. degree from National Taiwan University in 1993 and Ph.D. degree from University of Michigan in 1998. His major research areas include machine learning, data mining, and numerical optimization. He is best known for his work on support vector machines (SVM) for data classification. His software LIBSVM is one of the most widely used and cited SVM packages. For his research work he has received many awards, including best paper awards in some top computer science conferences. He is an IEEE fellow, a AAAI fellow, and an ACM fellow for his contribution to machine learning algorithms and software design. More information about him can be found at http://www.csie.ntu.edu.tw/~cjlin.

 

See 

https://www.csie.ntu.edu.tw/~cjlin/

and FAQ for potential students

https://www.csie.ntu.edu.tw/~cjlin/mlgroup/faqstudents.html


Research Topics
  • Efficient training and inference of machine learning methods
  • Algorithm and software for machine learning methods
  • Theoretical and empirical study of optimization methods for machine learning

Honor
  • Outstanding paper award: ACL 2023 (with students).
  • Best paper award, Asian Conference on Machine Learning 2018 (with two students).
  • ACM fellow, 2015.
  • AAAI fellow, 2014.
  • Best paper award, ACM RecSys 2013 (with three students).
  • ACM distinguished scientist, 2011.
  • IEEE fellow, 2011.
  • Best research paper award, ACM KDD 2010 (with three students).
  • Member of the NTU team to win the first prize of KDD Cup 2010, 2011, and 2013. 
     

Educational Background

Ph.D., Industrial & Operations Engineering, University of Michigan.
 


Job Description

Limited access to large-scale computing infrastructure hinders the deployment of modern AI systems, particularly large language models (LLMs). To broaden adoption under such constraints, improving computational efficiency is essential. Leveraging sparsity is a promising approach, as many datasets and model components in applications like natural language processing, information retrieval, and recommender systems are inherently sparse. However, efficiently handling sparse data is challenging: it requires specialized data structures, and general-purpose libraries often deliver suboptimal performance without task-specific optimization.

To address this, we will analyze computational bottlenecks in key sparse operations—such as sparse matrix multiplication in feedforward networks and masked attention—and develop optimized methods for core tasks including clustering, classification, and LLM inference.
 

Preferred Intern Educational Level

Any level is fine

Skill sets or Qualities

Passion on doing deep research