Paul (Kuo-Ming) Huang

Currently, I am a research assistant at the Computational Learning Lab at National Taiwan University, advised by Prof. Hsuan-Tien Lin. My research interests mainly lie at the intersection of computer vision and generative AI, especially in improving our controllability over generative AI and the effectiveness of generated data.

I received my undergraduate degree in Computer Science and Information Engineering at National Taiwan University, where I completed my Bachelor's Thesis under the supervision of Prof. Hsuan-Tien Lin. During my undergraduate study, I also had the honor of working with Prof. Hung-Yi Lee.

I was also very fortunate to attend Caltech Summer Undergraduate Research Fellowships (SURF) in 2024, where I was mentored by Prof. Yaser S. Abu-Mostafa at Learning Systems Group.

Email  /  CV  /  Scholar  /  Github

profile photo
Me holding a baby alligator.

Publications

Editing the Mind of Giants: An In-Depth Exploration of Pitfalls of Knowledge Editing in Large Language Models
Cheng-Hsun Hsueh*, Paul Kuo-Ming Huang*, Tzu-Han Lin*, Che-Wei Liao*,
Hung-Chieh Fang*, Chao-Wei Huang, Yun-Nung Chen
EMNLP 2024

A survey paper on pitfalls of knowledge editing in large language models. We organized the literatures and benchmarks on the drawbacks of knowledge editing to provide a unified view on this matter.

Prompting and Adapter Tuning for Self-supervised Encoder-Decoder Speech Model
Kai-Wei Chang, Ming-Hsin Chen, Yun-Ping Lin, Jing Neng Hsu,
Paul Kuo-Ming Huang, Chien-yu Huang, Shang-Wen Li, Hung-yi Lee
ASRU 2023

We demonstrate how parameter-efficient tuning can be applied to large pre-trained sequence-to-sequence speech models and its effectiveness, especially in low-resourced settings.

Score-based Conditional Generation with Fewer Labeled Data by Self-calibrating Classifier Guidance
Paul Kuo-Ming Huang, Si-An Chen, Hsuan-Tien Lin
Pre-Print, 2023

We demonstrate the potential of classifier guidance in semi-supervised settings and how it can be further improved through the proposed self-calibration technique, which aligns the probability estimation of classifiers with that of score-based models.

Improving Conditional Score-Based Generation with Calibrated Classification and Joint Training
Paul Kuo-Ming Huang, Si-An Chen, Hsuan-Tien Lin
NeurIPS 2022 Workshop on Score-Based Methods

We demonstrate the benefit of joint training and classifier calibration to classifier guidance.


Source code from Jon Barron.