Philhoon Oh

I am currently a master’s student at the Kim Jaechul Graduate School of AI, where I am fortunate to be advised by Prof. James Thorne and Prof. Jinwoo, Shin . Before joining KAIST, I was a NLP engineer in SK Inc. C&C, and I received my bachelor’s degree at UC Berkeley, majoring in Statistics.

I believe language plays a pivotal role in transferring knowledge. While humans appeared on Earth millions of years ago, civilization has only been built for thousands of years. It has only been in the last few hundred years that humans have made significant strides by accumulating knowledge through language. In other words, language enables people to think collectively beyond time and space. If we can expedite the sharing of knowledge by incorporating language into machines, it will be possible for humans to make further progress.

My primary interests lie in Information Retrieval (IR) and Retrieval Augmented Generation(RAG). I believe that any form of communication (conversations, audio, text, images) can be reformulated as query + (external knowledge) = response format. With this perspective, I am particularly eager to delve deeper into areas described below:

  1. Retrieval Augmented Generation(RAG)
    • RAG has been effective tool to deal with long-tail knowledge, hallucinations, and up-to-date information.
    • My interests include how to improve the RAG pipeline and overcome the limitation of RAG for better user experience.
  2. Information Retrieval
    • Information Retrieval encompasses a range of intriguing topics, such as Open-Domain Question Answering (ODQA), Multi-hop reasoning, and more.
    • Let’s search for information that can help align with the user’s intention.
  3. LLMs (Large Langauge Models)
    • Interested in efficient utilization of LLMs in various domains (Multi-Modal, Tabular, etc)
    • Some potential methods include prompting, ICL (In-Context Learning), and the utilization of external knowledge sources.

Email: philhoonoh@kaist.ac.kr / vlfgns5@gmail.com

News

Feb 20, 2024 1 paper accepted to LREC-COLING 2024
Jan 2, 2024 Started as Research Intern at NAVER Search
Oct 7, 2023 2 papers accepted to EMNLP 2023
Aug 22, 2023 2nd Place in the 2023 AI Graduate School Challenge (KT CTO Award)
Feb 27, 2023 Joined XFACT Lab at KAIST as a Master student.

Publications

  1. Under Review
    Parallel Key-Value Cache Fusion for Position Invariant RAG
    Philhoon Oh, Jinwoo Shin, and James Thorne
    2025
  2. LREC-COLING
    CLIcK: A Benchmark Dataset of Cultural and Linguistic Intelligence in Korean
    Eunsu Kim, Juyoung Suk, Philhoon Oh, and 3 more authors
    In Proceedings of the 2024 Joint International Conference on Computational Linguistics, Language Resources and Evaluation (LREC-COLING 2024), May 2024
  3. EMNLP
    Detrimental Contexts in Open-Domain Question Answering
    Philhoon Oh, and James Thorne
    In Findings of the Association for Computational Linguistics: EMNLP 2023, Dec 2023
  4. EMNLP
    Knowledge Corpus Error in Question Answering
    Yejoon Lee, Philhoon Oh, and James Thorne
    In Findings of the Association for Computational Linguistics: EMNLP 2023, Dec 2023