Projects

Research projects in NLP, Information Retrieval, and RAG.

WBL: World Best LLM Project

  • Led the data team and established a query clarity evaluation framework using engineered GPT-5.2 prompts, training a Qwen3-4B model as a proprietary tagger to selectively curate datasets with diverse clarity levels.
  • Established a robust response filtering pipeline for off-policy SFT data by ensembling three distinct reward models, applying score fusion techniques to accurately evaluate and retain high-quality responses.
  • Curated and refined large-scale alignment samples by integrating the score fusion pipeline with rigorous LLM-as-a-judge and Code Execution metrics to effectively eliminate repetitive and low-quality responses.

KURE: Korea University Retrieval Embedding Model   GitHub | HuggingFace

  • Led the lab’s flagship Korean retrieval project; curated large-scale training datasets and trained a dense retriever that achieved State-of-the-Art (1st place) on the MTEB-ko-retrieval leaderboard (as of Aug. 2025).
  • Designed and maintained MTEB-ko-retrieval, establishing a comprehensive evaluation suite and standardized public leaderboard for the Korean IR community.
  • Open-sourced the framework, achieving 200+ GitHub stars and 1.1M+ cumulative downloads on Hugging Face.
  • Awarded Best Oral Presentation at HCLT 2025.

Korean ColBERT & Sparse Retrievers   colbert-ko-v1 | splade-ko-v1 | inference-free-splade-ko-v1

  • Trained and open-sourced Korean ColBERT and SPLADE variants, achieving State-of-the-Art performance among corresponding architectures (as of Feb. 2026) on the Korean Retrieval Benchmark.
  • Outperformed existing multilingual and Korean fine-tuned models, providing highly optimized, reproducible pipelines to advance dense-sparse hybrid retrieval experiments.

  • Developed an end-to-end training recipe for a Korean legal-domain LLM; defined real-world judicial tasks, curated expert-written and synthetic alignment datasets, and optimized training strategies to preserve general capabilities.
  • This research directly contributed to KT winning a $10.42 million contract to build an AI platform for the South Korean Supreme Court. [News]
  • Published the training methodologies and data pipelines as LEGALMIDM at the ICLR 2026 Data-FM Workshop.

PreRanker   GitHub | HuggingFace

  • Built a lightweight reranker to narrow down candidate tools, reducing tool-call scope for LLM agents.

URACLE-Korea University Collaborative Research

  • Trained Korean-English cross-lingual retrieval embedding model and analyzed language-pair trade-offs; used model merging to recover mono-lingual retrieval while retaining CLIR gains. [Paper]