Hello! ๐Ÿ‘‹

I'm Junseok Oh (์˜ค์ค€์„) aka june-oh / RiceBerry

์„œ๊ฐ•๋Œ€ํ•™๊ต ์ปดํ“จํ„ฐ๊ณตํ•™๊ณผ ๋ฐ•์‚ฌ๊ณผ์ • ์ˆ˜๋ฃŒ

View CV / Resume

About Me

์„œ๊ฐ•๋Œ€ํ•™๊ต ์ปดํ“จํ„ฐ๊ณตํ•™๊ณผ ๋ฐ•์‚ฌ๊ณผ์ • ์ˆ˜๋ฃŒ (์ง€๋„๊ต์ˆ˜: ๊น€์ง€ํ™˜). ์ฃผ์š” ์—ฐ๊ตฌ ๋ถ„์•ผ๋Š” End-to-End ์Œ์„ฑ ์ธ์‹(ASR), ์Œ์„ฑ ๋ถ„์„ ๋ฐ ํ‰๊ฐ€, ๋ฌธ๋งฅ ์ธ์‹ ๋ฐ ๋„๋ฉ”์ธ ํŠนํ™” ASR, ๋Œ€๊ทœ๋ชจ ์–ธ์–ด ๋ชจ๋ธ(LLM)๊ณผ ์Œ์„ฑ ๊ธฐ์ˆ ์˜ ํ†ตํ•ฉ์ž…๋‹ˆ๋‹ค. ๊ฐ•๊ฑดํ•œ ์ €์ง€์—ฐ ์ŠคํŠธ๋ฆฌ๋ฐ ASR ์‹œ์Šคํ…œ ๊ฐœ๋ฐœ๊ณผ ์ž๋™ ๋งํ•˜๊ธฐ ํ‰๊ฐ€ ํ”„๋ ˆ์ž„์›Œํฌ ๊ตฌ์ถ•์— ์ฃผ๋ ฅํ•˜๊ณ  ์žˆ์Šต๋‹ˆ๋‹ค.

Tech Stack

Programming & Tools

Tech Stack Icons

Speech & ML Frameworks

NVIDIA NeMo Kaldi KenLM Hugging Face Transformers & PEFT Whisper Wav2Vec FastConformer

Research Interests

์Œ์„ฑ ์ธ์‹
์ŠคํŠธ๋ฆฌ๋ฐ ASR ๊ฐ•๊ฑด ASR ์ปจํ…์ŠคํŠธ ๋ฐ”์ด์–ด์‹ฑ ๋„๋ฉ”์ธ ์ ์‘
Speech + LLM
Speech LLM LLM ๊ธฐ๋ฐ˜ ASR ๋ฉ€ํ‹ฐ๋ชจ๋‹ฌ AI
์Œ์„ฑ ๋ถ„์„
๋งํ•˜๊ธฐ ์ž๋™ ํ‰๊ฐ€ ๋งˆ๋น„๋ง์žฅ์•  ๋ถ„์„ ์˜ค๋””์˜ค ์ด๋ฒคํŠธ ๊ฒ€์ถœ

Research Projects

01

๋ฐ•์‚ฌ ํ•™์œ„ ๋…ผ๋ฌธ: Adapter-Only Speechโ€“LLM Bridging

2025.10 - ํ˜„์žฌ

Whisper(์Œ์„ฑ ์ธ์ฝ”๋”)์™€ Gemma(LLM) ๋ชจ๋‘ ๋™๊ฒฐ ์ƒํƒœ์—์„œ ์ „์ฒด์˜ 0.44% ๊ฒฝ๋Ÿ‰ ์–ด๋Œ‘ํ„ฐ๋งŒ ํ•™์Šต. ํ•™์ˆ  ๊ฐ•์—ฐ 4๊ฐœ ๋ถ„์•ผ ํ‰๊ท  26.8% WER ๊ฐ์†Œ. ์ž์—ฐ์–ด ๋„๋ฉ”์ธ ํ”„๋กฌํ”„ํŒ…์œผ๋กœ ์ „๋ฌธ ์–ดํœ˜ F1 +7.2%p ํ–ฅ์ƒ. PAKDD 2026 Accepted (Oral).

PyTorchWhisperGemmaAdapterDomain Adaptation
02

SEAM: ์Œ์„ฑโ€“LLM ์‹œ๊ฐ„์˜๋ฏธ๋ก ์  ์ •๋ ฌ

2025.05 - 2026.01

๊ต์ฐจ ์–ดํ…์…˜ ๊ธฐ๋ฐ˜ ๊ฐ€๋ณ€ ์†๋„ ์ƒ์„ฑ ์ธ์ฝ”๋”-๋””์ฝ”๋” ๋ชจ๋“ˆ. ์Œ์„ฑ ์ธ์ฝ”๋” ๋™๊ฒฐ + LLM LoRA. LibriSpeech 2.6%/5.2% WER, ๊ต์ฐจ ๋„๋ฉ”์ธ TED-LIUM-v2 4.7% WER ๋‹ฌ์„ฑ. EACL 2026 Findings Accepted.

PyTorchWhisperLLMLoRAASR
03

์Œ์„ฑ/๊ฐ€์ฐฝ ๋ณ€ํ™˜ (SVC)

2025
Partner:๊ฐœ์ธ ํ”„๋กœ์ ํŠธ

so-vits-svc ๋ฐ whisper-vits-svc๋ฅผ ํ™œ์šฉํ•œ End-to-end SVC ํŒŒ์ดํ”„๋ผ์ธ. ์•ฝ 10์‹œ๊ฐ„ ํ™”์ž ๋ฐ์ดํ„ฐ ์ˆ˜์ง‘, UVR5 ๋ณด์ปฌ ์ถ”์ถœ, RTX A5000 ํ•™์Šต. ์Œ์„ฑ ๋ณ€ํ™˜ ๋ฐ ๊ฐ€์ฐฝ ๋ณ€ํ™˜ ์ถ”๋ก  ์ˆ˜ํ–‰.

so-vits-svcWhisperUVR5SVCTTS
04

End-to-End ํ•œ๊ตญ์–ด ์Œ์„ฑ์ธ์‹ ํ”„๋กœ์ ํŠธ

2024 - 2025.04
Partner:์‚ฐํ•™ํ˜‘๋ ฅ: ์Šค๋งˆ์ผ๊ฒŒ์ดํŠธ

Hybrid FastConformer RNNT+CTC ๊ธฐ๋ฐ˜ ๋ฒ”์šฉ ํ•œ๊ตญ์–ด ASR ์‹œ์Šคํ…œ ๊ฐœ๋ฐœ. Cache-aware ์ŠคํŠธ๋ฆฌ๋ฐ ์ €์ง€์—ฐ ์ถ”๋ก  ๊ตฌํ˜„. ๊ฒŒ์ž„ ๋„๋ฉ”์ธ ์–ดํœ˜ ์ปจํ…์ŠคํŠธ ๋ฐ”์ด์–ด์‹ฑ ์ ์šฉ์œผ๋กœ ์ธ์‹ ์„ฑ๋Šฅ ํ–ฅ์ƒ.

NVIDIA NeMoFastConformerRNNTCTC
05

์ „ํ™”๋ง(8kHz) ํ™˜๊ฒฝ End-to-End ์Œ์„ฑ์ธ์‹

2024.04 - 2024.12
Partner:์‚ฐํ•™ํ˜‘๋ ฅ: ๋กฏ๋ฐ์ด๋…ธ๋ฒ ์ดํŠธ

8kHz ์ „ํ™”๋ง ๋ฐ์ดํ„ฐ์— ์ตœ์ ํ™”๋œ ์ŠคํŠธ๋ฆฌ๋ฐ/๋น„์ŠคํŠธ๋ฆฌ๋ฐ ํ•œ๊ตญ์–ด ASR ํŒŒ์ดํ”„๋ผ์ธ ๊ฐœ๋ฐœ (FastConformer-CTC). ๋„๋ฉ”์ธ ์‹œํ”„ํŠธ ๋Œ€์‘ ๋™์  ์ปจํ…์ŠคํŠธ ๋ฐ”์ด์–ด์‹ฑ ๋ชจ๋“ˆ ๊ตฌํ˜„.

FastConformerCTCStreaming ASR
06

(์™ธ๊ตญ์ธ์„ ์œ„ํ•œ) 2024๋…„ ํ•œ๊ตญ์–ด ํ‰๊ฐ€ ๋งํ•˜๊ธฐ, ์“ฐ๊ธฐ ์ž๋™ ์ฑ„์  ์—ฐ๊ตฌ

2024.05 - 2024.12
Partner:์ฃผ๊ด€: ๋ฌธํ™”์ฒด์œก๊ด€๊ด‘๋ถ€

Wav2Vec ๊ธฐ๋ฐ˜ ๋ฉ€ํ‹ฐ ํƒœ์Šคํฌ ํ•™์Šต์œผ๋กœ ๋ฐœ์Œ, ์œ ์ฐฝ์„ฑ, ๋‚ด์šฉ์„ ๊ณต๋™ ๋ชจ๋ธ๋งํ•˜๋Š” L2-ํ•œ๊ตญ์–ด ๋งํ•˜๊ธฐ ํ‰๊ฐ€ ํ”„๋ ˆ์ž„์›Œํฌ ๊ฐœ๋ฐœ. Conformer-CTC ASR + LLaMa ๊ฒฐํ•ฉ ๋‹ค์ธก๋ฉด ์ž๋™ ์ฑ„์ .

Wav2VecConformerLLaMa
07

๋Œ€ํ™” ๊ธฐ๋ฐ˜ ์„ค๋ช…๊ฐ€๋Šฅ์„ฑ์„ ๋ฉ€ํ‹ฐ๋ชจ๋‹ฌ๋กœ ์ œ๊ณตํ•˜๋Š” ์ธ๊ณต์ง€๋Šฅ ๊ธฐ์ˆ  ๊ฐœ๋ฐœ

2022.04 - ํ˜„์žฌ
Partner:์ฃผ๊ด€: ๊ณผํ•™๊ธฐ์ˆ ์ •๋ณดํ†ต์‹ ๋ถ€/IITP

๋งˆ๋น„๋ง์žฅ์•  ์ค‘์ฆ๋„ ๋ถ„๋ฅ˜๋ฅผ ์œ„ํ•œ AI ํ”„๋ ˆ์ž„์›Œํฌ ๊ฐœ๋ฐœ. ์Œํ–ฅ/์–ธ์–ด ํŠน์„ฑ ๋ถ„์„ ๊ธฐ๋ฐ˜ ํ•ด์„ ๊ฐ€๋Šฅํ•œ ์ง„๋‹จ ๋ชจ๋“ˆ ๊ตฌํ˜„. ๋Œ€ํ™”ํ˜• ๋ฉ€ํ‹ฐ๋ชจ๋‹ฌ ํ™˜๊ฒฝ์—์„œ AI ์ถœ๋ ฅ๊ณผ ์‚ฌ์šฉ์ž ์ดํ•ด ๊ฐ„ ํ•ด์„ ๊ฒฉ์ฐจ ์™„ํ™”.

Explainable AIMulti-modalSpeech Analysis
08

๋ฌด๋ถ„๋ณ„์  ์œ ํ•ด์„ฑ ๋ฏธ๋””์–ด ๋ฐฐํฌ ๋ฌธ์ œ ํ•ด๊ฒฐ์„ ์œ„ํ•œ ์ง€๋Šฅ์  ๋ถ„์„/๋ถ„๋ฅ˜ ๊ธฐ๋ฐ˜ ์ฝ˜ํ…์ธ  ๋“ฑ๊ธ‰๋ถ„๋ฅ˜ ๊ธฐ์ˆ  ๊ฐœ๋ฐœ

2022 - 2024
Partner:์ฃผ๊ด€: ๊ณผํ•™๊ธฐ์ˆ ์ •๋ณดํ†ต์‹ ๋ถ€

์ž๋™ ์˜์ƒ ๋“ฑ๊ธ‰ ํŒ์ • ํ”„๋ ˆ์ž„์›Œํฌ ๋‚ด ์Œ์„ฑ์ธ์‹ ๋ฐ ์˜ค๋””์˜ค ๋ถ„์„ ๋‹ด๋‹น. ์Œํ–ฅ ์ด๋ฒคํŠธ ๊ฒ€์ถœ(SED) ๋ชจ๋ธ ์„ค๊ณ„. ๋„๋ฉ”์ธ ์˜์ƒ ์ฝ”ํผ์Šค์— ๋งž์ถฐ Whisper ASR ํŒŒ์ธํŠœ๋‹.

WhisperSound Event DetectionFine-tuning
09

(์™ธ๊ตญ์ธ์„ ์œ„ํ•œ) 2023๋…„ ํ•œ๊ตญ์–ด ํ‰๊ฐ€ ๋งํ•˜๊ธฐ, ์“ฐ๊ธฐ ์ž๋™ ์ฑ„์  ์—ฐ๊ตฌ

2023.05 - 2023.12
Partner:์ฃผ๊ด€: ๋ฌธํ™”์ฒด์œก๊ด€๊ด‘๋ถ€

Conformer-CTC ASR ์ถœ๋ ฅ + BERT ๊ธฐ๋ฐ˜ ์˜๋ฏธ ์ ์ˆ˜ํ™”๋ฅผ ๊ฒฐํ•ฉํ•œ L2-ํ•œ๊ตญ์–ด ํ‰๊ฐ€ ํŒŒ์ดํ”„๋ผ์ธ ๊ตฌ์ถ•. ๋ฐœ์Œ ์ •ํ™•๋„, ๋ฐœํ™” ์†๋„, ๊ตฌ๋ฌธ์  ์ •ํ™•์„ฑ ์ •๋Ÿ‰ํ™” ์•Œ๊ณ ๋ฆฌ์ฆ˜ ๊ฐœ๋ฐœ.

ConformerCTCBERT
010

๋น„๋””์˜ค ํŠœ๋ง ํ…Œ์ŠคํŠธ๋ฅผ ํ†ต๊ณผํ•  ์ˆ˜์ค€์˜ ๋น„๋””์˜ค ์Šคํ† ๋ฆฌ ์ดํ•ด ๊ธฐ๋ฐ˜ ์งˆ์˜์‘๋‹ต ๊ธฐ์ˆ  ๊ฐœ๋ฐœ

2017.09 - 2019.12
Partner:์ฃผ๊ด€: ๊ณผํ•™๊ธฐ์ˆ ์ •๋ณดํ†ต์‹ ๋ถ€

Kaldi ๋ฌธ์žฅ ๋‹จ์œ„ ๋””์ฝ”๋”๋ฅผ ์ˆ˜์ •ํ•˜์—ฌ ์‹ค์‹œ๊ฐ„ ๋น„๋””์˜ค QA์—์„œ RT 1.0 ๋ฏธ๋งŒ ๋‹ฌ์„ฑ. ํƒ€๊นƒ ๋น„๋””์˜ค์—์„œ ๋„๋ฉ”์ธ ํŠนํ™” ์ฝ”ํผ์Šค ์ˆ˜์ง‘/์ •์ œ, ์Œํ–ฅ/์–ธ์–ด ๋ชจ๋ธ ์ตœ์ ํ™”.

KaldiLanguage ModelReal-time ASR

Experience

๋ฐ•์‚ฌ ์—ฐ๊ตฌ์›

์„œ๊ฐ•๋Œ€ํ•™๊ต Auditory Intelligence Lab

2022.03 - ํ˜„์žฌ
  • ์ง€๋„๊ต์ˆ˜: ๊น€์ง€ํ™˜
  • End-to-End ASR, ์Œ์„ฑ ๋ถ„์„, LLM ํ†ตํ•ฉ ๊ธฐ์ˆ  ์—ฐ๊ตฌ
  • FastConformer RNNT+CTC ๊ธฐ๋ฐ˜ ์ŠคํŠธ๋ฆฌ๋ฐ ASR ์‹œ์Šคํ…œ ๊ฐœ๋ฐœ
  • EACL, PAKDD, TIIS ๋“ฑ ๊ตญ์ œ ํ•™์ˆ ๋Œ€ํšŒ/์ €๋„ ๋…ผ๋ฌธ ๊ฒŒ์žฌ

์„์‚ฌ ์—ฐ๊ตฌ์›

์„œ๊ฐ•๋Œ€ํ•™๊ต Auditory Intelligence Lab

2017.09 - 2019.08
  • ์ง€๋„๊ต์ˆ˜: ๊น€์ง€ํ™˜
  • ํ•™์œ„ ๋…ผ๋ฌธ: ์•ฝํ•œ ๋ ˆ์ด๋ธ” ๋ง๋ญ‰์น˜๋ฅผ ์ด์šฉํ•œ ํ•œ๊ตญ์–ด ์‹ค์‹œ๊ฐ„ ์ž๋™ ์ „์‚ฌ ์‹œ์Šคํ…œ
  • Kaldi ๋””์ฝ”๋” ์ˆ˜์ •์œผ๋กœ sub-1.0 RT ์‹ค์‹œ๊ฐ„ ์ถ”๋ก  ๋‹ฌ์„ฑ
  • ๋„๋ฉ”์ธ ํŠนํ™” ์ฝ”ํผ์Šค ์ˆ˜์ง‘ ๋ฐ ์Œํ–ฅ/์–ธ์–ด ๋ชจ๋ธ ์ตœ์ ํ™”

Education

์ปดํ“จํ„ฐ๊ณตํ•™๊ณผ ๋ฐ•์‚ฌ๊ณผ์ • ์ˆ˜๋ฃŒ

์„œ๊ฐ•๋Œ€ํ•™๊ต

2022.03 - ํ˜„์žฌ
  • ์ง€๋„๊ต์ˆ˜: ๊น€์ง€ํ™˜
  • ์—ฐ๊ตฌ ๋ถ„์•ผ: End-to-End ASR, ์Œ์„ฑ ๋ถ„์„, LLM ํ†ตํ•ฉ
  • PAKDD 2026 Accepted (Oral Presentation)
  • EACL 2026 Findings Accepted

์ปดํ“จํ„ฐ๊ณตํ•™๊ณผ ๊ณตํ•™์„์‚ฌ

์„œ๊ฐ•๋Œ€ํ•™๊ต

2017.09 - 2019.08
  • ์ง€๋„๊ต์ˆ˜: ๊น€์ง€ํ™˜
  • ํ•™์œ„ ๋…ผ๋ฌธ: ์•ฝํ•œ ๋ ˆ์ด๋ธ” ๋ง๋ญ‰์น˜๋ฅผ ์ด์šฉํ•œ ํ•œ๊ตญ์–ด ์‹ค์‹œ๊ฐ„ ์ž๋™ ์ „์‚ฌ ์‹œ์Šคํ…œ

์ปดํ“จํ„ฐ๊ณตํ•™๊ณผ ๊ณตํ•™ํ•™์‚ฌ

์„œ๊ฐ•๋Œ€ํ•™๊ต

2010.03 - 2017.08

    Publications

    International Journals

    1. [1]

      J. Oh, J. Nam, and J.-H. Kim, "HiTCA: Fusing Hierarchical Text and Contextual Audio for Accurate VCR," EURASIP Journal on Audio, Speech, and Music Processing, 2025.SCIE, Under Review

    2. [2]

      S. Ma, J. Oh, M. Kim, and J.-H. Kim, "Survey on Deep Learning-based Speech Technologies in Voice Chatbot Systems," KSII Transactions on Internet & Information Systems (TIIS), vol. 19, no. 5, pp. 1406-1440, 2025.SCIE

    3. [3]

      J. Oh, E. Cho, and J.-H. Kim, "Integration of WFST language model in pre-trained Korean E2E ASR model," KSII Transactions on Internet and Information Systems (TIIS), vol. 18, no. 6, pp. 1692โ€“1705, 2024.SCIE

    4. [4]

      S. Seo, J. Oh, E. Cho, H. Park, G. Kim, and J.-H. Kim, "TP-MobNet: A Two-pass Mobile Network for Low-complexity Classification of Acoustic Scene," Computers, Materials & Continua, vol. 73, no. 2, 2022.SCIE

    5. [5]

      M. Lim, D. Lee, H. Park, Y. Kang, J. Oh, J.-S. Park, G.-J. Jang, and J.-H. Kim, "Convolutional neural network based audio event classification," KSII Transactions on Internet and Information Systems (TIIS), vol. 12, no. 6, pp. 2748โ€“2760, 2018.SCIE

    International Conferences

    1. [1]

      J. Oh and J.-H. Kim, "Adapter-Only Bridging of Frozen Speech Encoder and Frozen LLM for ASR," in Pacific-Asia Conference on Knowledge Discovery and Data Mining (PAKDD), 2026.Accepted, Oral

    2. [2]

      J. Oh and J.-H. Kim, "SEAM: Bridging the Temporal-Semantic Granularity Gap for LLM-based Speech Recognition," in Findings of the Association for Computational Linguistics: EACL 2026, pp. 2135โ€“2144, 2026.

    3. [3]

      J. Oh, H. Park, and J.-H. Kim, "Speech Intelligibility Prediction of Dysarthria Using Deep Convolutional Networks," in Proc. Asia Pacific International Conference on Information Science and Technology (APIC-IST), pp. 236โ€“237, 2023.

    4. [4]

      M. Kim, J. Oh, and J.-H. Kim, "Automated Dysarthria Severity Classification Using Diadochokinetic test and Speech Intelligibility Based on LightGBM," in Proc. Asia Pacific International Conference on Information Science and Technology (APIC-IST), pp. 12โ€“13, 2023.

    5. [5]

      S. Seo, M. Lim, D. Lee, H. Park, J. Oh, D. J. Rim, and J.-H. Kim, "Environmental noise robustness for Korean fricatives using speech enhancement generative adversarial networks," in Proc. IEEE Int. Conf. Big Data and Smart Computing (BigComp), pp. 1โ€“4, 2019.

    6. [6]

      S. Seo, D. J. Rim, M. Lim, D. Lee, H. Park, J. Oh, C. Kim, and J.-H. Kim, "Shortcut connections based deep speaker embeddings for end-to-end speaker verification system," in Proc. Interspeech, pp. 2928โ€“2932, 2019.

    Domestic Journals

    1. [1]

      ์ด์ •ํ•„, ์žฅ์žฌํ›„, ๊น€์ง€ํ˜„, ๊น€๋ฏผ์„ญ, ๊น€์„ฑ์ค€, ๊น€๋ฏผ์„œ, ๊น€ํ•˜์˜, ์˜ค์ค€์„, ์ •์›, ๊น€์žฅ์—ฐ ์™ธ, "์Œ์„ฑ์— ๊ธฐ๋ฐ˜ํ•œ ๋งˆ๋น„๋ง์žฅ์•  ์ง„๋‹จ๊ณผ ์„ค๋ช…์ด ๊ฐ€๋Šฅํ•œ ์‹œ์Šคํ…œ," ์ •๋ณด๊ณผํ•™ํšŒ์ง€, vol. 42, no. 4, pp. 45โ€“56, 2024.KCI

    2. [2]

      H. Park, Y. Kang, M. Lim, D. Lee, J. Oh, and J.-H. Kim, "LFMMI-based acoustic modeling by using external knowledge," The Journal of the Acoustical Society of Korea, vol. 38, no. 5, pp. 607โ€“613, 2019.KCI

    Teaching

    Teaching Assistant Experience

    ๊ฐ•์˜ ์กฐ๊ตCSE5109/CSEG109/AIEG109/AIE5109

    ์ƒ์„ฑํ˜• AI ๊ธฐ๋ฐ˜ ์˜ค๋””์˜ค์ธ์‹ ๋ฐ ํ•ฉ์„ฑ/๋ณ€ํ™˜

    2024๋…„ 2ํ•™๊ธฐ
    ์„œ๊ฐ•๋Œ€ํ•™๊ต ยท ๊น€์ง€ํ™˜ ๊ต์ˆ˜

    ์˜ค๋””์˜ค ์ฒ˜๋ฆฌ, ๋”ฅ๋Ÿฌ๋‹ ๊ธฐ์ดˆ, ์–ธ์–ด ๋ชจ๋ธ, FastSpeech2 TTS ์‹ค์Šต

    Lab Materials
    ๊ฐ•์˜ ์กฐ๊ตSamsung AI Academy

    ๋”ฅ๋Ÿฌ๋‹ ๊ธฐ๋ฐ˜ ์ž๋™ ์Œ์„ฑ ์ธ์‹

    2023๋…„ ์—ฌ๋ฆ„
    ์„œ๊ฐ•๋Œ€ํ•™๊ต ร— ์‚ผ์„ฑ์ „์ž ยท ๊น€์ง€ํ™˜ ๊ต์ˆ˜

    ๊น€์ง€ํ™˜ ๊ต์ˆ˜ ์ดˆ์ฒญ ๊ฐ•์˜ ์กฐ๊ต. ์˜ค๋””์˜ค ์ฒ˜๋ฆฌ, MLP, CTC, Whisper, NeMo ํŒŒ์ธํŠœ๋‹, WFST ์‹ค์Šต.

    Lab Materials
    ๊ฐ•์˜ ์กฐ๊ตCSE5109/CSEG109/AIEG109/AIE5109

    ์˜ค๋””์˜ค์ธ์‹ ๋ฐ ํ•ฉ์„ฑ๋ณ€ํ™˜

    2023๋…„ 2ํ•™๊ธฐ
    ์„œ๊ฐ•๋Œ€ํ•™๊ต ยท ๊น€์ง€ํ™˜ ๊ต์ˆ˜

    ์˜ค๋””์˜ค ์ฒ˜๋ฆฌ, PyTorch, RNN/CNN/Seq2Seq, FastSpeech2/VocGAN TTS ์‹ค์Šต

    Lab Materials
    ๊ฐ•์˜ ์กฐ๊ตCSE5311/CSEG311/GITA370

    ๋Œ€ํ™”ํ˜• ์‚ฌ์šฉ์ž ์ธํ„ฐํŽ˜์ด์Šค๊ฐœ๋ก 

    2022๋…„
    ์„œ๊ฐ•๋Œ€ํ•™๊ต ยท ๊น€์ง€ํ™˜ ๊ต์ˆ˜

    ๋Œ€ํ™” ์‹œ์Šคํ…œ ๋ฐ ๋Œ€ํ™”ํ˜• AI ์ธํ„ฐํŽ˜์ด์Šค ์„ค๊ณ„ ์‹ค์Šต

    Achievements

    Awards

    ์žฅ๋ ค์ƒ2023

    ํ•œ๊ตญ์–ด AI ๊ฒฝ์ง„๋Œ€ํšŒ

    Track2-1, ์ƒ๋‹ด ์Œ์„ฑ์ธ์‹

    Team '์ƒ๋‹ด ONE': ์˜ค์ค€์„, ๊น€๋ฏผ์„œ, ๋‚จ์ฃผํ˜•

    ์ฃผ๊ด€: ํ•œ๊ตญ์ง€๋Šฅ์ •๋ณด์‚ฌํšŒ์ง„ํฅ์›(NIA)

    ์ตœ์šฐ์ˆ˜์ƒ / ๋„ค์ด๋ฒ„ ๋Œ€ํ‘œ (1์œ„)2022

    ํ•œ๊ตญ์–ด ์ธ๊ณต์ง€๋Šฅ ๊ฒฝ์ง„๋Œ€ํšŒ

    ๊ธฐ์—…ํ˜„์•ˆ (ํšŒ์˜์Œ์„ฑ)

    Team 'SGCSE': ์˜ค์ค€์„, ๊น€ํ•˜์˜

    ์ฃผ๊ด€: ํ•œ๊ตญ์ง€๋Šฅ์ •๋ณด์‚ฌํšŒ์ง„ํฅ์›(NIA)

    ์ตœ์šฐ์ˆ˜์ƒ (1์œ„)2021

    ์Œ์ ˆ์ธ์‹๋ฅ  ์ธก์ • ์•Œ๊ณ ๋ฆฌ์ฆ˜ ๊ฐœ๋ฐœ ๋Œ€ํšŒ

    ์ˆซ์ž ํฌํ•จ ํŒจํ„ด๋ฐœํ™” ์Œ์„ฑ ๋ฐ์ดํ„ฐ์…‹ ํ™œ์šฉ

    Team '๊ฒ€์€์‚ฌ์ผ€๋™': ๋ฐ•ํ˜ธ์„ฑ, ์˜ค์ค€์„, ์กฐ์€์ˆ˜

    ์ฃผ๊ด€: KT alpha

    Patents

    KR 10-2699607 (B1) - ์ฝ”ํผ์Šค ๊ตฌ์ถ• ์„œ๋น„์Šค ์ œ๊ณต ์„œ๋ฒ„ ๋ฐ ๋ฐฉ๋ฒ• (๋“ฑ๋ก: 2024.08)

    Certificates

    NVIDIA Deep Learning Institute - Building Conversational AI Applications (2022)