Academic works: Difference between revisions
No edit summary |
No edit summary |
||
Line 1: | Line 1: | ||
I am Wan Ziyu, a current doctoral course student at [https://hci-lab.jp/ Human-Computer Interaction Laboratory], Hokkaido University. I have relatively formal education on opto-electrical engineering and some computer science, and informal self-education on linguistics and phonetics. | |||
I '''am not very proud''' '''to say''' that I have no publications as of now (whenever you are looking at this page, that is). | |||
However, I feel that it is equally important to introduce what I am currently doing and what I have done in the slight case that you are interested. | However, I feel that it is equally important to introduce what I am currently doing and what I have done in the slight case that you are interested. | ||
== Language-independent speech recognition == | == Research strengths and, well, abilities == | ||
To capture the phonemes (sound) rather than text, | I don't have any specifically notable certificates or qualifications so take everything I list here with a pinch of salt. | ||
* Digital signal processing, with Python (librosa / numpy / scipy) and a some C++ (iPlug2 for creating VST apps) | |||
* Neural network (rather basic) with Tensorflow / Keras. Also a tad of PyTorch, but I hate migrating between toolkits. | |||
* Some HTML / JavaScript for unpretty utility webpages. | |||
* Some bash / Python for task automation. | |||
* LLM prompt composition, generic LLM utilisation. | |||
* Basic welding and electrician skills. | |||
== Research topics == | |||
=== Language-independent speech recognition === | |||
To capture the phonemes (sound) rather than text, a language-independent speech recognition system is proposed. | |||
''See: [[Language-independent speech recognition]]'' | ''See: [[Language-independent speech recognition]]'' | ||
=== LLM powered conversational (car) navigation agent === | |||
Based on the half-faith, half-fact (proportions may vary) that vehicle navigation is a social, conversational task, I think it might be good to enable navigation softwares to have some conversation with the driver as well. | |||
''See: [[LLM navigation agent]]'' |
Revision as of 06:26, 6 October 2023
I am Wan Ziyu, a current doctoral course student at Human-Computer Interaction Laboratory, Hokkaido University. I have relatively formal education on opto-electrical engineering and some computer science, and informal self-education on linguistics and phonetics.
I am not very proud to say that I have no publications as of now (whenever you are looking at this page, that is).
However, I feel that it is equally important to introduce what I am currently doing and what I have done in the slight case that you are interested.
Research strengths and, well, abilities
I don't have any specifically notable certificates or qualifications so take everything I list here with a pinch of salt.
- Digital signal processing, with Python (librosa / numpy / scipy) and a some C++ (iPlug2 for creating VST apps)
- Neural network (rather basic) with Tensorflow / Keras. Also a tad of PyTorch, but I hate migrating between toolkits.
- Some HTML / JavaScript for unpretty utility webpages.
- Some bash / Python for task automation.
- LLM prompt composition, generic LLM utilisation.
- Basic welding and electrician skills.
Research topics
Language-independent speech recognition
To capture the phonemes (sound) rather than text, a language-independent speech recognition system is proposed.
See: Language-independent speech recognition
Based on the half-faith, half-fact (proportions may vary) that vehicle navigation is a social, conversational task, I think it might be good to enable navigation softwares to have some conversation with the driver as well.
See: LLM navigation agent