HeeSeung Jung

Machine Learning Engineer at Samsung Research

About Me

I am a Machine Learning Engineer at Samsung Research, the advanced R&D hub of Samsung’s Device eXperience (DX) Division to prepare the future of Samsung Electronics. I achived the M.S. degree at GIST AI graduate school, Gwangju, Korea, working with Prof. Kangil Kim at Intelligence Representation & Reasoning Lab. I am interested in generalization of machine learning (e.g. regularization, penalization) and natural language processing (e.g. machine translation, machine reading comprehension).

Experience

Samsung Electronics | Samsung Research

Security Assurance Lab. | Artificial Intelligence

July 2022 - present

https://research.samsung.com/

Samsung Research is the advanced R&D hub of Samsung’s Device eXperience (DX) Division to prepare the future of Samsung Electronics. We lead the development of the future technologies with about 10,000 researchers and developers working in overseas R&D centers. Under the vision of “Shape the Future with Innovation and Intelligence”, Samsung Research is actively conducting research and development to identify new future growth areas and secure advanced technologies to create new value and improve people’s lives.

In the future, AI technology will become much more prevalent and we will interact with smart devices on a daily basis. These devices will understand our needs via spoken commands, visual gestures, and other cues and be able to help and guide us appropriately. New AI technology in the home and in the workplace will enable us to carry out our routines more efficiently; however, we need to ensure that these technologies operate in a safe and secure manner.

Education

Gwangju Institute of Science and Technology (GIST)

M.S. in AI graduate school

September 2020 - August 2022

https://www.gist.ac.kr

Research for Master Dissertation, Machine Learning and Deep Learning, Neural Engineering, Artificial Intelligence, Intelligent Agent

Government-sponsered full tuition scholarship including residency stipends (2020-2022)

Konkuk University

B.S. in Department of Computer Science and Engineering

March 2016 - February 2020

https://www.konkuk.ac.kr

Data Structure, Algorithm, Linear Algebra, Network Programming, Operating System, Artificial Intelligence, Embedded System Software & Hardware

Institutional Scholarship: Recieved for a grading semester for performing academic excellence (2017, 2018)

Publications

Learning from Matured Dumb Teacher for Fine Generalization

link to paper

IEEE Transaction on Neural Networks and Learning Systems (Waiting for the acceptance)

The flexibility of decision boundaries in neural networks that are unguided by training data is a well-known problem typically resolved with generalization methods. A surprising result from recent knowledge distillation (KD) literature is that random, untrained, and equally structured teacher networks can also vastly improve generalization performance. It raises the possibility of existence of undiscovered assumptions useful for generalization on an uncertain region. In this paper, we shed light on the assumptions by analyzing decision boundaries and confidence distributions of both simple and KD-based generalization methods. Assuming that a decision boundary exists to represent the most general tendency of distinction on an input sample space (i.e., the simplest hypothesis), we show the various limitations of methods when using the hypothesis. To resolve these limitations, we propose matured dumb teacher based KD, conservatively transferring the hypothesis for generalization of the student without massive destruction of trained information. In practical experiments on feed-forward and convolution neural networks for image classification tasks on MNIST, CIFAR-10, and CIFAR-100 datasets, the proposed method shows stable improvement to the best test performance in the grid search of hyperparameters. The analysis and results imply that the proposed method can provide finer generalization than existing methods.

Impact of Sentence Representation Matching in Neural Machine Translation

link to paper

Applied Science

Most neural machine translation models are implemented as a conditional language model framework composed of encoder and decoder models. This framework learns complex and long-distant dependencies, but its deep structure causes inefficiency in training. Matching vector representations of source and target sentences improves the inefficiency by shortening the depth from parameters to costs and generalizes NMTs with different perspective to cross-entropy loss. In this paper, we propose matching methods to derive the cost based on constant word embedding vectors of source and target sentences. To find the best method, we analyze impact of the methods with varying structures, distance metrics, and model capacity in a French to English translation task. An optimally configured method is applied to English from and to French, Spanish, and German translation tasks. In the tasks, the method showed performance improvement by 3.23 BLEU in maximum, 0.71 in average. We evaluated the robustness of this method to various embedding distributions and models as conventional gated structures and transformer network, and empirical results showed that it has higher chance to improve performance in those variety.

합성 샴 신경망을 이용한 화자 인식 (Speaker Recognition Using Convolutional Siamese Neural Networks)

link to paper

[The bachelor thesis] The Korean Institute of Electrical Engineers

Recently, machine learning has been applied in a variety of fields. Speaker recognition is one of attractive applications of machine learning. In this paper, we propose a convolutional Siamese neural network for speaker recognition. The proposed model generates feature vectors through the identical two convolutional neural networks for speech data of two speakers. The similarity is measured by calculating the Euclidean distance of two output feature vectors. If the calculated similarity is less than the threshold, it is judged that two speakers are the same.
The experimental result of the proposed speaker recognition based on the convolutional Siamese neural network showed its accuracy was achieved up to 96%. The accuracy of one-shot classification using the trained convolutional Siamese neural network was evaluated also.
For the evaluation, the 10-way one-shot classification for 10 speakers not used for learning stages were tested, resulting in 92% accuracy.