Output list
Conference paper
Language modeling through Long-Term memory network
Published 2019
2019 International Joint Conference on Neural Networks (IJCNN)
International Joint Conference on Neural Networks (IJCNN) 2019, 14/07/2019–19/07/2019, Budapest, Hungary
Recurrent Neural Networks (RNN), Long Short-Term Memory Networks (LSTM), and Memory Networks which contain memory are popularly used to learn patterns in sequential data. Sequential data has long sequences that hold relationships. RNN can handle long sequences but suffers from the vanishing and exploding gradient problems. While LSTM and other memory networks address this problem, they are not capable of handling long sequences (50 or more data points long sequence patterns). Language modelling requiring learning from longer sequences are affected by the need for more information in memory. This paper introduces Long Term Memory network (LTM), which can tackle the exploding and vanishing gradient problems and handles long sequences without forgetting. LTM is designed to scale data in the memory and gives a higher weight to the input in the sequence. LTM avoid overfitting by scaling the cell state after achieving the optimal results. The LTM is tested on Penn treebank dataset, and Text8 dataset and LTM achieves test perplexities of 83 and 82 respectively. 650 LTM cells achieved a test perplexity of 67 for Penn treebank, and 600 cells achieved a test perplexity of 77 for Text8. LTM achieves state of the art results by only using ten hidden LTM cells for both datasets.
Conference paper
Evolutionary algorithm for Sinhala to English translation
Published 2019
2019 National Information Technology Conference (NITC)
National Information Technology Conference (NITC) 2019, 08/10/2019–10/10/2019, Colombo, Sri Lanka
Machine Translation (MT) is an area in natural language processing, which focuses on translating from one language to another. Many approaches ranging from statistical methods to deep learning approaches are used in order to achieve MT. However, these methods either require a large number of data or a clear understanding about the language. Sinhala language has less digital text which could be used to train a deep neural network. Furthermore, Sinhala has complex rules, and therefore, it is harder to create statistical rules in order to apply statistical methods in MT. This research focuses on Sinhala to English translation using an Evolutionary Algorithm (EA). EA is used to identifying the correct meaning of Sinhala text and to translate it into English. The Sinhala text is passed to identify the meaning in order to get the correct meaning of the sentence. With the use of the EA the translation is carried out. The translated text is passed on to grammatically correct the sentence. This has shown to achieve accurate results.
Conference paper
Enhancing semantic word representations by embedding deep word relationships
Published 2019
Proceedings of the 2019 11th International Conference on Computer and Automation Engineering - ICCAE 2019
11th International Conference on Computer and Automation Engineering, 23/02/2019–25/02/2019, Perth, Western Australia
Word representations are created using analogy context-based statistics and lexical relations on words. Word representations are inputs for the learning models in Natural Language Understanding (NLU) tasks. However, to understand language, knowing only the context is not sufficient. Reading between the lines is a key component of NLU. Embedding deeper word relationships which are not represented in the context enhances the word representation. This paper presents a word embedding which combines an analogy, context-based statistics using Word2Vec, and deeper word relationships using Conceptnet, to create an expanded word representation. In order to fine-tune the word representation, Self-Organizing Map is used to optimize it. The proposed word representation is compared with semantic word representations using Simlex 999. Furthermore, the use of 3D visual representations has shown to be capable of representing the similarity and association between words. The proposed word representation shows a Spearman correlation score of 0.886 and provided the best results when compared to the current state-of-the-art methods, and exceed the human performance of 0.78.
Conference paper
The fuzzy misclassification analysis with deep neural network for handling class noise problem
Published 2018
Neural Information Processing, 11304
International Conference on Neural Information Processing (ICONIP) 2018, 13/12/2018–16/12/2018, Siem Reap, Cambodia
Most of the real world data is embedded with noise, and noise can negatively affect the classification learning models which are used to analyse data. Therefore, noisy data should be handled in order to avoid any negative effect on the learning algorithm used to build the analysis model. Deep learning algorithm has shown to outperform general classification algorithms. However, it has undermined by noisy data. This paper proposes a Fuzzy misclassification the analysis with deep neural networks (FAD) to handle the noise in classification ion data. By combining the fuzzy misclassification analysis with the deep neural network, it can improve the classification confidence by better handling the noisy data. The FAD has tested on Ionosphere, Pima, German and Yeast3 datasets by randomly adding 40% of noise to the data. The FAD has shown to consistently provide good results when compared to other noise removal techniques. FAD has outperformed CMTF-SVM by an average of 3.88% in the testing datasets.