Efficient online training algorithms for recurrent neural networks
buir.advisor | Kozat, Süleyman Serdar | |
dc.contributor.author | Vural, Nuri Mert | |
dc.date.accessioned | 2021-01-27T09:30:26Z | |
dc.date.available | 2021-01-27T09:30:26Z | |
dc.date.copyright | 2020-12 | |
dc.date.issued | 2020-12 | |
dc.date.submitted | 2021-01-26 | |
dc.description | Cataloged from PDF version of article. | en_US |
dc.description | Thesis (Master's): Bilkent University, Department of Electrical and Electronics Engineering, İhsan Doğramacı Bilkent University, 2020. | en_US |
dc.description | Includes bibliographical references (leaves 49-53). | en_US |
dc.description.abstract | Recurrent Neural Networks (RNNs) are widely used for online regression due to their ability to learn nonlinear temporal dependencies. As an RNN model, Long-Short-Term-Memory Networks (LSTMs) are commonly preferred in prac-tice, since these networks are capable of learning long-term dependencies while avoiding the exploding gradient problem. On the other hand, the performance improvement of LSTMs usually comes with the price of their large parameter size, which makes their training significantly demanding in terms of computational and data requirements. In this thesis, we address the computational challenges of LSTM training. We introduce two training algorithms, designed for obtaining the online regression performance of LSTMs with less computational requirements than the state-of-the-art. The introduced algorithms are truly online, i.e., they do not assume any underlying data generating process and future information, except that the dataset is bounded. We discuss theoretical guarantees of the introduced algo-rithms, along with their asymptotic convergence behavior. Finally, we demon-strate their performance through extensive numerical studies on real and synthetic datasets, and show that they achieve the regression performance of LSTMs with significantly shorter training times. | en_US |
dc.description.provenance | Submitted by Betül Özen (ozen@bilkent.edu.tr) on 2021-01-27T09:30:26Z No. of bitstreams: 1 10375991.pdf: 1751814 bytes, checksum: 52e338270467005d65194417f1c498a7 (MD5) | en |
dc.description.provenance | Made available in DSpace on 2021-01-27T09:30:26Z (GMT). No. of bitstreams: 1 10375991.pdf: 1751814 bytes, checksum: 52e338270467005d65194417f1c498a7 (MD5) Previous issue date: 2021-01 | en |
dc.description.statementofresponsibility | by Nuri Mert Vural | en_US |
dc.embargo.release | 2021-07-22 | |
dc.format.extent | xii, 74 leaves ; 30 cm. | en_US |
dc.identifier.itemid | B150715 | |
dc.identifier.uri | http://hdl.handle.net/11693/54922 | |
dc.language.iso | English | en_US |
dc.rights | info:eu-repo/semantics/openAccess | en_US |
dc.subject | Long-short-term-memory | en_US |
dc.subject | Recurrent neural networks | en_US |
dc.subject | Online opti-mization | en_US |
dc.subject | Kalman filtering | en_US |
dc.subject | Sequential learning | en_US |
dc.title | Efficient online training algorithms for recurrent neural networks | en_US |
dc.title.alternative | Yineleyici sinir ağları için verimli çevrimici eğitim algoritmaları | en_US |
dc.type | Thesis | en_US |
thesis.degree.discipline | Electrical and Electronic Engineering | |
thesis.degree.grantor | Bilkent University | |
thesis.degree.level | Master's | |
thesis.degree.name | MS (Master of Science) |