Dergi makalesi Açık Erişim

Online Training of LSTM Networks in Distributed Systems for Variable Length Data Sequences

Ergen, Tolga; Kozat, Suleyman S.


Dublin Core

<?xml version='1.0' encoding='utf-8'?>
<oai_dc:dc xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:oai_dc="http://www.openarchives.org/OAI/2.0/oai_dc/" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xsi:schemaLocation="http://www.openarchives.org/OAI/2.0/oai_dc/ http://www.openarchives.org/OAI/2.0/oai_dc.xsd">
  <dc:creator>Ergen, Tolga</dc:creator>
  <dc:creator>Kozat, Suleyman S.</dc:creator>
  <dc:date>2018-01-01</dc:date>
  <dc:description>In this brief, we investigate online training of long short term memory (LSTM) architectures in a distributed network of nodes, where each node employs an LSTM-based structure for online regression. In particular, each node sequentially receives a variable length data sequence with its label and can only exchange information with its neighbors to train the LSTM architecture. We first provide a generic LSTM-based regression structure for each node. In order to train this structure, we put the LSTM equations in a nonlinear state-space form for each node and then introduce a highly effective and efficient distributed particle filtering (DPF)-based training algorithm. We also introduce a distributed extended Kalman filtering-based training algorithm for comparison. Here, our DPF-based training algorithm guarantees convergence to the performance of the optimal LSTM coefficients in the mean square error sense under certain conditions. We achieve this performance with communication and computational complexity in the order of the first-order gradient-based methods. Through both simulated and reallife examples, we illustrate significant performance improvements with respect to the state-of-the-art methods.</dc:description>
  <dc:identifier>https://aperta.ulakbim.gov.trrecord/29013</dc:identifier>
  <dc:identifier>oai:zenodo.org:29013</dc:identifier>
  <dc:rights>info:eu-repo/semantics/openAccess</dc:rights>
  <dc:rights>http://www.opendefinition.org/licenses/cc-by</dc:rights>
  <dc:source>IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 29(10) 5159-5165</dc:source>
  <dc:title>Online Training of LSTM Networks in Distributed Systems for Variable Length Data Sequences</dc:title>
  <dc:type>info:eu-repo/semantics/article</dc:type>
  <dc:type>publication-article</dc:type>
</oai_dc:dc>
6
3
görüntülenme
indirilme
Görüntülenme 6
İndirme 3
Veri hacmi 597 Bytes
Tekil görüntülenme 6
Tekil indirme 3

Alıntı yap