TY - CHAP U1 - Konferenzveröffentlichung A1 - Schall, Martin A1 - Schambach, Marc-Peter A1 - Franz, Matthias O. T1 - Improving gradient-based LSTM training for offline handwriting recognition by careful selection of the optimization method T2 - 3rd Baden-Württemberg Center of Applied Research Symposium on Information and Communication Systems - SInCom 2016 - Karlsruhe, December 2nd, 2016 N2 - Recent years have seen the proposal of several different gradient-based optimization methods for training artificial neural networks. Traditional methods include steepest descent with momentum, newer methods are based on per-parameter learning rates and some approximate Newton-step updates. This work contains the result of several experiments comparing different optimization methods. The experiments were targeted at offline handwriting recognition using hierarchical subsampling networks with recurrent LSTM layers. We present an overview of the used optimization methods, the results that were achieved and a discussion of why the methods lead to different results. Y1 - 2016 UR - https://opus.hs-offenburg.de/1786 SN - 978-3-943301-21-2 SB - 978-3-943301-21-2 SP - 11 EP - 16 ER -