Воспроизведение и регуляризация SCRN модели

Translated title of the contribution: Reproducing and regularizing the SCRN model

Olzhas Kabdolov, Zhenisbek Assylbekov, Rustem Takhanov

Research output: Chapter in Book/Report/Conference proceedingConference contribution

Abstract

We reproduce the Structurally Constrained Recurrent Network (SCRN) model, and then regularize it using the existing widespread techniques, such as naïve dropout, variational dropout, and weight tying. We show that when regularized and optimized appropriately the SCRN model can achieve performance comparable with the ubiquitous LSTM model in language modeling task on English data, while outperforming it on non-English data.

Translated title of the contributionReproducing and regularizing the SCRN model
Original languageRussian
Title of host publicationCOLING 2018 - 27th International Conference on Computational Linguistics, Proceedings
EditorsEmily M. Bender, Leon Derczynski, Pierre Isabelle
PublisherAssociation for Computational Linguistics (ACL)
Pages1705-1716
Number of pages12
ISBN (Electronic)9781948087506
Publication statusPublished - 2018
Event27th International Conference on Computational Linguistics, COLING 2018 - Santa Fe, United States
Duration: Aug 20 2018Aug 26 2018

Publication series

NameCOLING 2018 - 27th International Conference on Computational Linguistics, Proceedings

Conference

Conference27th International Conference on Computational Linguistics, COLING 2018
Country/TerritoryUnited States
CitySanta Fe
Period8/20/188/26/18

ASJC Scopus subject areas

  • Language and Linguistics
  • Computational Theory and Mathematics
  • Linguistics and Language

Fingerprint

Dive into the research topics of 'Reproducing and regularizing the SCRN model'. Together they form a unique fingerprint.

Cite this