Advanced Search

Show simple item record

dc.contributor.authorYigit, G.
dc.contributor.authorAmasyali, M.F.
dc.date.accessioned2023-10-19T15:05:33Z
dc.date.available2023-10-19T15:05:33Z
dc.date.issued2021
dc.identifier.isbn9781665436038
dc.identifier.urihttps://doi.org/10.1109/INISTA52262.2021.9548535
dc.identifier.urihttps://hdl.handle.net/20.500.12469/4942
dc.descriptionKocaeli University;Kocaeli University Technoparken_US
dc.description2021 International Conference on INnovations in Intelligent SysTems and Applications, INISTA 2021 --25 August 2021 through 27 August 2021 -- --172175en_US
dc.description.abstractRecurrent Neural Network (RNN) is a widely used deep learning architecture applied to sequence learning problems. However, it is recognized that RNNs suffer from exploding and vanishing gradient problems that prohibit the early layers of the network from learning the gradient information. GRU networks are particular kinds of recurrent networks that reduce the short-comings of these problems. In this study, we propose two variants of the standard GRU with simple but effective modifications. We applied an empirical approach and tried to determine the effectiveness of the current units and recurrent units of gates by giving different coefficients. Interestingly, we realize that applying such minor and simple changes to the standard GRU provides notable improvements. We comparatively evaluate the standard GRU with the proposed two variants on four different tasks: (1) sentiment classification on the IMDB movie review dataset, (2) language modeling task on Penn TreeBank (PTB) dataset, (3) sequence to sequence addition problem, and (4) question answering problem on Facebook's bAbitasks dataset. The evaluation results indicate that the proposed two variants of GRU consistently outperform standard GRU. © 2021 IEEE.en_US
dc.description.sponsorshipACKNOWLEDGMENT G.Yigit is supported by TUB?TAK - B?DEB 2211/A national fellowship program for Ph.D. studies.en_US
dc.language.isoengen_US
dc.publisherInstitute of Electrical and Electronics Engineers Inc.en_US
dc.relation.ispartof2021 International Conference on INnovations in Intelligent SysTems and Applications, INISTA 2021 - Proceedingsen_US
dc.rightsinfo:eu-repo/semantics/closedAccessen_US
dc.subjectGated recurrent unitsen_US
dc.subjectRecurrent neural networksen_US
dc.subjectSeq2seqen_US
dc.subjectClassification (of information)en_US
dc.subjectModeling languagesen_US
dc.subjectMultilayer neural networksen_US
dc.subjectNetwork layersen_US
dc.subjectGated recurrent uniten_US
dc.subjectGradient informationsen_US
dc.subjectLearning architecturesen_US
dc.subjectLearning problemen_US
dc.subjectRecurrent networksen_US
dc.subjectSeq2seqen_US
dc.subjectSequence learningen_US
dc.subjectShort-comingsen_US
dc.subjectSimple++en_US
dc.subjectVanishing gradienten_US
dc.subjectRecurrent neural networksen_US
dc.titleSimple but effective GRU variantsen_US
dc.typeconferenceObjecten_US
dc.departmentN/Aen_US
dc.identifier.doi10.1109/INISTA52262.2021.9548535en_US
dc.identifier.scopus2-s2.0-85116609087en_US
dc.institutionauthorN/A
dc.relation.publicationcategoryKonferans Öğesi - Uluslararası - Kurum Öğretim Elemanıen_US
dc.authorscopusid57215312808
dc.authorscopusid55664402200
dc.khas20231019-Scopusen_US


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record