www.wikidata.uk-ua.nina.az
Ne plutati z Rekursivna nejronna merezha Rekure ntni nejro nni mere zhi RNM angl recurrent neural networks RNN ce klas shtuchnih nejronnih merezh u yakomu z yednannya mizh vuzlami utvoryuyut graf oriyentovanij u chasi Ce stvoryuye vnutrishnij stan merezhi sho dozvolyaye yij proyavlyati dinamichnu povedinku v chasi Na vidminu vid nejronnih merezh pryamogo poshirennya RNM mozhut vikoristovuvati svoyu vnutrishnyu pam yat dlya obrobki dovilnih poslidovnostej vhodiv Ce robit yih zastosovnimi do takih zadach yak rozpiznavannya nesegmentovanogo neperervnogo rukopisnogo tekstu 1 ta rozpiznavannya movlennya 2 Zmist 1 Arhitekturi 1 1 Povnorekurentna merezha 1 2 Rekursivni nejronni merezhi 1 3 Merezha Hopfilda 1 4 Merezhi Elmana ta Dzhordana 1 5 Merezha z vidlunnyam stanu 1 6 Nejronnij stiskach istoriyi 1 7 Dovga korotkochasna pam yat 1 8 Ventilnij rekurentnij vuzol 1 9 Dvonapravlena RNM 1 10 RNM neperervnogo chasu 1 11 Iyerarhichna RNM 1 12 Rekurentnij bagatosharovij perceptron 1 13 RNM drugogo poryadku 1 14 Model rekurentnoyi nejronnoyi merezhi kilkoh masshtabiv chasu 1 15 Poslidovni kaskadni merezhi Pollaka 1 16 Nejronni mashini Tyuringa 1 17 Nejromerezhevi magazinni avtomati 1 18 Dvospryamovana asociativna pam yat 2 Trenuvannya 2 1 Gradiyentnij spusk 2 2 Metodi globalnoyi optimizaciyi 3 Pov yazani galuzi ta modeli 4 Poshireni biblioteki RNM 5 Primitki 6 PosilannyaArhitekturi RedaguvatiPovnorekurentna merezha Redaguvati Ce osnovna arhitektura rozroblena v 1980 h rokah merezha nejronopodibnih vuzliv kozhen z oriyentovanim z yednannyam do kozhnogo inshogo vuzla dzherelo Kozhen z vuzliv maye zminnu v chasi dijsnoznachnu aktivaciyu Kozhne z yednannya maye zminyuvanu dijsnoznachnu vagu en Deyaki z vuzliv nazivayutsya vhodovimi vuzlami deyaki vihodovimi a reshta prihovanimi vuzlami Bilshist iz navedenih nizhche arhitektur ye okremimi vipadkami Dlya postanovok kerovanogo navchannya z diskretnim chasom trenuvalni poslidovnosti vhodovih vektoriv stayut poslidovnostyami aktivacij vhodovih vuzliv po odnomu vektoru na kozhen moment chasu proyasniti V kozhen zadanij moment chasu kozhen ne vhodovij vuzol obchislyuye svoyu potochnu aktivaciyu yak nelinijnu funkciyu vid zvazhenoyi sumi aktivacij vsih vuzliv vid yakih do nogo nadhodyat z yednannya proyasniti Dlya deyakih iz vihodovih vuzliv na pevnih taktah mozhut buti zadani vchitelem cilovi aktivaciyi Napriklad yaksho vhodova poslidovnist ye movlennyevim signalom sho vidpovidaye vimovlenij cifri to kincevij cilovij vihid u kinci poslidovnosti mozhe buti mitkoyu yaka klasifikuye cyu cifru Dlya kozhnoyi poslidovnosti yiyi pohibka ye sumoyu vidhilen usih cilovih signaliv vid vidpovidnih aktivacij obchislenih merezheyu Dlya trenuvalnogo naboru chislennih poslidovnostej zagalna pohibka ye sumoyu pohibok usih okremih poslidovnostej Algoritmi minimizaciyi ciyeyi pohibki zgadano v rozdili algoritmiv trenuvannya nizhche U postanovkah navchannya z pidkriplennyam ne isnuye vchitelya yakij nadavav bi cilovi signali dlya RNM natomist chas vid chasu zastosovuyetsya funkciya dopasovanosti abo funkciya vinagorodi dlya ocinyuvannya produktivnosti RNM yaka vplivaye na yiyi vhodovij potik cherez vihodovi vuzli z yednani z privodami sho vplivayut na seredovishe Znov taki zrobit porivnyannya v rozdili pro trenuvalni algoritmi nizhche Rekursivni nejronni merezhi Redaguvati Dokladnishe Rekursivna nejronna merezhaRekursivna nejronna merezha 3 stvoryuyetsya rekursivnim zastosuvannyam odnogo j togo zh naboru vag do diferencijovnoyi grafopodibnoyi strukturi shlyahom obhodu ciyeyi strukturi v topologichnij poslidovnosti Takozh taki merezhi zazvichaj trenuyut zvorotnim rezhimom avtomatichnogo diferenciyuvannya 4 5 Yih bulo vvedeno dlya navchannya rozpodilenih predstavlen strukturi takih yak termini logiki Okremim vipadkom rekursivnih nejronnih merezh ye sami RNM chiya struktura vidpovidaye linijnomu lancyuzhkovi Rekursivni nejronni merezhi zastosovuvali do obrobki prirodnoyi movi 6 Rekursivna nejronna tenzorna merezha angl Recursive Neural Tensor Network vikoristovuye funkciyu komponuvannya na osnovi tenzoriv dlya vsih vuzliv dereva 7 Merezha Hopfilda Redaguvati Dokladnishe Merezha HopfildaMerezha Hopfilda stanovit istorichnij interes hoch vona j ne ye zagalnoprijnyatoyu RNM oskilki yiyi pobudovano ne dlya obrobki poslidovnostej zrazkiv Vona natomist vimagaye stacionarnih vhodiv Vona ye RNM v yakij usi z yednannya ye simetrichnimi Vinajdena Dzhonom Hopfildom 1982 roku vona garantuye sho yiyi dinamika zbigatimetsya Yaksho z yednannya trenuyutsya iz zastosuvannyam gebbovogo navchannya to merezha Hopfilda mozhe pracyuvati yak robastna asociativna pam yat stijka do zmin z yednan Odnim iz variantiv merezhi Hopfilda ye dvospryamovana asociativna pam yat DAP angl bidirectional associative memory BAM DAP maye dva shari kozhen z yakih mozhna vikoristovuvati yak vhodovij shobi viklikati asociaciyu j virobiti vihid na inshomu shari 8 Merezhi Elmana ta Dzhordana Redaguvati Dokladnishe Merezha Elmana ta Merezha Dzhordana nbsp Merezha ElmanaNastupnij okremij vipadok navedenoyi vishe osnovnoyi arhitekturi bulo zastosovano Dzheffom Elmanom en Vikoristovuyetsya trisharova merezha vporyadkovana na ilyustraciyi po gorizontali yak x y ta z z dodavannyam naboru kontekstnih vuzliv angl context units na ilyustraciyi u Isnuyut z yednannya z serednogo prihovanogo sharu z cimi kontekstnimi vuzlami z nezminnimi odinichnimi vagami 9 Na kozhnomu takti vhid poshiryuyetsya standartnim pryamim chinom a potim zastosovuyetsya pravilo navchannya Nezminni zvorotni z yednannya prizvodyat do togo sho kontekstni vuzli zavzhdi zberigayut kopiyu poperednih znachen prihovanih vuzliv oskilki voni poshiryuyutsya z yednannyami do zastosuvannya pravila navchannya Takim chinom merezha Elmana mozhe zberigati svogo rodu stan sho dozvolyaye yij vikonuvati taki zadachi yak peredbachennya poslidovnostej sho ye za mezhami mozhlivostej standartnogo bagatosharovogo perceptronu Merezhi Dzhordana sho zavdyachuyut svoyeyu nazvoyu Majklovi I Dzhordanu en podibni do merezh Elmana Prote podavannya na kontekstni vuzli jde z vihodovogo sharu zamist prihovanogo Kontekstni vuzli v merezhi Dzhordana takozh nazivayut sharom stanu i voni mayut rekurentni z yednannya sami z soboyu bez inshih vuzliv na cih z yednannyah 9 Merezhi Elmana ta Dzhordana vidomi takozh yak prosti rekurentni merezhi PRM angl simple recurrent networks SRN Merezha Elmana 10 h t s h W h x t U h h t 1 b h y t s y W y h t b y displaystyle begin aligned h t amp sigma h W h x t U h h t 1 b h y t amp sigma y W y h t b y end aligned nbsp Merezha Dzhordana 11 h t s h W h x t U h y t 1 b h y t s y W y h t b y displaystyle begin aligned h t amp sigma h W h x t U h y t 1 b h y t amp sigma y W y h t b y end aligned nbsp Zminni ta funkciyi x t displaystyle x t nbsp vhodovij vektor h t displaystyle h t nbsp vektor prihovanogo sharu y t displaystyle y t nbsp vihodovij vektor W displaystyle W nbsp U displaystyle U nbsp ta b displaystyle b nbsp matrici ta vektor parametriv s h displaystyle sigma h nbsp ta s y displaystyle sigma y nbsp funkciyi aktivaciyiMerezha z vidlunnyam stanu Redaguvati Dokladnishe Merezha z vidlunnyam stanuMerezha z vidlunnyam stanu angl echo state network ESN ce rekurentna nejronna merezha z rozridzheno z yednanim vipadkovim prihovanim sharom Yedinoyu chastinoyu merezhi sho mozhe zminyuvatisya j trenuvatisya ye vagi vihodovih nejroniv Taki merezhi ye dobrimi dlya vidtvorennya pevnih chasovih ryadiv 12 Variant dlya nejroniv z potencialom diyi en vidomij yak ridki skinchenni avtomati 13 Nejronnij stiskach istoriyi Redaguvati Problemu znikannya gradiyentu 14 avtomatichnogo diferenciyuvannya ta zvorotnogo poshirennya v nejronnih merezhah bulo chastkovo podolano 1992 roku rannoyu porodzhuvalnoyu modellyu nazvanoyu nejronnim stiskachem istoriyi realizovanoyu yak nekerovana stopka rekurentnih nejronnih merezh RNM 15 RNM na vhodovomu rivni navchayetsya peredbachuvati svij nastupnij vhid z istoriyi poperednih vhodiv Lishe neperedbachuvani vhodi pevnoyi RNM v cij iyerarhiyi stayut vhodami do RNM nastupnogo vishogo rivnya yaka vidtak pereobchislyuye svij vnutrishnij stan lishe zridka Kozhna RNM vishogo rivnya vidtak navchayetsya stisnenogo predstavlennya informaciyi v nizhchij RNM Ce zdijsnyuyetsya takim chinom sho vhidnu poslidovnist mozhe buti tochno vidtvoreno z predstavlennya ciyeyi poslidovnosti na najvishomu rivni Cya sistema diyevo minimizuye dovzhinu opisu abo vid yemnij logarifm imovirnosti danih 16 Yaksho v poslidovnosti vhidnih danih ye bagato peredbachuvanosti yaka piddayetsya navchannyu to RNM najvishogo rivnya mozhe vikoristovuvati kerovane navchannya shobi legko klasifikuvati navit gliboki poslidovnosti z duzhe trivalimi chasovimi intervalami mizh vazhlivimi podiyami 1993 roku taka sistema vzhe rozv yazala zadachu duzhe glibokogo navchannya angl Very Deep Learning yaka vimagaye ponad 1 000 poslidovnih shariv v rozgornutij u chasi RNM 17 Takozh mozhlivo peregnati vsyu cyu iyerarhiyu RNM v lishe dvi RNM sho nazivayut svidomim fragmentatorom angl conscious chunker vishij riven ta pidsvidomim avtomatizatorom angl subconscious automatizer nizhchij riven 15 Shojno fragmentator navchivsya peredbachuvati ta stiskati vhodi yaki ye vse she ne peredbachuvanimi avtomatizatorom to v nastupnij fazi navchannya mozhna navantazhiti avtomatizator peredbachuvannyam abo imituvannyam cherez osoblivi dodatkovi vuzli prihovanih vuzliv povilnishe zminyuvanogo fragmentatora Ce polegshuye avtomatizatorovi navchannya dorechnih ridko zminyuvanih spogadiv protyagom duzhe trivalih promizhkiv chasu Ce svoyeyu chergoyu dopomagaye avtomatizatorovi robiti bagato z jogo kolis neperedbachuvanih vhodiv peredbachuvanimi takim chinom sho fragmentator mozhe zosereditisya na reshti vse she neperedbachuvanih podij shobi stiskati dani she silnishe 15 Dovga korotkochasna pam yat Redaguvati Dokladnishe Dovga korotkochasna pam yatChislenni doslidniki nini vikoristovuyut RNM glibokogo navchannya yaku nazivayut merezheyu dovgoyi korotkochasnoyi pam yati DKChP angl long short term memory LSTM opublikovanu Hohrajterom en ta Shmidguberom 1997 roku 18 Ce sistema glibokogo navchannya yaka na vidminu vid tradicijnih RNM ne maye problemi znikannya gradiyentu porivnyajte v rozdili algoritmiv trenuvannya nizhche DKChP v normi ye dopovnenoyu rekurentnimi ventilyami yaki nazivayut zabuvalnimi angl forget gates 19 RNM DKChP zapobigayut znikannyu ta vibuhannyu zvorotno poshiryuvanih pohibok 14 Natomist pohibki mozhut plinuti v zvorotnomu napryamku kriz neobmezhene chislo virtualnih shariv rozgornutoyi v prostori RNM DKChP Tobto DKChP mozhe vchitisya zavdan duzhe glibokogo navchannya angl Very Deep Learning 20 yaki vimagayut spogadiv pro podiyi sho trapilisya tisyachi abo navit miljoni taktiv tomu Mozhlivo rozvivati DKChP podibni problemno oriyentovani topologiyi 21 DKChP pracyuye navit za trivalih zatrimok i mozhe obroblyati signali sho mayut sumish nizko ta visokochastotnih skladovih Nini bagato zastosunkiv vikoristovuyut stopki RNM DKChP 22 i trenuyut yih nejromerezhevoyu chasovoyu klasifikaciyeyu NChK angl Connectionist Temporal Classification CTC 23 dlya znahodzhennya takoyi vagovoyi matrici RNM yaka maksimizuye jmovirnist poslidovnostej mitok u trenuvalnomu nabori dlya zadanih vidpovidnih vhidnih poslidovnostej NChK dosyagaye yak virivnyuvannya tak i rozpiznavannya Blizko 2007 roku DKChP pochali revolyuciyuvati rozpiznavannya movlennya perevershuyuchi tradicijni modeli v deyakih movlennyevih zastosuvannyah 24 2009 roku DKChP trenovana NChK stala pershoyu RNM yaka peremogla v zmagannyah iz rozpiznavannya obraziv koli vona vigrala kilka zmagan iz neperervnogo rukopisnogo rozpiznavannya 20 25 2014 roku kitajskij poshukovij gigant Baidu zastosuvav RNM trenovani NChK shobi perevershiti etalon rozpiznavannya movlennya Switchboard Hub5 00 bez zastosuvannya zhodnih tradicijnih metodiv obrobki movlennya 26 DKChP takozh polipshila veliko slovnikove rozpiznavannya movlennya 27 28 sintez movlennya z tekstu 29 takozh i dlya Google Android 20 30 i foto realistichni golovi sho rozmovlyayut 29 2015 roku v rozpiznavanni movlennya Google yak povidomlyayetsya stavsya rizkij 49 vidsotkovij dzherelo stribok produktivnosti zavdyaki NChK trenovanij DKChP yaka teper dostupna cherez Google Voice Search vsim koristuvacham smartfoniv 31 DKChP takozh stala duzhe populyarnoyu v galuzi obrobki prirodnoyi movi Na vidminu vid poperednih modelej na osnovi PMM ta podibnih ponyat DKChP mozhe vchitisya rozpiznavati kontekstno chutlivi movi en 32 DKChP polipshila mashinnij pereklad 33 modelyuvannya mov 34 ta bagatomovnu obrobku mov 35 DKChP u poyednanni zi zgortkovimi nejronnimi merezhami ZNM takozh polipshila avtomatichnij opis zobrazhen 36 i bezlich inshih zastosuvan Ventilnij rekurentnij vuzol Redaguvati Dokladnishe Ventilnij rekurentnij vuzolVentilnij rekurentnij vuzol angl gated recurrent unit ye odniyeyu z rekurentnih nejronnih merezh predstavlenih 2014 roku Dvonapravlena RNM Redaguvati Dokladnishe Dvonapravleni rekurentni nejronni merezhiVinajdena Shusterom ta Palivalom 1997 roku 37 dvonapravlena RNM angl bi directional RNN abo DRNM angl BRNN vikoristovuye skinchennu poslidovnist shobi peredbachuvati abo mititi kozhen element ciyeyi poslidovnosti na osnovi yak minulogo tak i majbutnogo kontekstu cogo elementu Ce zdijsnyuyetsya shlyahom z yednannya vihodiv dvoh RNM odna z yakih obroblyaye poslidovnist zliva napravo a insha sprava nalivo Poyednani vihodi ye peredbachennyami zadanih uchitelem cilovih signaliv Cya metodika viyavilasya osoblivo korisnoyu pri poyednanni z RNM DKChP 38 RNM neperervnogo chasu Redaguvati Rekurentna nejronna merezha neperervnogo chasu RNMNCh angl continuous time recurrent neural network CTRNN ce model dinamichnih sistem biologichnih nejronnih merezh Dlya modelyuvannya vpliviv na nejron vhodovogo lancyuzhka aktivacij RNMNCh zastosovuye sistemu zvichajnih diferencialnih rivnyan Dlya nejronu i displaystyle i nbsp v merezhi z potencialom diyi y i displaystyle y i nbsp temp zmini zbudzhennya zadayetsya yak t i y i y i j 1 n w j i s y j 8 j I i t displaystyle tau i dot y i y i sum j 1 n w ji sigma y j Theta j I i t nbsp de t i displaystyle tau i nbsp chasova stala postsinaptichnogo vuzla y i displaystyle y i nbsp zbudzhennya postsinaptichnogo vuzla y i displaystyle dot y i nbsp temp zmini zbudzhennya postsinaptichnogo vuzla w j i displaystyle w ji nbsp vaga z yednannya vid pre do postsinaptichnogo vuzla s x displaystyle sigma x nbsp sigmoyida x displaystyle x nbsp napriklad s x 1 1 e x displaystyle sigma x 1 1 e x nbsp y j displaystyle y j nbsp zbudzhennya presinaptichnogo vuzla 8 j displaystyle Theta j nbsp uperedzhennya presinaptichnogo vuzla I i t displaystyle I i t nbsp vhid yaksho ye do vuzlaRNMNCh chasto zastosovuvali v galuzi evolyucijnoyi robototehniki en de yih vikoristovuvali shobi bratisya za napriklad bachennya 39 vzayemodiyu 40 ta minimalno piznavalnu povedinku 41 Zauvazhte sho za teoremoyu vidlikiv Shennona rekurentni nejronni merezhi diskretnogo chasu mozhna rozglyadati yak rekurentni nejronni merezhi neperervnogo chasu v yakih diferencialne rivnyannya bulo peretvoreno na rivnoznachne rizniceve rivnyannya pislya togo yak funkciyi zbudzhennya postsinaptichnih vuzliv y i t displaystyle y i t nbsp bulo propusheno cherez nizkochastotnij filtr pered diskretizaciyeyu Iyerarhichna RNM Redaguvati Isnuye bagato prikladiv iyerarhichnih RNM angl hierarchical RNN chiyi elementi z yednano riznimi sposobami dlya rozkladu iyerarhichnoyi povedinki na korisni pidprogrami 15 42 Rekurentnij bagatosharovij perceptron Redaguvati Yak pravilo rekurentnij bagatosharovij perceptron RBShP angl Recurrent Multi Layer Perceptron RMLP skladayetsya z ryadu kaskadovanih pidmerezh kozhna z yakih skladayetsya z dekilkoh shariv vuzliv Kozhna z cih pidmerezh ye merezheyu pryamogo poshirennya povnistyu krim ostannogo sharu yakij mozhe mati zvorotni zv yazki vseredini sebe Kozhna z cih pidmerezh pid yednuyetsya lishe zv yazkami pryamogo poshirennya 43 RNM drugogo poryadku Redaguvati RNM drugogo poryadku angl second order RNN vikoristovuyut vagi vishih poryadkiv w i j k displaystyle w ijk nbsp zamist standartnih vagiv w i j displaystyle w ij nbsp a vhodi ta stani mozhut buti dobutkom Ce umozhlivlyuye pryame vidobrazhennya na skinchennij avtomat yak u trenuvanni stijkosti tak i v predstavlenni 44 45 Dovga korotkochasna pam yat ye prikladom cogo krim togo sho vona ne maye takih formalnih vidobrazhen ta dovedennya stijkosti Model rekurentnoyi nejronnoyi merezhi kilkoh masshtabiv chasu Redaguvati Model rekurentnoyi nejronnoyi merezhi kilkoh masshtabiv chasu angl multiple timescales recurrent neural network MTRNN ye mozhlivoyu obchislyuvalnoyu modellyu na nejronnij osnovi yaka do deyakoyi miri imituye diyalnist golovnogo mozku 46 47 Vona maye zdatnist imituvati funkcijnu iyerarhiyu golovnogo mozku cherez samoorganizaciyu yaka zalezhit ne lishe vid prostorovih zv yazkiv mizh nejronami a j vid okremih tipiv nejronnoyi aktivnosti kozhnogo z okremimi chasovimi vlastivostyami Za takih riznih nejronnih aktivnostej neperervni poslidovnosti bud yakoyi mnozhini povedinki segmentuyutsya na pridatni do povtornogo vikoristannya primitivi yaki svoyeyu chergoyu gnuchko vbudovuyutsya do riznomanitnih poslidovnostej povedinki Biologichne pidtverdzhennya takogo tipu iyerarhiyi obgovoryuvalosya v teoriyi pam yati peredbachuvannya funkciyuvannya mozku Dzheffom Gokinsom en u jogo knizi Pro intelekt Poslidovni kaskadni merezhi Pollaka Redaguvati angl Pollack s sequential cascaded networks Nejronni mashini Tyuringa Redaguvati Dokladnishe Nejronna mashina TyuringaNejronni mashini Tyuringa NMT angl Neural Turing machine NTM ce metod rozshirennya mozhlivostej rekurentnih nejronnih merezh shlyahom z yednannya yih iz zovnishnimi resursami pam yati z yakimi voni mozhut vzayemodiyati za dopomogoyu procesiv zoseredzhennya uvagi Taka ob yednana sistema analogichna mashini Tyuringa abo arhitekturi fon Nejmana ale ye diferencijovnoyu z krayu v kraj sho dozvolyaye yij produktivno trenuvatisya za dopomogoyu gradiyentnogo spusku 48 Nejromerezhevi magazinni avtomati Redaguvati Nejromerezhevi magazinni avtomati angl Neural network pushdown automata NNPDA analogichni NMT ale strichki zaminyuyutsya analogovimi stekami yaki ye diferencijovnimi i trenuyutsya dlya keruvannya Takim chinom voni podibni za skladnistyu do rozpiznavachiv kontekstno vilnih gramatik 49 Dvospryamovana asociativna pam yat Redaguvati Dokladnishe Dvospryamovana asociativna pam yatPredstavleni vpershe Bartom Kosko 50 merezhi dvospryamovanoyi asociativnoyi pam yati DAP angl bidirectional associative memory BAM zberigayut asociativni dani yak vektor Dvospryamovanist pohodit vid peredavannya informaciyi matriceyu ta yiyi transpoziciyeyu Yak pravilo dlya dvijkovogo koduvannya par asociacij viddayut perevagu bipolyarnomu koduvannyu en Neshodavno stohastichni modeli DAP z markovskim krokom bulo optimizovano dlya vishoyi stijkosti merezhi ta vidpovidnosti dlya realnih zastosuvan 51 Trenuvannya RedaguvatiGradiyentnij spusk Redaguvati Shobi minimizuvati zagalnu pohibku mozhe zastosovuvatisya gradiyentnij spusk dlya zmini kozhnoyi vagi proporcijno pohidnij pohibki po vidnoshennyu do ciyeyi vagi za umovi sho nelinijni funkciyi aktivaciyi ye diferencijovnimi Dlya zdijsnennya cogo v 1980 h i na pochatku 1990 h rokiv bulo rozrobleno rizni metodi Polom Verbosom en Ronaldom Vilyamsom en Toni Robinsonom Yurgenom Shmidguberom Zeppom Hohrajterom en Barakom Perlmutterom ta inshimi Standartnij metod nazivayetsya zvorotne poshirennya v chasi angl backpropagation through time abo ZPCh angl BPTT i ye uzagalnennyam zvorotnogo poshirennya dlya merezh pryamogo poshirennya 52 53 i yak i toj metod ye zrazkom avtomatichnogo diferenciyuvannya v rezhimi zvorotnogo nakopichennya abo principu minimumu Pontryagina Obchislyuvalno bilsh vitratnij interaktivnij variant nazivayetsya realnochasove rekurentne navchannya angl Real Time Recurrent Learning abo RChRN angl RTRL 54 55 i ye zrazkom avtomatichnogo diferenciyuvannya v rezhimi poslidovnogo nakopichennya zi skladenimi vektorami tangensiv Na vidminu vid ZPCh cej algoritm ye lokalnim v chasi ale ne lokalnim u prostori V comu konteksti lokalnij u prostori oznachaye sho vektor vag vuzla mozhe buti utochneno lishe iz zastosuvannyam informaciyi sho zberigayetsya v z yednanih vuzlah ta samomu vuzli tak sho skladnist utochnennya odnogo vuzla ye linijnoyu po vidnoshennyu do rozmirnosti vektoru vag Lokalnij v chasi oznachaye sho utochnennya vidbuvayutsya neperervno interaktivno i zalezhat lishe vid najneshodavnishogo taktu a ne vid dekilkoh taktiv u mezhah zadanogo obriyu chasu yak u ZPCh Biologichni nejronni merezhi vidayutsya lokalnimi yak u chasi tak i v prostori 56 57 Nedolikom RNRCh ye te sho dlya rekursivnogo obchislennya chastkovih pohidnih vid maye chasovu skladnist O kilkist prihovanih kilkist vag na takt dlya obchislennya matric Yakobi todi yak ZPCh zajmaye lishe O kilkist vag na takt cinoyu prote zberigannya vsih pryamih aktivacij v mezhah zadanogo obriyu chasu 58 Isnuye takozh interaktivnij gibrid ZPCh i RNRCh z promizhnoyu skladnistyu 59 60 i ye varianti dlya neperervnogo chasu 61 Golovnoyu problemoyu gradiyentnogo spusku dlya standartnih arhitektur RNM ye te sho gradiyenti pohibki znikayut eksponencijno shvidko z rozmirom chasovoyi zatrimki mizh vazhlivimi podiyami 14 62 Yak sprobu podolannya cih problem bulo zaproponovano arhitekturu dovgoyi korotkochasnoyi pam yati razom z gibridnim metodom navchannya ZPCh RNRCh 18 Krim togo interaktivnij algoritm sho nazivayetsya prichinnim rekursivnim zvorotnim poshirennyam PRZP angl causal recursive backpropagation CRBP realizuye ta poyednuye razom paradigmi ZPCh ta RNRCh dlya lokalnoyi rekurentnoyi merezhi 63 Vin pracyuye z najzagalnishimi lokalno rekurentnimi merezhami Algoritm PRZP mozhe minimizuvati globalnu pohibku cej fakt prizvodit do polipshenoyi stijkosti algoritmu zabezpechuyuchi ob yednavchij poglyad na metodiki gradiyentnih obchislen dlya rekurentnih merezh iz lokalnim zvorotnim zv yazkom Cikavij pidhid do obchislennya gradiyentnoyi informaciyi v RNM dovilnih arhitektur sho zaproponuvali Van ta Bufe 64 gruntuyetsya na diagramnomu vivedenni grafiv plinu signalu dlya otrimannya paketnogo algoritmu ZPCh todi yak Kampoluchchi Unchini ta Piacca zaproponuvali jogo shvidku interaktivnu versiyu 65 na osnovi teoremi Li 66 dlya obchislennya chutlivosti merezh Metodi globalnoyi optimizaciyi Redaguvati Trenuvannya vag u nejronnij merezhi mozhlivo modelyuvati yak nelinijnu zadachu globalnoyi optimizaciyi Cilovu funkciyu dlya ocinki dopasovanosti abo pohibki pevnogo vagovogo vektora mozhe buti sformovano takim chinom Spershu vagi v merezhi vstanovlyuyutsya vidpovidno do cogo vagovogo vektora Dali merezha ocinyuyetsya za trenuvalnoyu poslidovnistyu Yak pravilo dlya predstavlennya pohibki potochnogo vagovogo vektora vikoristovuyut sumu kvadrativ riznic mizh peredbachennyami ta cilovimi znachennyami vkazanimi v trenuvalnij poslidovnosti Potim dlya minimizaciyi ciyeyi cilovoyi funkciyi mozhe buti zastosovano dovilni metodiki globalnoyi optimizaciyi Najuzhivanishim metodom globalnoyi optimizaciyi dlya trenuvannya RNM ye genetichni algoritmi osoblivo v nestrukturovanih merezhah 67 68 69 Spochatku genetichnij algoritm koduyetsya vagami nejronnoyi merezhi v napered viznachenomu poryadku koli odin gen u hromosomi predstavlyaye odne zvazhene z yednannya i tak dali vsya merezha predstavlyayetsya yedinoyu hromosomoyu Funkciya dopasovanosti obchislyuyetsya nastupnim chinom 1 kozhna vaga zakodovana v hromosomi priznachayetsya vidpovidnomu zvazhenomu z yednannyu merezhi 2 potim trenuvalnij nabir zrazkiv predstavlyayetsya merezhi yaka poshiryuye vhidni signali dali 3 do funkciyi dopasovanosti povertayetsya serednokvadratichna pohibka 4 cya funkciya potim vede proces genetichnogo vidboru Populyaciyu skladayut bagato hromosom takim chinom bagato riznih nejronnih merezh evolyuciyuyut poki ne bude dosyagnuto kriteriyu zupinki Poshirenoyu shemoyu zupinki ye 1 koli nejronna merezha zasvoyila pevnij vidsotok trenuvalnih danih abo 2 koli dosyagnuto minimalnogo znachennya serednokvadratichnoyi pohibki abo 3 koli bulo dosyagnuto maksimalnogo chisla trenuvalnih pokolin Kriterij zupinki ocinyuyetsya funkciyeyu dopasovanosti pri otrimanni neyu obernenogo znachennya serednokvadratichnoyi pohibki z kozhnoyi z nejronnih merezh pid chas trenuvannya Otzhe metoyu genetichnogo algoritmu ye maksimizuvati funkciyu dopasovanosti znizivshi takim chinom serednokvadratichnu pohibku Dlya poshuku dobrogo naboru vag mozhut zastosovuvatisya j inshi metodiki globalnoyi ta abo evolyucijnoyi optimizaciyi taki yak imitaciya vidpalu ta metod royu chastok Pov yazani galuzi ta modeli RedaguvatiRNM mozhut povoditisya haotichno V takih vipadkah dlya analizu mozhna vikoristovuvati teoriyu dinamichnih sistem Rekurentni nejronni merezhi naspravdi ye rekursivnimi nejronnimi merezhami z pevnoyu strukturoyu takoyu yak v linijnogo lancyuzhka V toj chas yak rekursivni nejronni merezhi pracyuyut na bud yakij iyerarhichnij strukturi poyednuyuchi dochirni predstavlennya v batkivski rekurentni nejronni merezhi diyut na linijnij poslidovnosti chasu poyednuyuchi poperednij takt i prihovane predstavlennya v predstavlennya potochnogo taktu Rekurentni nejronni merezhi zokrema mozhna predstavlyati yak nelinijni versiyi filtriv zi skinchennoyu ta neskinchennoyu impulsnoyu harakteristikoyu en a takozh yak nelinijnu avtoregresijnu ekzogennu model angl nonlinear autoregressive exogenous model NARX 70 Poshireni biblioteki RNM RedaguvatiApache Singa en Caffe Stvorena Centrom bachennya ta navchannya Berkli angl Berkeley Vision and Learning Center BVLC Pidtrimuye yak CP tak i GP Rozroblena movoyu C maye obgortki dlya Python ta MATLAB Deeplearning4j Gliboke navchannya v Java ta Scala na Spark z pidtrimkoyu bagatoh GP Biblioteka glibokogo navchannya Arhivovano 30 bereznya 2016 u Wayback Machine zagalnogo priznachennya dlya produktovogo steka JVM sho pracyuye na rushiyi naukovih obchislen C Arhivovano 11 chervnya 2018 u Wayback Machine Dozvolyaye stvoryuvati specialni shari Integruyetsya z Hadoop ta Kafka Keras Microsoft Cognitive Toolkit TensorFlow Theano podibna biblioteka z licenziyeyu Apache 2 0 z pidtrimkoyu CP GP ta zapatentovanih kompaniyeyu Google TP 71 mobilnih Theano Etalonna biblioteka glibokogo navchannya dlya Python z PPI znachnoyu miroyu sumisnim z populyarnoyu bibliotekoyu NumPy Dozvolyaye koristuvacham pisati simvolichni matematichni virazi potim avtomatichno porodzhuye yihni pohidni vberigayuchi koristuvacha vid obov yazku koduvati gradiyenti abo zvorotne poshirennya Ci simvolichni virazi avtomatichno kompilyuyutsya v CUDA dlya otrimannya shvidkoyi realizaciyi na GP Torch www torch ch Arhivovano 9 lipnya 2016 u Wayback Machine Naukovij obchislyuvalnij karkas iz shirokoyu pidtrimkoyu algoritmiv mashinnogo navchannya napisanij movami C ta lua Golovnim avtorom ye Ronan Kollobert narazi zastosovuyetsya u Facebook AI Research ta Twitter Primitki Redaguvati A Graves M Liwicki S Fernandez R Bertolami H Bunke J Schmidhuber A Novel Connectionist System for Improved Unconstrained Handwriting Recognition IEEE Transactions on Pattern Analysis and Machine Intelligence vol 31 no 5 2009 angl H Sak and A W Senior and F Beaufays Long short term memory recurrent neural network architectures for large scale acoustic modeling Proc Interspeech pp338 342 Singapore Sept 201 angl Goller C Kuchler A Learning task dependent distributed representations by backpropagation through structure Neural Networks 1996 IEEE doi 10 1109 ICNN 1996 548916 angl Seppo Linnainmaa en 1970 The representation of the cumulative rounding error of an algorithm as a Taylor expansion of the local rounding errors Master s Thesis in Finnish Univ Helsinki 6 7 fin Griewank Andreas and Walther A Principles and Techniques of Algorithmic Differentiation Second Edition SIAM 2008 angl Socher Richard Lin Cliff Ng Andrew Y Manning Christopher D Parsing Natural Scenes and Natural Language with Recursive Neural Networks The 28th International Conference on Machine Learning ICML 2011 angl Socher Richard Perelygin Alex Y Wu Jean Chuang Jason D Manning Christopher Y Ng Andrew Potts Christopher Recursive Deep Models for Semantic Compositionality Over a Sentiment Treebank EMNLP 2013 Arhiv originalu za 28 grudnya 2016 Procitovano 17 kvitnya 2017 angl Raul Rojas 1996 Neural networks a systematic introduction Springer s 336 ISBN 978 3 540 60505 8 angl a b Cruse Holk Neural Networks as Cybernetic Systems Arhivovano 20 zhovtnya 2016 u Wayback Machine 2nd and revised edition angl Elman Jeffrey L 1990 Finding Structure in Time Cognitive Science 14 2 179 211 doi 10 1016 0364 0213 90 90002 E angl Jordan Michael I 1986 Serial Order A Parallel Distributed Processing Approach angl H Jaeger Harnessing nonlinearity Predicting chaotic systems and saving energy in wireless communication Science 304 78 80 2004 angl W Maass T Natschlager and H Markram A fresh look at real time computation in generic recurrent neural circuits Technical report Institute for Theoretical Computer Science TU Graz 2002 angl a b v Sepp Hochreiter en 1991 Untersuchungen zu dynamischen neuronalen Netzen Arhivovano 6 bereznya 2015 u Wayback Machine Diploma thesis Institut f Informatik Technische Univ Munich Advisor J Schmidhuber nim a b v g Jurgen Schmidhuber Learning complex extended sequences using the principle of history compression Neural Computation 4 2 234 242 Online angl Jurgen Schmidhuber 2015 Deep Learning Scholarpedia 10 11 32832 Section on Unsupervised Pre Training of RNNs and FNNs Arhivovano 19 kvitnya 2016 u Wayback Machine angl Jurgen Schmidhuber 1993 Habilitation thesis TUM 1993 Page 150 ff demonstrates credit assignment across the equivalent of 1 200 layers in an unfolded RNN Online angl a b Hochreiter Sepp en and Schmidhuber Jurgen Long Short Term Memory Neural Computation 9 8 1735 1780 1997 angl Felix Gers Nicholas Schraudolph and Jurgen Schmidhuber 2002 Learning precise timing with LSTM recurrent networks Journal of Machine Learning Research 3 115 143 angl a b v Jurgen Schmidhuber 2015 Deep learning in neural networks An overview Neural Networks 61 2015 85 117 ArXiv Arhivovano 8 travnya 2017 u Wayback Machine angl Justin Bayer Daan Wierstra Julian Togelius and Jurgen Schmidhuber 2009 Evolving memory cell structures for sequence learning Proceedings of ICANN 2 pp 755 764 angl Santiago Fernandez Alex Graves and Jurgen Schmidhuber 2007 Sequence labelling in structured domains with hierarchical recurrent neural networks Proceedings of IJCAI angl Alex Graves Santiago Fernandez Faustino Gomez and Jurgen Schmidhuber 2006 Connectionist temporal classification Labelling unsegmented sequence data with recurrent neural nets Proceedings of ICML 06 pp 369 376 angl Santiago Fernandez Alex Graves and Jurgen Schmidhuber 2007 An application of recurrent neural networks to discriminative keyword spotting Proceedings of ICANN 2 pp 220 229 angl Graves Alex and Schmidhuber Jurgen Offline Handwriting Recognition with Multidimensional Recurrent Neural Networks in Bengio Yoshua Schuurmans Dale Lafferty John Williams Chris K I and Culotta Aron eds Advances in Neural Information Processing Systems 22 NIPS 22 December 7th 10th 2009 Vancouver BC Neural Information Processing Systems NIPS Foundation 2009 pp 545 552 angl Awni Hannun Carl Case Jared Casper Bryan Catanzaro Greg Diamos Erich Elsen Ryan Prenger Sanjeev Satheesh Shubho Sengupta Adam Coates Andrew Ng 2014 Deep Speech Scaling up end to end speech recognition arXiv 1412 5567 Arhivovano 21 grudnya 2016 u Wayback Machine angl Hasim Sak and Andrew Senior and Francoise Beaufays 2014 Long Short Term Memory recurrent neural network architectures for large scale acoustic modeling Proceedings of Interspeech 2014 angl Xiangang Li Xihong Wu 2015 Constructing Long Short Term Memory based Deep Recurrent Neural Networks for Large Vocabulary Speech Recognition arXiv 1410 4281 Arhivovano 26 veresnya 2017 u Wayback Machine angl a b Bo Fan Lijuan Wang Frank K Soong and Lei Xie 2015 Photo Real Talking Head with Deep Bidirectional LSTM In Proceedings of ICASSP 2015 angl Heiga Zen and Hasim Sak 2015 Unidirectional Long Short Term Memory Recurrent Neural Network with Recurrent Output Layer for Low Latency Speech Synthesis In Proceedings of ICASSP pp 4470 4474 angl Hasim Sak Andrew Senior Kanishka Rao Francoise Beaufays and Johan Schalkwyk September 2015 Google voice search faster and more accurate Arhivovano 9 bereznya 2016 u Wayback Machine angl Felix A Gers and Jurgen Schmidhuber LSTM Recurrent Networks Learn Simple Context Free and Context Sensitive Languages IEEE TNN 12 6 1333 1340 2001 angl Ilya Sutskever Oriol Vinyals and Quoc V Le 2014 Sequence to Sequence Learning with Neural Networks arXiv Arhivovano 29 kvitnya 2017 u Wayback Machine angl Rafal Jozefowicz Oriol Vinyals Mike Schuster Noam Shazeer Yonghui Wu 2016 Exploring the Limits of Language Modeling arXiv Arhivovano 6 chervnya 2017 u Wayback Machine angl Dan Gillick Cliff Brunk Oriol Vinyals Amarnag Subramanya 2015 Multilingual Language Processing From Bytes arXiv Arhivovano 26 lipnya 2017 u Wayback Machine angl Oriol Vinyals Alexander Toshev Samy Bengio and Dumitru Erhan 2015 Show and Tell A Neural Image Caption Generator arXiv Arhivovano 4 kvitnya 2017 u Wayback Machine angl Bidirectional recurrent neural networks IEEE Transactions on Signal Processing 45 2673 81 November 1997 angl A Graves and J Schmidhuber Framewise phoneme classification with bidirectional LSTM and other neural network architectures Neural Networks 18 602 610 2005 angl Harvey Inman Husbands P Cliff D 1994 Seeing the light Artificial evolution real vision Proceedings of the third international conference on Simulation of adaptive behavior from animals to animats 3 392 401 angl Quinn Matthew 2001 Evolving communication without dedicated communication channels Advances in Artificial Life Lecture Notes in Computer Science 2159 357 366 ISBN 978 3 540 42567 0 doi 10 1007 3 540 44811 X 38 angl Beer R D 1997 The dynamics of adaptive behavior A research program Robotics and Autonomous Systems 20 2 4 257 289 doi 10 1016 S0921 8890 96 00063 2 angl R W Paine J Tani How hierarchical control self organizes in artificial adaptive systems Adaptive Behavior 13 3 211 225 2005 angl CiteSeerX Recurrent Multilayer Perceptrons for Identification and Control The Road to Applications Citeseerx ist psu edu Arhiv originalu za 28 grudnya 2013 Procitovano 3 sichnya 2014 angl C L Giles C B Miller D Chen H H Chen G Z Sun Y C Lee Learning and Extracting Finite State Automata with Second Order Recurrent Neural Networks Arhivovano 15 kvitnya 2021 u Wayback Machine Neural Computation 4 3 p 393 1992 angl C W Omlin C L Giles Constructing Deterministic Finite State Automata in Recurrent Neural Networks Arhivovano 18 kvitnya 2017 u Wayback Machine Journal of the ACM 45 6 937 972 1996 angl Y Yamashita J Tani Emergence of functional hierarchy in a multiple timescale neural network model a humanoid robot experiment PLoS Computational Biology 4 11 e1000220 211 225 2008 http journals plos org ploscompbiol article id 10 1371 journal pcbi 1000220 Arhivovano 5 bereznya 2022 u Wayback Machine angl Alnajjar F Yamashita Y Tani J 2013 The Hierarchical and Functional Connectivity of Higher order Cognitive Mechanisms Neurorobotic Model to Investigate the Stability and Flexibility of Working Memory Frontiers in Neurorobotics 7 2 doi 10 3389 fnbot 2013 00002 PubMed angl Graves Alex Wayne Greg Danihelka Ivo 2014 Neural Turing Machines arXiv 1410 5401 Arhiv originalu za 7 chervnya 2017 Procitovano 17 kvitnya 2017 angl Guo Zheng Sun C Lee Giles Hsing Hen Chen The Neural Network Pushdown Automaton Architecture Dynamics and Training Adaptive Processing of Sequences and Data Structures Lecture Notes in Computer Science Volume 1387 296 345 1998 angl Kosko B 1988 Bidirectional associative memories IEEE Transactions on Systems Man and Cybernetics 18 1 49 60 doi 10 1109 21 87054 angl Rakkiyappan R Chandrasekar A Lakshmanan S Park Ju H 2 sichnya 2015 Exponential stability for markovian jumping stochastic BAM neural networks with mode dependent probabilistic time varying delays and impulse control Complexity 20 3 39 65 doi 10 1002 cplx 21503 angl P J Werbos Generalization of backpropagation with application to a recurrent gas market model Neural Networks 1 1988 angl David E Rumelhart Geoffrey E Hinton Ronald J Williams Learning Internal Representations by Error Propagation angl A J Robinson and F Fallside The utility driven dynamic error propagation network Technical Report CUED F INFENG TR 1 Cambridge University Engineering Department 1987 angl R J Williams and D Zipser Gradient based learning algorithms for recurrent networks and their computational complexity In Back propagation Theory Architectures and Applications Hillsdale NJ Erlbaum 1994 angl J Schmidhuber A local learning algorithm for dynamic feedforward and recurrent networks Connection Science 1 4 403 412 1989 angl Neural and Adaptive Systems Fundamentals through Simulation J C Principe N R Euliano W C Lefebvre angl Ollivier Y and Tallec C and Charpiat G 2015 Training recurrent networks online without backtracking arXiv Arhivovano 6 lipnya 2017 u Wayback Machine angl J Schmidhuber A fixed size storage O n3 time complexity learning algorithm for fully recurrent continually running networks Neural Computation 4 2 243 248 1992 angl R J Williams Complexity of exact gradient computation algorithms for recurrent neural networks Technical Report Technical Report NU CCS 89 27 Boston Northeastern University College of Computer Science 1989 angl B A Pearlmutter Learning state space trajectories in recurrent neural networks Neural Computation 1 2 263 269 1989 angl S Hochreiter en Y Bengio P Frasconi and J Schmidhuber Gradient flow in recurrent nets the difficulty of learning long term dependencies In S C Kremer and J F Kolen editors A Field Guide to Dynamical Recurrent Neural Networks IEEE Press 2001 angl P Campolucci A Uncini F Piazza B D Rao 1999 On Line Learning Algorithms for Locally Recurrent Neural Networks IEEE Transaction On Neural Networks 10 2 253 271 cherez http ieeexplore ieee org document 750549 angl E A Wan F Beaufays 1996 Diagrammatic derivation of gradient algorithms for neural networks Neural Computation 8 182 201 cherez http www mitpressjournals org doi abs 10 1162 neco 1996 8 1 182 journalCode neco WJ14jYWcGOw angl P Campolucci A Uncini F Piazza 2000 A Signal Flow Graph Approach to On line Gradient Calculation Neural Computation 12 1901 1927 cherez http www mitpressjournals org doi abs 10 1162 089976600300015196 journalCode neco WJ16EoWcGOw angl A Y Lee 1974 Signal Flow Graphs Computer Aided System Analysis and Sensitivity Calculations IEEE Transactions on Circuits and Systems 21 209 216 cherez http ieeexplore ieee org document 1083832 angl F J Gomez and R Miikkulainen Solving non Markovian control tasks with neuroevolution Proc IJCAI 99 Denver CO 1999 Morgan Kaufmann angl Applying Genetic Algorithms to Recurrent Neural Networks for Learning Network Parameters and Architecture Arhiv originalu za 6 grudnya 2010 Procitovano 17 kvitnya 2017 angl F Gomez J Schmidhuber R Miikkulainen Accelerated Neural Evolution through Cooperatively Coevolved Synapses Journal of Machine Learning Research JMLR 9 937 965 2008 angl Hava T Siegelmann Bill G Horne C Lee Giles Computational capabilities of recurrent NARX neural networks IEEE Transactions on Systems Man and Cybernetics Part B 27 2 208 215 1997 angl Cade Metz 18 travnya 2016 Google Built Its Very Own Chips to Power Its AI Bots Wired Arhiv originalu za 13 sichnya 2018 Procitovano 17 kvitnya 2017 angl Mandic Danilo P amp Chambers Jonathon A 2001 Recurrent Neural Networks for Prediction Learning Algorithms Architectures and Stability Wiley ISBN 0 471 49517 4 Posilannya RedaguvatiRNNSharp Arhivovano 4 serpnya 2017 u Wayback Machine CRF na osnovi rekurentnih nejronnih merezh C NET Recurrent Neural Networks Arhivovano 2 sichnya 2014 u Wayback Machine z ponad 60 pracyami z RNM vid grupi Yurgena Shmidgubera v Instituti doslidzhennya shtuchnogo intelektu Dalle Molle en angl Realizaciya nejronnoyi merezhi Elmana Arhivovano 2 lyutogo 2017 u Wayback Machine dlya WEKA Rekurentni nejronni merezhi ta DKChP v Java angl Otrimano z https uk wikipedia org w index php title Rekurentna nejronna merezha amp oldid 40676156