www.wikidata.uk-ua.nina.az
Generative pre trained transformer GPT variant shtuchnogo intelektu 1 2 ce simejstvo movnih modelej yaki zazvichaj navchayutsya na velikomu korpusi tekstovih danih dlya stvorennya tekstu shozhogo na tekst lyudini Voni pobudovani z vikoristannyam dekilkoh blokiv transformnoyi arhitekturi Yih mozhna tochno nalashtuvati dlya vikonannya riznomanitnih zavdan obrobki prirodnoyi movi takih yak generaciya tekstu pereklad movi ta klasifikaciya tekstu Poperednye navchannya v jogo nazvi vidnositsya do pochatkovogo procesu navchannya na velikomu tekstovomu korpusi pid chas yakogo model vchitsya peredbachati nastupne slovo v urivku sho zabezpechuye nadijnu osnovu dlya uspishnoyi roboti modeli v nastupnih zavdannyah z obmezhenoyu kilkistyu danih sho stosuyutsya konkretnogo zavdannya Model GPTZastosuvannya RedaguvatiChatGPT Chat Generative Pre trained Transformer 3 ce chat bot zapushenij OpenAI u listopadi 2022 roku Vin vikoristovuye GPT 3 5 i nalashtovanij pidhid do perenesennya navchannya 4 yak na kerovane navchannya tak i navchannya z pidkriplennyam BioGPT ce GPT yakij fokusuyetsya na vidpovidyah na biomedichni zapitannya 5 Vin rozroblenij Microsoft 6 ProtGPT2 ce GPT yakij zoseredzhuyetsya na bilkovomu dizajni 7 Istoriya RedaguvatiDokladnishe OpenAI Cej rozdil mistit neperekladeni fragmenti anglijskoyu movoyu Vi mozhete dopomogti proyektu pereklavshi yih ukrayinskoyu 11 chervnya 2018 roku OpenAI opublikuvav stattyu pid nazvoyu Pokrashennya rozuminnya movi za dopomogoyu generativnogo poperednogo navchannya v yakomu voni predstavili Generative Pre trained Transformer GPT 8 Na toj moment najefektivnishi nejronni modeli NLP v osnovnomu vikoristovuvali kerovane navchannya z velikih obsyagiv danih poznachenih vruchnu Cya zalezhnist vid kerovanogo navchannya obmezhuvala yih vikoristannya v naborah danih yaki ne buli dobre anotovanimi a takozh robila nadto dorogim i trudomistkim navchannya nadzvichajno velikih modelej 8 9 bagato mov napriklad suahili chi gayityanska kreolska vazhko perekladati ta interpretuvati za dopomogoyu takih modelej cherez brak dostupnogo tekstu dlya pobudovi korpusu 9 Na vidminu vid cogo napivkerovanij pidhid GPT vklyuchav dva etapi nekerovanij generativnij etap poperednogo navchannya na yakomu meta modelyuvannya movi vikoristovuvalasya dlya vstanovlennya pochatkovih parametriv i kerovanij etap diskriminacijnogo tonkogo nalashtuvannya na yakomu ci parametri buli adaptovani do cilovogo zavdannya 8 Versiyi GPT Arhitektura Kilkist parametriv Training data Data relizu Training costGPT 1 12 level 12 headed Transformer decoder no encoder followed by linear softmax 0 12 mlrd BookCorpus 10 4 5 GB tekstu from 7000 unpublished knig riznih zhanriv 11 06 2018 1 misyac na 8 GPU 1 7e 19 FLOP GPT 2 GPT 1 ale z modifikovanoyu normalization 1 5 mlrd WebText 40 GB tekstu 8 mln dokumentiv from 45 mln vebstorinok upvoted na Reddit 14 02 2019 obmezhena versiya 05 11 2019 povna versiya Desyatki petaflops s day 1 5e 21 FLOP GPT 3 GPT 2 but with modification to allow larger scaling 175 mlrd 570 GB plaintext 0 4 trillion tokens Mostly CommonCrawl WebText Anglijska Vikipediya and two books corpora Books1 and Books2 28 05 2020 3640 petaflops s day 3 1e 23 FLOP GPT 3 5 Zasekrechena 175 mlrd Informaciya zasekrechena 15 03 2022 Informaciya zasekrechenaGPT 4 Zasekrechena Nevidomo Informaciya zasekrechena 14 03 2023 Informaciya zasekrechena oriyentovno 2 1e 25 FLOP Primitki Redaguvati Henry Kissinger Eric Schmidt and Daniel Huttenlocher A Robot Wrote This Book Review https konkurent ua publication 112626 shtuchniy intelekt scho bude yakscho ludini ne vdastsya prokontroluvati tehnologiu Shtuchnij intelekt sho bude yaksho lyudini ne vdastsya prokontrolyuvati tehnologiyu Roose Kevin 5 grudnya 2022 The Brilliance and Weirdness of ChatGPT The New York Times amer Arhiv originalu za 18 sichnya 2023 Procitovano 26 grudnya 2022 Like those tools ChatGPT which stands for generative pre trained transformer landed with a splash Quinn Joanne 2020 Dive into deep learning tools for engagement Thousand Oaks California s 551 ISBN 9781544361376 Arhiv originalu za 10 sichnya 2023 Procitovano 10 sichnya 2023 Luo R Sun L Xia Y Qin T Zhang S Poon H 2022 BioGPT generative pre trained transformer for biomedical text generation and mining Brief Bioinform 23 6 PMID 36156661 doi 10 1093 bib bbac409 Matthias Bastian 29 sichnya 2023 BioGPT is a Microsoft language model trained for biomedical tasks The Decoder Ferruz N Schmidt S amp Hocker B 2022 ProtGPT2 is a deep unsupervised language model for protein design Nature Communications volume 13 doi 10 1038 s41467 022 32007 7 a b v Radford Alec Narasimhan Karthik Salimans Tim Sutskever Ilya 11 chervnya 2018 Improving Language Understanding by Generative Pre Training OpenAI s 12 Arhiv originalu za 26 sichnya 2021 Procitovano 23 sichnya 2021 a b Tsvetkov Yulia 22 chervnya 2017 Opportunities and Challenges in Working with Low Resource Languages Carnegie Mellon University Arhiv originalu za 31 bereznya 2020 Procitovano 23 sichnya 2021 Zhu Yukun Kiros Ryan Zemel Rich Salakhutdinov Ruslan Urtasun Raquel Torralba Antonio Fidler Sanja 2015 Aligning Books and Movies Towards Story Like Visual Explanations by Watching Movies and Reading Books s 19 27 Otrimano z https uk wikipedia org w index php title Generative pre trained transformer amp oldid 40638341