大規模言語モデル
[Wikipedia|▼Menu]
□記事を途中から表示しています
[最初から表示]


^ “Papers with Code - MassiveText Dataset” (英語). paperswithcode.com. 2023年4月26日閲覧。
^ Shijie Wu, Ozan Irsoy, Steven Lu, Vadim Dabravolski, Mark Dredze, Sebastian Gehrmann, Prabhanjan Kambadur, David Rosenberg, Gideon Mann. "BloombergGPT: A Large Language Model for Finance". arXiv:2303.17564。
^ Jesse Dodge, Maarten Sap, Ana Marasovi?, William Agnew, Gabriel Ilharco, Dirk Groeneveld, Margaret Mitchell, Matt Gardner. "Documenting Large Webtext Corpora: A Case Study on the Colossal Clean Crawled Corpus". arXiv:2104.08758。
^ Villalobos, Pablo; Sevilla, Jaime; Heim, Lennart; Besiroglu, Tamay; Hobbhahn, Marius; Ho, Anson (25 October 2022). "Will we run out of data? An analysis of the limits of scaling datasets in Machine Learning". arXiv:2211.04325 [cs.LG]。
^ Tom B. Brown, Benjamin Mann, Nick Ryder, Melanie Subbiah, Jared Kaplan, Prafulla Dhariwal, Arvind Neelakantan, Pranav Shyam, Girish Sastry, Amanda Askell, Sandhini Agarwal, Ariel Herbert-Voss, Gretchen Krueger, Tom Henighan, Rewon Child, Aditya Ramesh, Daniel M. Ziegler, Jeffrey Wu, Clemens Winter, Christopher Hesse, Mark Chen, Eric Sigler, Mateusz Litwin, Scott Gray, Benjamin Chess, Jack Clark, Christopher Berner, Sam McCandlish, Alec Radford, Ilya Sutskever, Dario Amodei. "Language Models are Few-Shot Learners". arXiv:2005.14165。
^ Hoffmann, Jordan; Borgeaud, Sebastian; Mensch, Arthur; Buchatskaya, Elena; Cai, Trevor; Rutherford, Eliza; Casas, Diego de Las; Hendricks, Lisa Anne et al. (2022-03-29). ⇒“Training Compute-Optimal Large Language Models”. arXiv:2203.15556 [cs]. ⇒http://arxiv.org/abs/2203.15556
^ a b Kaplan, Jared; McCandlish, Sam; Henighan, Tom; Brown, Tom B.; Chess, Benjamin; Child, Rewon; Gray, Scott; Radford, Alec et al. (2020). “Scaling Laws for Neural Language Models”. CoRR abs/2001.08361. arXiv:2001.08361. 
^ Caballero, Ethan; Gupta, Kshitij; Rish, Irina; Krueger, David (2022). Broken Neural Scaling Laws. International Conference on Learning Representations (ICLR), 2023.
^ Ornes, Stephen (2023年3月16日). “The Unpredictable Abilities Emerging From Large AI Models”. Quanta Magazine. 2023年5月13日閲覧。
^ Schaeffer, Rylan; Miranda, Brando; Koyejo, Sanmi (1 April 2023). "Are Emergent Abilities of Large Language Models a Mirage?". arXiv:2304.15004 [cs.AI]。
^ Elman, Jeffrey L. (March 1990). ⇒“Finding Structure in Time” (英語). Cognitive Science 14 (2): 179?211. doi:10.1207/s15516709cog1402_1. ⇒http://doi.wiley.com/10.1207/s15516709cog1402_1
^ Sutskever, Ilya; Vinyals, Oriol; Le, Quoc V (2014). “Sequence to Sequence Learning with Neural Networks”. Advances in Neural Information Processing Systems (Curran Associates, Inc.) 27. https://proceedings.neurips.cc/paper_files/paper/2014/hash/a14ac55a4f27472c5d894ec1c3c743d2-Abstract.html. 
^ Cho, Kyunghyun; van Merrienboer, Bart; Bahdanau, Dzmitry; Bengio, Yoshua (2014). “On the Properties of Neural Machine Translation: Encoder?Decoder Approaches”. Proceedings of SSST-8, Eighth Workshop on Syntax, Semantics and Structure in Statistical Translation (Stroudsburg, PA, USA: Association for Computational Linguistics). doi:10.3115/v1/w14-4012. https://doi.org/10.3115/v1/w14-4012. 
^ a b Bahdanau, Dzmitry; Cho, Kyunghyun; Bengio, Yoshua (2014-09-01). Neural Machine Translation by Jointly Learning to Align and Translate. https://ui.adsabs.harvard.edu/abs/2014arXiv1409.0473B. 
^ Lewis-Kraus, Gideon (2016年12月14日). “The Great A.I. Awakening” (英語). The New York Times. ISSN 0362-4331. オリジナルの2023年5月24日時点におけるアーカイブ。. https://web.archive.org/web/20230524052626/https://www.nytimes.com/2016/12/14/magazine/the-great-ai-awakening.html 2023年6月22日閲覧。 
^ Wu, Yonghui; Schuster, Mike; Chen, Zhifeng; Le, Quoc V.; Norouzi, Mohammad; Macherey, Wolfgang; Krikun, Maxim; Cao, Yuan et al. (2016-09-01). Google's Neural Machine Translation System: Bridging the Gap between Human and Machine Translation. https://ui.adsabs.harvard.edu/abs/2016arXiv160908144W. 
^ Vaswani, Ashish; Shazeer, Noam; Parmar, Niki; Uszkoreit, Jakob; Jones, Llion; Gomez, Aidan N; Kaiser, ?ukasz; Polosukhin, Illia (2017). “Attention is All you Need”. Advances in Neural Information Processing Systems (Curran Associates, Inc.) 30. https://proceedings.neurips.cc/paper_files/paper/2017/hash/3f5ee243547dee91fbd053c1c4a845aa-Abstract.html. 
^ Devlin, Jacob; Chang, Ming-Wei; Lee, Kenton; Toutanova, Kristina (11 October 2018). "BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding". arXiv:1810.04805v2 [cs.CL]。
^ “Improving language understanding with unsupervised learning” (英語). openai.com (2018年6月11日). 2023年3月18日時点のオリジナルよりアーカイブ。2023年3月18日閲覧。
^ finetune-transformer-lm, OpenAI, (June 11, 2018), https://github.com/openai/finetune-transformer-lm 2023年5月1日閲覧。 
^ Shazeer, Noam; Mirhoseini, Azalia; Maziarz, Krzysztof; Davis, Andy; Le, Quoc; Hinton, Geoffrey; Dean, Jeff (2017-01-01). Outrageously Large Neural Networks: The Sparsely-Gated Mixture-of-Experts Layer. https://ui.adsabs.harvard.edu/abs/2017arXiv170106538S. 
^ Lepikhin, Dmitry; Lee, HyoukJoong; Xu, Yuanzhong; Chen, Dehao; Firat, Orhan; Huang, Yanping; Krikun, Maxim; Shazeer, Noam et al. (2021-01-12) (英語). GShard: Scaling Giant Models with Conditional Computation and Automatic Sharding. https://openreview.net/forum?id=qrwe7XHTmYb. 
^ a b c “More Efficient In-Context Learning with GLaM” (英語). ai.googleblog.com (2021年12月9日). 2023年3月9日閲覧。
^ “OpenAI API” (英語). platform.openai.com. 2023年4月23日時点のオリジナルよりアーカイブ。2023年4月30日閲覧。
^ “OpenAI API” (英語). platform.openai.com. 2023年6月16日時点のオリジナルよりアーカイブ。2023年6月20日閲覧。
^ LeCun, Yann (2023年4月28日). “A survey of LLMs with a practical guide and evolutionary tree” (英語). Twitter. 2023年6月23日時点のオリジナルよりアーカイブ。2023年6月23日閲覧。
^ Zaib, Munazza; Sheng, Quan Z.; Emma Zhang, Wei (4 February 2020). “A Short Survey of Pre-trained Language Models for Conversational AI-A New Age in NLP”. Proceedings of the Australasian Computer Science Week Multiconference: 1?4. arXiv:2104.10810. doi:10.1145/3373017.3373028. ISBN 9781450376976. https://www.researchgate.net/publication/338931711. 
^ a b c d e f Jurafsky, Dan; Martin, James H. (7 January 2023). Speech and Language Processing (3rd edition draft ed.). https://web.stanford.edu/~jurafsky/slp3/ed3book_jan72023.pdf 2022年5月24日閲覧。 
^ Zhu, Yukun; Kiros, Ryan; Zemel, Rich; Salakhutdinov, Ruslan; Urtasun, Raquel; Torralba, Antonio; Fidler, Sanja (December 2015). “Aligning Books and Movies: Towards Story-Like Visual Explanations by Watching Movies and Reading Books”. 2015 IEEE International Conference on Computer Vision (ICCV): 19?27. arXiv:1506.06724. doi:10.1109/ICCV.2015.11. ISBN 978-1-4673-8391-2. https://www.cv-foundation.org/openaccess/content_iccv_2015/papers/Zhu_Aligning_Books_and_ICCV_2015_paper.pdf 2023年4月11日閲覧。
次ページ
記事の検索
おまかせリスト
▼オプションを表示
ブックマーク登録
mixiチェック!
Twitterに投稿
オプション/リンク一覧
話題のニュース
列車運行情報
暇つぶしWikipedia

Size:164 KB
出典: フリー百科事典『ウィキペディア(Wikipedia)
担当:undef