Большая языковая модель (Russian Wikipedia)

Analysis of information sources in references of the Wikipedia article "Большая языковая модель" in Russian language version.

refsWebsite
Global rank Russian rank
1st place
1st place
69th place
148th place
383rd place
298th place
1,559th place
1,840th place
1,272nd place
1,197th place
187th place
479th place
77th place
166th place
179th place
213th place
234th place
157th place
low place
low place
616th place
628th place
low place
low place
153rd place
177th place
low place
low place
low place
low place
878th place
60th place
low place
low place
low place
low place
105th place
247th place

amazon.com

aws.amazon.com

amazon.science

anthropic.com

arxiv.org

  • Wei, Jason Emergent Abilities of Large Language Models. Дата обращения: 16 марта 2023. Архивировано 16 марта 2023 года.
  • Kaplan, Jared; McCandlish, Sam; Henighan, Tom; Brown, Tom B.; Chess, Benjamin; Child, Rewon; Gray, Scott; Radford, Alec; Wu, Jeffrey; Amodei, Dario (2020). "Scaling Laws for Neural Language Models". CoRR. abs/2001.08361. arXiv:2001.08361. Архивировано 15 марта 2023. Дата обращения: 16 марта 2023. {{cite journal}}: |archive-date= / |archive-url= несоответствие временной метки; предлагается 15 марта 2023 (справка)
  • Devlin, Jacob; Chang, Ming-Wei; Lee, Kenton; Toutanova, Kristina (11 октября 2018). "BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding". arXiv:1810.04805v2 [cs.CL].
  • Gao, Leo; Biderman, Stella; Black, Sid; Golding, Laurence; Hoppe, Travis; Foster, Charles; Phang, Jason; He, Horace; Thite, Anish; Nabeshima, Noa; Presser, Shawn; Leahy, Connor (31 декабря 2020). "The Pile: An 800GB Dataset of Diverse Text for Language Modeling". arXiv:2101.00027.
  • Smith, Shaden; Patwary, Mostofa; Norick, Brandon; LeGresley, Patrick; Rajbhandari, Samyam; Casper, Jared; Liu, Zhun; Prabhumoye, Shrimai; Zerveas, George; Korthikanti, Vijay; Zhang, Elton; Child, Rewon; Aminabadi, Reza Yazdani; Bernauer, Julie; Song, Xia (2022-02-04). "Using DeepSpeed and Megatron to Train Megatron-Turing NLG 530B, A Large-Scale Generative Language Model". arXiv:2201.11990 Архивная копия от 15 апреля 2023 на Wayback Machine
  • Askell, Amanda; Bai, Yuntao; Chen, Anna; et al. (9 декабря 2021). "A General Language Assistant as a Laboratory for Alignment". arXiv:2112.00861.
  • Bai, Yuntao; Kadavath, Saurav; Kundu, Sandipan; et al. (15 декабря 2022). "Constitutional AI: Harmlessness from AI Feedback". arXiv:2212.08073.
  • Hoffmann, Jordan; Borgeaud, Sebastian; Mensch, Arthur; Buchatskaya, Elena; Cai, Trevor; Rutherford, Eliza; Casas, Diego de Las; Hendricks, Lisa Anne; Welbl, Johannes; Clark, Aidan; Hennigan, Tom; Noland, Eric; Millican, Katie; Driessche, George van den; Damoc, Bogdan; Guy, Aurelia; Osindero, Simon; Simonyan, Karen; Elsen, Erich; Rae, Jack W.; Vinyals, Oriol; Sifre, Laurent (29 марта 2022). "Training Compute-Optimal Large Language Models". arXiv:2203.15556.
  • Zhang, Susan; Roller, Stephen; Goyal, Naman; Artetxe, Mikel; Chen, Moya; Chen, Shuohui; Dewan, Christopher; Diab, Mona; Li, Xian; Lin, Xi Victoria; Mihaylov, Todor; Ott, Myle; Shleifer, Sam; Shuster, Kurt; Simig, Daniel; Koura, Punit Singh; Sridhar, Anjali; Wang, Tianlu; Zettlemoyer, Luke (21 июня 2022). "OPT: Open Pre-trained Transformer Language Models". arXiv:2205.01068.
  • Soltan, Saleh; Ananthakrishnan, Shankar; FitzGerald, Jack; et al. (3 августа 2022). "AlexaTM 20B: Few-Shot Learning Using a Large-Scale Multilingual Seq2Seq Model". arXiv:2208.01448.

deepmind.com

facebook.com

ai.facebook.com

forefront.ai

github.com

googleblog.com

ai.googleblog.com

huggingface.co

lambdalabs.com

microsoft.com

nature.com

openai.com

openai.com

cdn.openai.com

stanford.edu

web.stanford.edu

techcrunch.com

venturebeat.com

web.archive.org

yandex.ru