大型语言模型 (Chinese Wikipedia)

Analysis of information sources in references of the Wikipedia article "大型语言模型" in Chinese language version.

refsWebsite
Global rank Chinese rank
1st place
1st place
69th place
254th place
2nd place
23rd place
low place
low place
low place
low place
5th place
12th place
low place
low place
1,559th place
848th place
11th place
332nd place
18th place
57th place
415th place
500th place
low place
low place
low place
low place
274th place
320th place
low place
low place
14th place
18th place
179th place
275th place
low place
low place
7th place
31st place
187th place
481st place
low place
low place
3,464th place
5,489th place
652nd place
712th place
12th place
60th place
612th place
2,396th place
1,943rd place
2,036th place
2,263rd place
5,757th place
low place
low place
616th place
838th place
2,503rd place
1,088th place
low place
low place
low place
low place
8,920th place
7,729th place
2,218th place
5,303rd place
low place
low place
120th place
337th place
3,700th place
4,616th place
1,131st place
1,050th place
low place
low place
234th place
227th place
low place
low place
9th place
2nd place
1,185th place
809th place
low place
low place
1,497th place
54th place

aclanthology.org (Global: low place; Chinese: low place)

acm.org (Global: 1,185th place; Chinese: 809th place)

dl.acm.org

amacad.org (Global: 3,464th place; Chinese: 5,489th place)

anthropic.com (Global: low place; Chinese: low place)

archive.today (Global: 14th place; Chinese: 18th place)

arxiv.org (Global: 69th place; Chinese: 254th place)

  • Bommasani, Rishi; Hudson, Drew A.; Adeli, Ehsan; Altman, Russ; Arora, Simran; von Arx, Matthew; Bernstein, Michael S.; Bohg, Jeannette; Bosselut, Antoine; Brunskill, Emma. On the Opportunities and Risks of Foundation Models. 2021. arXiv:2108.07258可免费查阅. 
  • Brown, Tom B.; Mann, Benjamin; Ryder, Nick; Subbiah, Melanie; Kaplan, Jared; Dhariwal, Prafulla; Neelakantan, Arvind; Shyam, Pranav; Sastry, Girish; Askell, Amanda. Language Models are Few-Shot Learners. 2020. arXiv:2005.14165可免费查阅 [cs.CL]. 
  • Kaplan, Jared; McCandlish, Sam; Henighan, Tom; Brown, Tom B.; Chess, Benjamin; Child, Rewon; Gray, Scott; Radford, Alec; Wu, Jeffrey; Amodei, Dario. Scaling Laws for Neural Language Models. 2020. arXiv:2001.08361可免费查阅 [cs.LG]. 
  • Brown, Tom B.; Mann, Benjamin; Ryder, Nick; Subbiah, Melanie; Kaplan, Jared; Dhariwal, Prafulla; Neelakantan, Arvind; Shyam, Pranav; Sastry, Girish; Askell, Amanda; Agarwal, Sandhini; Herbert-Voss, Ariel; Krueger, Gretchen; Henighan, Tom; Child, Rewon; Ramesh, Aditya; Ziegler, Daniel M.; Wu, Jeffrey; Winter, Clemens; Hesse, Christopher; Chen, Mark; Sigler, Eric; Litwin, Mateusz; Gray, Scott; Chess, Benjamin; Clark, Jack; Berner, Christopher; McCandlish, Sam; Radford, Alec; Sutskever, Ilya; Amodei, Dario. Larochelle, H.; Ranzato, M.; Hadsell, R.; Balcan, M.F.; Lin, H. , 编. Language Models are Few-Shot Learners (PDF). Advances in Neural Information Processing Systems (Curran Associates, Inc.). Dec 2020, 33: 1877–1901 [2023-03-14]. arXiv:2005.14165可免费查阅. doi:10.1145/3582269.3615599. (原始内容存档 (PDF)于2023-11-17). 
  • Devlin, Jacob; Chang, Ming-Wei; Lee, Kenton; Toutanova, Kristina. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. 2018. arXiv:1810.04805可免费查阅 [cs.CL]. 
  • Goodman, Joshua, A Bit of Progress in Language Modeling, 2001-08-09, Bibcode:2001cs........8005G, arXiv:cs/0108005可免费查阅 
  • Bahdanau, Dzmitry; Cho, Kyunghyun; Bengio, Yoshua. Neural Machine Translation by Jointly Learning to Align and Translate. 2014. arXiv:1409.0473可免费查阅 [cs.CL]. 
  • Rogers, Anna; Kovaleva, Olga; Rumshisky, Anna. A Primer in BERTology: What We Know About How BERT Works. Transactions of the Association for Computational Linguistics. 2020, 8: 842–866 [2024-01-21]. S2CID 211532403. arXiv:2002.12327可免费查阅. doi:10.1162/tacl_a_00349. (原始内容存档于2022-04-03). 
  • Movva, Rajiv; Balachandar, Sidhika; Peng, Kenny; Agostini, Gabriel; Garg, Nikhil; Pierson, Emma. Topics, Authors, and Institutions in Large Language Model Research: Trends from 17K arXiv Papers. Proceedings of the 2024 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 1: Long Papers). 2024: 1223–1243 [2024-12-08]. arXiv:2307.10700可免费查阅. doi:10.18653/v1/2024.naacl-long.67. (原始内容存档于2025-04-12). 
  • Movva, Rajiv; Balachandar, Sidhika; Peng, Kenny; Agostini, Gabriel; Garg, Nikhil; Pierson, Emma. Topics, Authors, and Institutions in Large Language Model Research: Trends from 17K arXiv Papers. Proceedings of the 2024 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 1: Long Papers). 2024: 1223–1243 [2024-12-08]. arXiv:2307.10700可免费查阅. doi:10.18653/v1/2024.naacl-long.67. (原始内容存档于2025-04-12). 
  • Peng, Bo; et al. RWKV: Reinventing RNNS for the Transformer Era. 2023. arXiv:2305.13048可免费查阅 [cs.CL]. 
  • Gu, Albert; Dao, Tri, Mamba: Linear-Time Sequence Modeling with Selective State Spaces, 2023-12-01, arXiv:2312.00752可免费查阅 
  • Petrov, Aleksandar; Malfa, Emanuele La; Torr, Philip; Bibi, Adel. Language Model Tokenizers Introduce Unfairness Between Languages. NeurIPS. June 23, 2023 [September 16, 2023]. arXiv:2305.15425可免费查阅. (原始内容存档于December 15, 2023) –通过openreview.net. 
  • Kaushal, Ayush; Mahowald, Kyle, What do tokens know about their characters and how do they know it?, 2022-06-06, arXiv:2206.02608可免费查阅 
  • Petrov, Aleksandar; Emanuele La Malfa; Torr, Philip H. S.; Bibi, Adel. Language Model Tokenizers Introduce Unfairness Between Languages. 2023. arXiv:2305.15425可免费查阅 [cs.CL]. 
  • Dodge, Jesse; Sap, Maarten; Marasović, Ana; Agnew, William; Ilharco, Gabriel; Groeneveld, Dirk; Mitchell, Margaret; Gardner, Matt. Documenting Large Webtext Corpora: A Case Study on the Colossal Clean Crawled Corpus. 2021. arXiv:2104.08758可免费查阅 [cs.CL]. 
  • Li, Yuanzhi; Bubeck, Sébastien; Eldan, Ronen; Del Giorno, Allie; Gunasekar, Suriya; Lee, Yin Tat, Textbooks Are All You Need II: phi-1.5 technical report, 2023-09-11, arXiv:2309.05463可免费查阅 
  • Lin, Zhenghao; Gou, Zhibin; Gong, Yeyun; Liu, Xiao; Shen, Yelong; Xu, Ruochen; Lin, Chen; Yang, Yujiu; Jiao, Jian. Rho-1: Not All Tokens Are What You Need. 2024-04-11. arXiv:2404.07965可免费查阅 [cs.CL]. 
  • Brown, Tom B.; et al. Language Models are Few-Shot Learners. 2020. arXiv:2005.14165可免费查阅 [cs.CL]. 
  • Abdin, Marah; Jacobs, Sam Ade; Awan, Ammar Ahmad; Aneja, Jyoti; Awadallah, Ahmed; Awadalla, Hany; Bach, Nguyen; Bahree, Amit; Bakhtiari, Arash. Phi-3 Technical Report: A Highly Capable Language Model Locally on Your Phone. 2024-04-23. arXiv:2404.14219可免费查阅 [cs.CL]. 
  • Zaib, Munazza; Sheng, Quan Z.; Emma Zhang, Wei. A Short Survey of Pre-trained Language Models for Conversational AI-A New Age in NLP. Proceedings of the Australasian Computer Science Week Multiconference. 4 February 2020: 1–4. ISBN 9781450376976. S2CID 211040895. arXiv:2104.10810可免费查阅. doi:10.1145/3373017.3373028. 
  • Shazeer, Noam; Mirhoseini, Azalia; Maziarz, Krzysztof; Davis, Andy; Le, Quoc; Hinton, Geoffrey; Dean, Jeff. Outrageously Large Neural Networks: The Sparsely-Gated Mixture-of-Experts Layer. 2017-01-01. arXiv:1701.06538可免费查阅 [cs.LG]. 
  • Lepikhin, Dmitry; Lee, HyoukJoong; Xu, Yuanzhong; Chen, Dehao; Firat, Orhan; Huang, Yanping; Krikun, Maxim; Shazeer, Noam; Chen, Zhifeng. GShard: Scaling Giant Models with Conditional Computation and Automatic Sharding. 2021-01-12. arXiv:2006.16668可免费查阅 [cs.CL]. 
  • Lewis, Patrick; Perez, Ethan; Piktus, Aleksandra; Petroni, Fabio; Karpukhin, Vladimir; Goyal, Naman; Küttler, Heinrich; Lewis, Mike; Yih, Wen-tau; Rocktäschel, Tim; Riedel, Sebastian; Kiela, Douwe. Retrieval-Augmented Generation for Knowledge-Intensive NLP Tasks. Advances in Neural Information Processing Systems (Curran Associates, Inc.). 2020, 33: 9459–9474 [2023-06-12]. arXiv:2005.11401可免费查阅. (原始内容存档于2023-06-12). 
  • Ouyang, Long; Wu, Jeff; Jiang, Xu; Almeida, Diogo; Wainwright, Carroll L.; Mishkin, Pamela; Zhang, Chong; Agarwal, Sandhini; Slama, Katarina; Ray, Alex; Schulman, John; Hilton, Jacob; Kelton, Fraser; Miller, Luke; Simens, Maddie; Askell, Amanda; Welinder, Peter; Christiano, Paul; Leike, Jan; Lowe, Ryan. Training language models to follow instructions with human feedback. 2022. arXiv:2203.02155可免费查阅 [cs.CL]. 
  • Sharir, Or; Peleg, Barak; Shoham, Yoav. The Cost of Training NLP Models: A Concise Overview. 2020. arXiv:2004.08900可免费查阅 [cs.CL]. 
  • Biderman, Stella; Schoelkopf, Hailey; Anthony, Quentin; Bradley, Herbie; Khan, Mohammad Aflah; Purohit, Shivanshu; Prashanth, USVSN Sai. Pythia: A Suite for Analyzing Large Language Models Across Training and Scaling. April 2023. arXiv:2304.01373可免费查阅 [cs.CL]. 
  • Maslej, Nestor; Fattorini, Loredana; Brynjolfsson, Erik; Etchemendy, John; Ligett, Katrina; Lyons, Terah; Manyika, James; Ngo, Helen; Niebles, Juan Carlos, Artificial Intelligence Index Report 2023, 2023-10-05, arXiv:2310.03715可免费查阅 
  • Section 2.1 and Table 1, Kaplan, Jared; McCandlish, Sam; Henighan, Tom; Brown, Tom B.; Chess, Benjamin; Child, Rewon; Gray, Scott; Radford, Alec; Wu, Jeffrey; Amodei, Dario. Scaling Laws for Neural Language Models. 2020. arXiv:2001.08361可免费查阅 [cs.LG]. 
  • Li, Junnan; Li, Dongxu; Savarese, Silvio; Hoi, Steven. BLIP-2: Bootstrapping Language-Image Pre-training with Frozen Image Encoders and Large Language Models. 2023-01-01. arXiv:2301.12597可免费查阅 [cs.CV]. 
  • Alayrac, Jean-Baptiste; Donahue, Jeff; Luc, Pauline; Miech, Antoine; Barr, Iain; Hasson, Yana; Lenc, Karel; Mensch, Arthur; Millican, Katherine; Reynolds, Malcolm; Ring, Roman; Rutherford, Eliza; Cabi, Serkan; Han, Tengda; Gong, Zhitao. Flamingo: a Visual Language Model for Few-Shot Learning. Advances in Neural Information Processing Systems. 2022-12-06, 35: 23716–23736 [2023-07-02]. arXiv:2204.14198可免费查阅. (原始内容存档于2023-07-02). 
  • Liu, Haotian; Li, Chunyuan; Wu, Qingyang; Lee, Yong Jae. Visual Instruction Tuning. 2023-04-01. arXiv:2304.08485可免费查阅 [cs.CV]. 
  • Zhang, Hang; Li, Xin; Bing, Lidong. Video-LLaMA: An Instruction-tuned Audio-Visual Language Model for Video Understanding. 2023-06-01. arXiv:2306.02858可免费查阅 [cs.CL]. 
  • OpenAI. GPT-4 Technical Report. 2023-03-27. arXiv:2303.08774可免费查阅 [cs.CL]. 
  • Queenie Luo; Michael J. Puett; Michael D. Smith. A Perspectival Mirror of the Elephant: Investigating Language Bias on Google, ChatGPT, Wikipedia, and YouTube. arXiv. (原始内容存档于2024-04-16). 
  • Lei Huang; Weijiang Yu; Weitao Ma. A Survey on Hallucination in Large Language Models: Principles, Taxonomy, Challenges, and Open Questions. arXiv. (原始内容存档于2024-11-28). 

blog.google (Global: 2,218th place; Chinese: 5,303rd place)

doi.org (Global: 2nd place; Chinese: 23rd place)

doi.org

dx.doi.org

eswc-conferences.org (Global: low place; Chinese: low place)

2024.eswc-conferences.org

euronews.com (Global: 612th place; Chinese: 2,396th place)

harvard.edu (Global: 18th place; Chinese: 57th place)

ui.adsabs.harvard.edu

huggingface.co (Global: low place; Chinese: low place)

ibm.com (Global: 1,131st place; Chinese: 1,050th place)

ieee.org (Global: 652nd place; Chinese: 712th place)

ieeexplore.ieee.org

jalammar.github.io (Global: low place; Chinese: low place)

llmbook-zh.github.io (Global: low place; Chinese: low place)

  • 赵鑫,李军毅,周昆,唐天一,文继荣. 大语言模型. 北京: 高等教育出版社. 2024. 

meta.com (Global: low place; Chinese: low place)

ai.meta.com

mit.edu (Global: 415th place; Chinese: 500th place)

direct.mit.edu

mittrchina.com (Global: low place; Chinese: low place)

mlr.press (Global: low place; Chinese: low place)

proceedings.mlr.press

  • Nagel, Markus; Amjad, Rana Ali; Baalen, Mart Van; Louizos, Christos; Blankevoort, Tijmen. Up or Down? Adaptive Rounding for Post-Training Quantization. Proceedings of the 37th International Conference on Machine Learning (PMLR). 2020-11-21: 7197–7206 [2023-06-14]. (原始内容存档于2023-06-14). 
  • Kiros, Ryan; Salakhutdinov, Ruslan; Zemel, Rich. Multimodal Neural Language Models. Proceedings of the 31st International Conference on Machine Learning (PMLR). 2014-06-18: 595–603 [2023-07-02]. (原始内容存档于2023-07-02). 

nature.com (Global: 234th place; Chinese: 227th place)

neurips.cc (Global: low place; Chinese: low place)

proceedings.neurips.cc

  • Brown, Tom B.; Mann, Benjamin; Ryder, Nick; Subbiah, Melanie; Kaplan, Jared; Dhariwal, Prafulla; Neelakantan, Arvind; Shyam, Pranav; Sastry, Girish; Askell, Amanda; Agarwal, Sandhini; Herbert-Voss, Ariel; Krueger, Gretchen; Henighan, Tom; Child, Rewon; Ramesh, Aditya; Ziegler, Daniel M.; Wu, Jeffrey; Winter, Clemens; Hesse, Christopher; Chen, Mark; Sigler, Eric; Litwin, Mateusz; Gray, Scott; Chess, Benjamin; Clark, Jack; Berner, Christopher; McCandlish, Sam; Radford, Alec; Sutskever, Ilya; Amodei, Dario. Larochelle, H.; Ranzato, M.; Hadsell, R.; Balcan, M.F.; Lin, H. , 编. Language Models are Few-Shot Learners (PDF). Advances in Neural Information Processing Systems (Curran Associates, Inc.). Dec 2020, 33: 1877–1901 [2023-03-14]. arXiv:2005.14165可免费查阅. doi:10.1145/3582269.3615599. (原始内容存档 (PDF)于2023-11-17). 
  • Vaswani, Ashish; Shazeer, Noam; Parmar, Niki; Uszkoreit, Jakob; Jones, Llion; Gomez, Aidan N; Kaiser, Łukasz; Polosukhin, Illia. Attention is All you Need (PDF). Advances in Neural Information Processing Systems (Curran Associates, Inc.). 2017, 30 [2024-01-21]. (原始内容存档 (PDF)于2024-02-21). 
  • Lewis, Patrick; Perez, Ethan; Piktus, Aleksandra; Petroni, Fabio; Karpukhin, Vladimir; Goyal, Naman; Küttler, Heinrich; Lewis, Mike; Yih, Wen-tau; Rocktäschel, Tim; Riedel, Sebastian; Kiela, Douwe. Retrieval-Augmented Generation for Knowledge-Intensive NLP Tasks. Advances in Neural Information Processing Systems (Curran Associates, Inc.). 2020, 33: 9459–9474 [2023-06-12]. arXiv:2005.11401可免费查阅. (原始内容存档于2023-06-12). 
  • Krizhevsky, Alex; Sutskever, Ilya; Hinton, Geoffrey E. ImageNet Classification with Deep Convolutional Neural Networks. Advances in Neural Information Processing Systems (Curran Associates, Inc.). 2012, 25 [2023-07-02]. (原始内容存档于2023-07-02). 
  • Alayrac, Jean-Baptiste; Donahue, Jeff; Luc, Pauline; Miech, Antoine; Barr, Iain; Hasson, Yana; Lenc, Karel; Mensch, Arthur; Millican, Katherine; Reynolds, Malcolm; Ring, Roman; Rutherford, Eliza; Cabi, Serkan; Han, Tengda; Gong, Zhitao. Flamingo: a Visual Language Model for Few-Shot Learning. Advances in Neural Information Processing Systems. 2022-12-06, 35: 23716–23736 [2023-07-02]. arXiv:2204.14198可免费查阅. (原始内容存档于2023-07-02). 

nvidia.com (Global: 2,503rd place; Chinese: 1,088th place)

blogs.nvidia.com

nytimes.com (Global: 7th place; Chinese: 31st place)

openai.com (Global: 1,559th place; Chinese: 848th place)

platform.openai.com

openai.com

cdn.openai.com

openreview.net (Global: low place; Chinese: low place)

ourworldindata.org (Global: 2,263rd place; Chinese: 5,757th place)

researchgate.net (Global: 120th place; Chinese: 337th place)

rgdoi.net (Global: low place; Chinese: low place)

semanticscholar.org (Global: 11th place; Chinese: 332nd place)

api.semanticscholar.org

springer.com (Global: 274th place; Chinese: 320th place)

link.springer.com

stanford.edu (Global: 179th place; Chinese: 275th place)

web.stanford.edu

techcrunch.com (Global: 187th place; Chinese: 481st place)

technologyreview.com (Global: 1,943rd place; Chinese: 2,036th place)

thecvf.com (Global: low place; Chinese: low place)

openaccess.thecvf.com

  • Antol, Stanislaw; Agrawal, Aishwarya; Lu, Jiasen; Mitchell, Margaret; Batra, Dhruv; Zitnick, C. Lawrence; Parikh, Devi. VQA: Visual Question Answering. ICCV. 2015: 2425–2433 [2023-07-02]. (原始内容存档于2023-07-02). 

theguardian.com (Global: 12th place; Chinese: 60th place)

thepaper.cn (Global: 1,497th place; Chinese: 54th place)

theregister.com (Global: 3,700th place; Chinese: 4,616th place)

towardsdatascience.com (Global: 8,920th place; Chinese: 7,729th place)

unite.ai (Global: low place; Chinese: low place)

venturebeat.com (Global: 616th place; Chinese: 838th place)

web.archive.org (Global: 1st place; Chinese: 1st place)

worldcat.org (Global: 5th place; Chinese: 12th place)

yenniejun.com (Global: low place; Chinese: low place)

blog.yenniejun.com

youtube.com (Global: 9th place; Chinese: 2nd place)