Funder
Andrew W. Mellon Foundation
Open Philanthropy Project
Long-Term Future Fund
Publisher
Springer Science and Business Media LLC
Reference56 articles.
1. Alain, G., & Bengio, Y. (2016). Understanding intermediate layers using linear classifier probes. arXiv:1610.01644.
2. Azaria, A., & Mitchell, T. (2023). The internal state of an llm knows when its lying.
3. Beery, S., van Horn, G., & Perona, P. (2018). Recognition in terra incognita.
4. Bender, E. M., Gebru, T., McMillan-Major, A., & Shmitchell, S. (2021). On the dangers of stochastic parrots: Can language models be too big? In Proceedings of the 2021 ACM conference on fairness, accountability, and transparency (pp. 610–623).
5. Bender, E. M., & Koller, A. (2020). Climbing towards nlu: On meaning, form, and understanding in the age of data. In Proceedings of the 58th annual meeting of the association for computational linguistics (pp. 5185–5198).
Cited by
3 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献
1. ChatGPT is bullshit;Ethics and Information Technology;2024-06
2. ChatGPT or Bard: Who is a better Certified Ethical Hacker?;Computers & Security;2024-05
3. Love the Way You Lie: Unmasking the Deceptions of LLMs;2023 IEEE 23rd International Conference on Software Quality, Reliability, and Security Companion (QRS-C);2023-10-22