Affiliation:
1. Pyatigorsk State University
Abstract
Today, large language models are very powerful, informational and analytical tools that significantly accelerate most of the existing methods and methodologies for processing informational processes. Scientific information is of particular importance in this capacity, which gradually involves the power of large language models. This interaction of science and qualitative new opportunities for working with information lead us to new, unique scientific discoveries, their great quantitative diversity. There is an acceleration of scientific research, a reduction in the time spent on its implementation – the freed up time can be spent both on solving new scientific problems and on scientific creativity, which, although it may not necessarily lead to a specific solution to a particular scientific problem, but is able to demonstrate the beauty of science in various disciplinary areas. As a result, the interaction of large language models and scientific information is at the same time a research for solutions to scientific problems, scientific problems, and scientific creativity. Solving scientific problems requires the ability to efficiently process big data, which cannot be done without an effective method – one of the significant methods was the Transformer architecture, introduced in 2017 and comprehensively integrated into the GPT‑3 model, which, as of September 2020, was the largest and most advanced language model in the world. Therefore, GPT‑3 can be called the basis of most scientific developments carried out in the context of using large language models. The interaction of science and large language models has become a factor in the emergence of a large number of questions, among which are: «Is the result of data analysis new knowledge?», «What are the prospects for scientific creativity in the era of big computing?». Currently, these issues are extremely important, because they allow us to develop the foundations for effective human‑computer interaction. Therefore, this study analyzes the issues presented.
Publisher
Pyatigorsk State University
Reference17 articles.
1. Romera‑Paredes, B., Barekatain, M., Novikov, A. et al. (2024) Mathematical discoveries from program search with large language models. Nature 625, P. 468‑475.
2. Trinh, T.H., Wu, Y., Le, Q.V. et al. (2024) Solving Olympiad geometry without human demonstrations. Nature 625, P. 476‑482.
3. Gonthier, G. et al. (2013) A Machine‑Checked Proof of the Odd Order Theorem. In: Blazy, S., Paulin‑Mohring, C., Pichardie, D. (eds) Interactive Theorem Proving. ITP 2013. Lecture Notes in Computer Science, vol. 7998. Springer, Berlin, Heidelberg.
4. Iten, R., Metger, T., Wilming, H., del Rio L., Renner R. (2020) Discovering Physical Concepts with Neural Networks. Phys. Rev. Lett. Vol. 124, P. 1‑6.
5. Fawzi, A., Balog, M., Huang, A. et al. (2022) Discovering faster matrix multiplication algorithms with reinforcement learning. Nature 610, P. 47‑53.