Preview

Key Issues of Contemporary Linguistics

Advanced search

Modeling linguo-creative strategies in generative language systems: transformation of intentional deviations into discursive patterns

https://doi.org/10.18384/2949-5075-2025-4-6-15

Abstract

Aim. The present research aims to identify the mechanisms of processing intensional linguistic deviations by large language models (LLMs) and to analyse their linguocreative strategies in digital discourse. The aim of the work is to develop a theoretical model explaining cognitive algorithms for recognising and transforming deviations in artificial intelligence systems.
Methodology. An integrated approach including corpus analysis of a diachronic slice of digital discourse (2019–2024), experimental prompts with controlled deviations for GPT-4, Gemini 1.5 and Claude 3 models, and discourse analysis of AI speech acts using a three-level annotation scale (replication/amplification/normalisation) was used as a methodological framework.
Results. The results of the research supported the hypothesis of the statistical nature of LLM linguocreativity, revealing a three-stage model of deviation processing: recognition through attention mechanisms, classification by degree of deviation from the norm, and strategic choice of response. The paradox of ‘creative conformism’ is established, manifested in AI's tendency to hypernormalise initially marginal linguistic innovations. Of particular practical interest are the documented effects of circulation of AI-generated neologisms in social media and the formation of ‘artificial linguistic taste’.
Research implications. The theoretical significance of the work lies in the development of the apparatus of cognitive linguistics of digital discourse and clarification of the ontology of intensional deviations. Practical value is related to applications in the field of NLP-systems development, digital linguodidactics and prediction of language changes. The data obtained opens prospects for further studying culturally specific deviations in multilingual models and developing metrics for assessing the linguocreative potential of AI.

About the Author

O. M. Akay
Saint-Petersburg State University
Russian Federation

Oksana M. Akay (St.  Petersburg) – Dr.  Sci. (Philology), Prof., Department of Foreign Languages in Economics and Law



References

1. Dürscheid, C. (2024). Grammatische und lexikalische Strukturen digital geschriebener Sprache. In: Androutsopoulos, J. & Vogel, F., eds. Handbuch Sprache und digitale Kommunikation. Berllin – Boston: De Gruyter, pp. 157–175. DOI: 10.1515/9783110744163-008.

2. Seargeant, P. & Tagg, C., eds. (2014). The language of social media: Identity and community on the Internet. New York: Palgrave Macmillan.

3. Zappavigna, M. & Logi, L. (2024). Emoji and social media paralanguage. Cambridge: Cambridge University Press.

4. Androutsopoulos, J. & Vogel, F., eds. (2024). Handbuch Sprache und digitale Kommunikation. Berllin – Boston: De Gruyter.

5. Tagliamonte, S. A. (2016). Teen talk: The language of adolescents. Cambridge: Cambridge University Press.

6. McCulloch, G. (2019). Because internet: Understanding the new rules of language. New York: Riverhead Books.

7. Thurlow, C. (2018). Digital Discourse: Locating Language in New/Social Media. In: Burgess, J., Poell, T. & Marwick, A., eds. The SAGE Handbook of Social Media. New York: Sage, pp. 135–145. DOI: 10.4135/9781473984066.n8.

8. Bender, E. M., Gebru, T., McMillan-Major, A. & Shmitchell, S. (2021). On the dangers of stochastic parrots: Can language models be too big? In: FAccT’ 21: Proceedings of the 2021 ACM Conference on Fairness, Accountability, and Transparency. New York: Association for Computing Machinery, pp. 610–623. DOI: 10.1145/3442188.3445922.

9. Bommasani, R., Hudson, D. A., Adeli, E., Altman, R., Arora, S., von Arx, S., Bernstein, M. S. et al. (2021). On the opportunities and risks of foundation models. In: ArXiv. URL: https://arxiv.org/abs/2108.07258 (accessed: 07.03.2025). DOI: 10.48550/arXiv.2108.07258.

10. Devlin, J., Chang, M.-W., Lee, K. & Toutanova, K. (2019). BERT: Pre-training of deep bidirectional transformers for language understanding. In: Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Vol. 1 (Long and Short Papers). Minneapolis, Minnesota: Association for Computational Linguistics, pp. 4171–4186. DOI: 10.18653/v1/N19-1423.

11. Bender, E. M. & Koller, A. (2020). Climbing towards NLU: On meaning, form, and understanding in the age of data. In: Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics. Minneapolis, Minnesota: Association for Computational Linguistics, pp. 5185–5198. DOI: 10.18653/v1/2020.acl-main.463.

12. Marcus, G. (2018). Deep Learning: A Critical Appraisal. In: ArXiv. URL: https://arxiv.org/abs/1801.00631 (accessed: 07.03.2025). DOI: 10.48550/arXiv.1801.00631.

13. Kovaleva, O., Romanov, A., Rogers, A. & Rumshisky ,A. (2019). Revealing the dark secrets of BERT. In: Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP). Hong Kong, China. Association for Computational Linguistics, pp. 4365–4374.

14. Acharjee, S., Aich, U. & Ali, A. (2025). Does Language Model Understand Language? In: ArXiv. URL: https://arxiv.org/abs/2509.12459 (accessed: 03.03.2025).

15. Van Hout, T. (2015). Book review: Jannis Androutsopoulos (ed.). Mediatization and Sociolinguistic Change (Linguae & Litterae 36). In: Journal of Sociolinguistics, 19 (5), 714–718. DOI: 10.1111/josl.12163.


Review

Views: 6


Creative Commons License
This work is licensed under a Creative Commons Attribution 4.0 License.


ISSN 2949-5059 (Print)
ISSN 2949-5075 (Online)