Vol. 4 No. 1 (2024): Journal of Machine Learning in Pharmaceutical Research
Articles

Advancements in Transfer Learning for Natural Language Processing Tasks

Dr. Olga Sokolova
Professor, Moscow State University, Ulitsa Kolmogorova, Moscow, Russia
Prof. Dmitri Volkov
Professor, Moscow State University, Ulitsa Kolmogorova, Moscow, Russia
Prof. Natasha Ivanova
Professor, Moscow State University, Ulitsa Kolmogorova, Moscow, Russia
Prof. Pavel Morozov
Professor, Moscow State University, Ulitsa Kolmogorova, Moscow, Russia
Cover

Published 12-04-2024

Keywords

  • BERT transformer,
  • Bidirectional Encoder Representations from Transformers

How to Cite

[1]
Dr. Olga Sokolova, Prof. Dmitri Volkov, Prof. Natasha Ivanova, and Prof. Pavel Morozov, “Advancements in Transfer Learning for Natural Language Processing Tasks”, Journal of Machine Learning in Pharmaceutical Research, vol. 4, no. 1, pp. 50–59, Apr. 2024, Accessed: Sep. 17, 2024. [Online]. Available: https://pharmapub.org/index.php/jmlpr/article/view/13

Abstract

Transfer learning is a new craze in the field of natural language processing, not just for beating task-specific performance records in many natural language processing tasks by huge computation and parameter settings with extensive training, but also with so much of model reusability. The transformer's rise, beginning with the attention is all you need paper by Vaswani et al. with the introduction of BERT transformer (Bidirectional Encoder Representations from Transformers) by the Google research team, has been spectacular. BERT has been designed to pre-train deep bidirectional representations from unlabelled text by jointly conditioning on left and right context in all layers and has subsequently paved way for many transformer-based models since November 2018.

 

Transfer learning is the new craze in modern natural language processing applications. Established by the concept of transfer learning, with more data and training time, much larger models have been built pre-trained on a large amount of training examples. These models are then fine-tuned with much smaller datasets for different linguistic tasks and applications. As an outcome of this modern trend, much better performance on the fine-tuning tasks has been observed. In this chapter, we look at various training paradigms of transfer learning. We explore a few fine-tuning methods and list several state-of-the-art results of different linguistic tasks. Finally, we talk about deployment in a real-world or for commercial natural language processing tasks.

Downloads

Download data is not yet available.

References

  1. Pulimamidi, Rahul. "To enhance customer (or patient) experience based on IoT analytical study through technology (IT) transformation for E-healthcare." Measurement: Sensors (2024): 101087.
  2. Pargaonkar, Shravan. "The Crucial Role of Inspection in Software Quality Assurance." Journal of Science & Technology 2.1 (2021): 70-77.
  3. Menaga, D., Loknath Sai Ambati, and Giridhar Reddy Bojja. "Optimal trained long short-term memory for opinion mining: a hybrid semantic knowledgebase approach." International Journal of Intelligent Robotics and Applications 7.1 (2023): 119-133.
  4. Singh, Amarjeet, and Alok Aggarwal. "Securing Microservices using OKTA in Cloud Environment: Implementation Strategies and Best Practices." Journal of Science & Technology 4.1 (2023): 11-39.
  5. Singh, Vinay, et al. "Improving Business Deliveries for Micro-services-based Systems using CI/CD and Jenkins." Journal of Mines, Metals & Fuels 71.4 (2023).
  6. Reddy, Surendranadha Reddy Byrapu. "Predictive Analytics in Customer Relationship Management: Utilizing Big Data and AI to Drive Personalized Marketing Strategies." Australian Journal of Machine Learning Research & Applications 1.1 (2021): 1-12.
  7. Thunki, Praveen, et al. "Explainable AI in Data Science-Enhancing Model Interpretability and Transparency." African Journal of Artificial Intelligence and Sustainable Development 1.1 (2021): 1-8.
  8. Reddy, Surendranadha Reddy Byrapu. "Ethical Considerations in AI and Data Science-Addressing Bias, Privacy, and Fairness." Australian Journal of Machine Learning Research & Applications 2.1 (2022): 1-12.