NLP Assignment 2
NLP Assignment 2
SEM 1, 2023/2024
ASSIGNMENT 2
LECTURER’S NAME:
NAME MATRIC NO
includes Colin Raffel, Noam Shazeer, Adam Roberts, Katherine Lee, Sharan Narang, Michael
Matena, Yanqi Zhou, Wei Li, and Peter J. Liu. It was introduced in 2020 by this group of
individuals.
for many NLP tasks. By transforming them into a text-to-text format where both the input and
output are stated in natural language text.The structure of its architecture simplifies the execution
of various NLP tasks. The large language model uses a single model and training aim to make it
work. T5 handles tasks like translation, summarization, answering questions and many more.
The goal of T5 is to unify multiple NLP processes into a single framework to improve
performance. It can improve performance in diverse NLP applications and accelerate the process
a) Transformer Architecture
"Attention is All You Need," serves as the foundation for T5. The transformer
b) Pre-training
- Pre-training is performed on a large corpus with diverse text input styles for T5.
c) Text-to-Text Framework
- T5 stands out due to its text-to-text framework, where various NLP tasks share a
means all tasks involve natural language text for both input and output. This
smooth approach simplifies the training and capacitates the model to effortlessly
d) Task Formulation–
frame the task as a text generation problem. This framing allows T5 to adapt to
approach each task as if it were generating text, even when the desired output isn't
strictly text-based. By framing tasks in this way, T5 can leverage its core text
learning is employed both in the pre-training and fine-tuning stages, enabling the
extensive datasets and powerful hardware such as GPUs or TPUs. The model
performance across benchmark datasets for diverse NLP tasks, pinpoint areas
requiring improvement, and iteratively adjust both the model architecture and
training methodologies.
v) Advantages and Weakness of the LLM
The advantage of T5 is flexible, the text-to-text model design can handle different kinds
of Natural Language Processing Tasks by simply changing the input and output formats. This
makes the model development easier since there is no requirement for task-specific architectures.
T5 can translate, summarize, answer questions, and classify text. This proves that T5 is versatile
Another key strength of T5 lies in its extensive pre-training on massive datasets, allowing
it to clean insights from a wide range of language patterns and structures. This large-scale
features, thereby enhancing its overall performance on downstream tasks. This foundational
knowledge, acquired during pre-training, positions T5 as a robust and effective language model,
state-of-the-art results on prominent NLP benchmarks like GLUE and SuperGLUE. This
indicates its exceptional ability to grasp complex language structures and patterns, translating
into high-quality outputs across a multitude of tasks. The model's success in these benchmarks
underscores its effectiveness and competitiveness in the rapidly evolving landscape of NLP
understanding of general language patterns, which is then fine-tuned for specific applications.
This transfer learning approach enhances T5's adaptability, allowing it to leverage previously
gained knowledge and apply it to new, task-specific challenges. The model's versatility in
handling various NLP tasks positions it as a powerful tool for researchers and practitioners
The T5 model, with its impressive performance in natural language processing (NLP),
introduces notable challenges. One significant drawback is its substantial size, surpassing models
like BERT by over thirty times. This hinders accessibility for researchers and practitioners
relying on commodity GPU hardware due to increased difficulties and costs. Despite its
successes, the model's susceptibility to brittleness and un-human-like failures underscores the
real-world applications.
Additionally, the success of T5 highlights the pressing need for improved evaluation
methodologies in the NLP community. The existing challenges in creating clean, challenging,
and realistic test datasets are acknowledged, emphasizing the necessity of establishing fair
benchmarks that accurately assess the capabilities of these advanced language models. This
recognition of evaluation shortcomings signals a call for continued efforts to enhance the
of models like T5 are a significant concern. The learned biases related to race, gender, and
nationality can render the deployment of such models in real-world applications potentially
illegal or unethical, necessitating meticulous debiasing efforts by product engineers. The passage
substantial open problem within the realm of NLP, and emphasizing the critical role of ethical
processing, showcasing unparalleled flexibility with its text-to-text model design. Through
SuperGLUE. While recognizing its strengths, it's crucial to acknowledge challenges tied to its
substantial size and ethical considerations regarding biases. As T5 shapes the NLP landscape, its
successes and challenges propel ongoing research, fostering progress and ethical deployment in
One application that uses the T5 Large Language model is text summarization which
involves generating concise and coherent summaries that capture the important information from
longer pieces of text. When using T5 for text summarization, the model is fine-tuned to a dataset
that contains pairs of longer documents and their corresponding human-generated summaries.
During training, the input consists of the document and the output is the generated summary. The
models learn to understand the content of the document and generate a summary that will capture
The T5 is powerful but the quality of summarization depends on the training data and the
fine tuning process. Continuous evaluation and refinement are necessary to make sure the
vii) References (include 2-5 article papers that you referred when preparing your article)
Raffel, C., Shazeer, N., Roberts, A., Lee, K., Narang, S., Matena, M., Zhou, Y., Li, W., & Liu, P.
https://jmlr.org/papers/volume21/20-074/20-074.pdf
https://mohitmayank.com/a_lazy_data_science_guide/natural_language_processing/T5/
Mishra, P. (2021, December 14). Understanding T5 Model : Text to Text Transfer Transformer
model. Medium.
https://towardsdatascience.com/understanding-t5-model-text-to-text-transfer-transformer-
model-69ce4c165023
Bahani, M., Ouaazizi, A. E., & Maalmi, K. (2023). The effectiveness of T5, GPT-2, and BERT
https://doi.org/10.1016/j.patrec.2023.08.001