Fine tune t5-small for text summarization
I am trying to fine tune t5-small for text summarization, and I have the following graph loss per batch:
and learning rate per batch:
Do you think loss graph is normal (regarding this use case) or there is a problem
Topic transformer automatic-summarization nlp
Category Data Science