Fine tune t5-small for text summarization

I am trying to fine tune t5-small for text summarization, and I have the following graph loss per batch:

and learning rate per batch:

Do you think loss graph is normal (regarding this use case) or there is a problem

Topic transformer automatic-summarization nlp

Category Data Science

About

Geeks Mental is a community that publishes articles and tutorials about Web, Android, Data Science, new techniques and Linux security.