How to version data science projects with large files

I am working on a project with large data files (~300MB). I want to version my work along with the data files so that it is always available online. I tried using git-lfs but it has a 1GB/month bandwidth limit, beyond which you're blocked for a month.

What are versioning tools using by data scientists for projects that use 100MB data files (both static and generated)?

Topic version-control

Category Data Science


I have used dvc. It has data versioning though I do not use that often. I tend to use the makefile (pipeline) features.

About

Geeks Mental is a community that publishes articles and tutorials about Web, Android, Data Science, new techniques and Linux security.