Tried this out again with v2 of the adapter training toolkit. Still the inference uses 26GB!!! And 40GB was used after one epoch. Really hope either there's improvements made to the toolkit or more clarity around the hardware specs needed for tuning the model
Topic:
Machine Learning & AI
SubTopic:
Foundation Models
Tags: