@Opsroller While that may be true, this is a 7B parameter model. 512GB is overkill for a 7B model. To fine-tune a 7B model with LORA, you should be able to get this done with 64GB RAM at most. Even on the VM I spun up, there were instances where my VM ran out of memory. I believe there's some issue with the sample code provided by Apple and potentially a memory leak that's causing excessive memory usage.
Topic:
Machine Learning & AI
SubTopic:
Foundation Models
Tags: