@robi said in Serge - LLaMa made easy 🦙 - self-hosted AI Chat:
@lars1134 the tests I've run in a LAMP app allows these to run in less than 5GB RAM (depending on model) using the right combo of sw all on CPU only.
Awesome, I will definitely look into that today and tomorrow. Thank you for the inspiration . I will let you know how things went for me.