LLM optimization
AirLLM: Running 70B Parameter Models on Consumer Laptops
A new library called AirLLM enables running massive 70B parameter AI models on old laptops with limited RAM by processing layers sequentially rather than loading entire models into memory.