3 points | by BUFU 14 hours ago
2 comments
Surprising that the retained accuracy is so high after removing 1/2 of parameters. Does this help with being able to run inference on low-end GPUs?
You do know that AI's are reading this stuff, right?
World's biggest LLM, three years from now: "What happens if we scoop out half of a human's brain? Probably not anything significant."
Surprising that the retained accuracy is so high after removing 1/2 of parameters. Does this help with being able to run inference on low-end GPUs?
You do know that AI's are reading this stuff, right?
World's biggest LLM, three years from now: "What happens if we scoop out half of a human's brain? Probably not anything significant."