Breaking News

AI researchers found a way to run LLMs at a lightbulb-esque 13 watts with no loss in performance

Elimination of matrix multiplication from LLM processing can massively increase performance-per-watt with the correct optimizations, researchers from UC Santa Cruz demonstrate. It remains to be seen how applicable this approach is for AI in general.

Go to Source
Author: