Litespark Inference on Consumer CPUs: Custom SIMD Kernels for Ternary Neural Networks
概要
arXiv:2605.06485v1 Announce Type: cross Abstract: Large language models (LLMs) have transformed artificial intelligence, but their computational requirements remain prohibitive for most users. Standard inference demands expensive datacenter GPUs or cloud API access, leaving over one billion persona…