Until now, AI services based on large language models (LLMs) have mostly relied on expensive data center GPUs. This has ...