Because Cerebras’ hardware focuses on AI training, it previously didn’t have an answer for customers’ AI inferencing needs. Now it does thanks to a new partnership with Qualcomm.
The two companies today said they have collaborated, so that the models trained on Cerebras’ hardware are optimized to run inferencing on Qualcomm’s Cloud A100 Ultra accelerator.
“They optimized the output of the big CS-3 machines to run really well on these very low-cost, low-power Qualcomm AI inferencing engines,” Freund said.