提高认知才能开阔眼界

认人待物,不断学习,提高商业和经济方敏锐的洞察力
个人资料
正文

三国: ASICs for AI (TPU)对GPU

(2025-12-14 09:58:02) 下一个

https://www.cnbc.com/2025/11/21/nvidia-gpus-google-tpus-aws-trainium-comparing-the-top-ai-chips.html

Custom ASICs, or application-specific integrated circuits, are now being designed by all the major hyperscalers, from?Google's?TPU to?Amazon's?Trainium and?OpenAI's?plans with?Broadcom. These chips are smaller, cheaper, accessible and could reduce these companies' reliance on Nvidia GPUs. Daniel Newman of the Futurum Group told CNBC that he sees custom ASICs "growing even faster than the GPU market over the next few years."

?

Besides GPUs and ASICs, there are also field-programmable gate arrays, which can be reconfigured with software after they're made for use in all sorts of applications, like signal processing, networking and AI. There's also an entire group of AI chips that power AI on devices rather than in the cloud.?Qualcomm,?Apple?and others have championed those on-device AI chips.

?

?

Google TPUs (Tensor Processing Units)
  • Strengths:?Extremely efficient for large-scale training inference of models like Gemini, using systolic arrays for massive matrix multiplication. Excellent cost-performance (e.g., 4x better for inference). Tightly integrated with Google's network for massive scaling.
  • Weaknesses:?Less flexible; designed for specific AI workloads, not general-purpose computing or HPC.
  • Best For:?Google's internal services (Search, YouTube), large model training, inference at massive scale.?
?
AWS (Trainium Inferentia)
  • Strengths:?Custom silicon (Trainium for training, Inferentia for inference) designed for performance/cost optimization in AWS, offering better efficiency than GPUs for many cloud workloads.
  • Weaknesses:?Like TPUs, less flexible than GPUs for novel research.
  • Best For:?AWS customers needing cost-effective, scalable AI compute within the AWS ecosystem.?
?
NVIDIA GPUs (e.g., H100)
  • Strengths:?Unmatched flexibility, broad software support (CUDA), runs on-prem/cloud/edge, ideal for RD, diverse models, and staying at the research frontier. The standard for most AI breakthroughs.
  • Weaknesses:?Higher power consumption and cost for highly specific, large-scale tasks where ASICs excel.
  • Best For:?General AI development, novel model architectures, hybrid cloud/on-prem deployments, research.?
[ 打印 ]
评论
目前还没有任何评论
登录后才可评论.