Nvidia's closest rival once again obliterates cloud giants in AI performance; Cerebras Inference is 75x faster than AWS, 32x faster than Google on Llama 3.1 405B

Cerebras hits 969 tokens/second on Llama 3.1 405B, 75x faster than AWSClaims industry-low 240ms latency, twice as fast as Google VertexCerebras Inference runs on the CS-3 with the WSE-3 AI…

Continue Reading Nvidia's closest rival once again obliterates cloud giants in AI performance; Cerebras Inference is 75x faster than AWS, 32x faster than Google on Llama 3.1 405B

OpenAI spent $80M to $100M training GPT-4; Chinese firm claims it trained its rival AI model for $3 million using just 2,000 GPUs

01.ai trained an AI model for $3 million using 2000 unnamed GPUS “Efficient engineering” allows 01.ai to compete globally, company claims 01.ai reduced inference costs to 10 cents per million…

Continue Reading OpenAI spent $80M to $100M training GPT-4; Chinese firm claims it trained its rival AI model for $3 million using just 2,000 GPUs

Photonic processor could enable ultrafast AI computations with extreme energy efficiency

The deep neural network models that power today’s most demanding machine-learning applications have grown so large and complex that they are pushing the limits of traditional electronic computing hardware.Photonic hardware,…

Continue Reading Photonic processor could enable ultrafast AI computations with extreme energy efficiency