Architecture Teardown: How Meta Trains LLMs for Code Generation on 100k GPU Clusters
In Q3 2024, Meta trained a 70B parameter code-specialized LLM on 100,000 Nvidia H100 GPUs, achieving 214 TFLOPS per GPU and 92% cluster utilization – …
Latest Web news from Tech News
In Q3 2024, Meta trained a 70B parameter code-specialized LLM on 100,000 Nvidia H100 GPUs, achieving 214 TFLOPS per GPU and 92% cluster utilization – …
The unwinding of Meta’s deal shows how tech founders struggle to cut China ties.
I have been writing notes for years. Architecture decisions, debugging war stories, things I wish someone told me before I shipped that thing to produ…
The Invisible Hand of the State: How Government Coercion is Rewriting the First Amendment in Silicon Valley Imagine a world where the President of the…
The 20,000th user was siru by invitation from bbbhltz . You may see all 20,000+ users and the invite tree by opening https://lobste.rs/users , but it …
I debug production systems for fun. I've traced memory leaks at 2am. I've stared into the void of a segfault and the void stared back. I once spent 6 …