Episode Details
Back to Episodes
📅 ThursdAI Feb 22nd - Groq near instant LLM calls, SDXL Lightning near instant SDXL, Google gives us GEMMA open weights and refuses to draw white people, Stability announces SD3 & more AI news
Description
Hey, this is Alex,
Ok let's start with the big news, holy crap this week was a breakthrough week for speed!
We had both Groq explode in popularity, and ByteDance release an updated SDXL model called Lightning, able to generate full blown SDXL 1024 images in 300ms.
I've been excited about seeing what real time LLM/Diffusion can bring, and with both of these news release the same week, I just had to go and test them out together:
Additionally, we had Google step into a big open weights role, and give us Gemma, 2 open weights models 2B and 7B (which is closer to 9B per Junyang) and it was great to see google committing to releasing at least some models in the open.
We also had breaking news, Emad from Stability announced SD3, which looks really great, Google to pay Reddit 200M for AI training on their data & a few more things.
TL;DR of all topics covered:
* Big CO LLMs + APIs
* Groq custom LPU inference does 400T/s Llama/Mistral generation (X, Demo)
* Google image generation is in Hot Waters and was reportedly paused (refuses to generate white people)
* Gemini 1.5 long context is very impressive to folks (Matt Shumer, Ethan Mollick)
* Open Weights LLMs
* Google releases GEMMA, open weights 2B and 7B models (Announcement, Models)
* Teknium releases Nous Hermes DPO (Announcement, HF)
* Vision & Video
* YoLo V9 - SOTA real time object detector is out (Announcement, Code)
* This weeks Buzz (What I learned in WandB this week)
* Went to SF to cohost an event with A16Z, Nous, Mistral (Thread, My Report)
* AI Art & Diffusion & 3D
* ByteDance presents SDXL-Lightning (Try here, Model)
* Stability announces Stable Diffusion 3 (Announcement)
* Tools
* Replit releases a new experimental Figma plugin for UI → Code (Announcement)
* Arc browser adds "AI pinch to understand" summarization (Announcement)
Big CO LLMs + APIs
Groq's new LPU show extreme performance for LLMs - up to 400T/s (example)
* Groq created a novel processing unit known as the Tensor Streaming Processor (TSP) which they categorize as a Linear Processor Unit (LPU). Unlike traditional GPUs that are parallel processors with hundreds of cores designed for graphics rendering, LPUs are architected to deliver deterministic perf