Episode Details
Back to Episodes
📅 ThursdAI - May 30 - 1000 T/s inference w/ SambaNova, <135ms TTS with Cartesia, SEAL leaderboard from Scale & more AI news
Description
Hey everyone, Alex here!
Can you believe it's already end of May? And that 2 huge AI companies conferences are behind us (Google IO, MSFT Build) and Apple's WWDC is just ahead in 10 days! Exciting!
I was really looking forward to today's show, had quite a few guests today, I'll add all their socials below the TL;DR so please give them a follow and if you're only in reading mode of the newsletter, why don't you give the podcast a try 🙂 It's impossible for me to add the density of knowledge that's being shared on stage for 2 hours here in the newsletter!
Also, before we dive in, I’m hosting a free workshop soon, about building evaluations from scratch, if you’re building anything with LLMs in production, more than welcome to join us on June 12th (it’ll be virtual)
TL;DR of all topics covered:
* Open Source LLMs
* Mistral open weights Codestral - 22B dense coding model (X, Blog)
* Nvidia open sources NV-Embed-v1 - Mistral based SOTA embeddings (X, HF)
* HuggingFace Chat with tool support (X, demo)
* Aider beats SOTA on Swe-Bench with 26% (X, Blog, Github)
* OpenChat - Sota finetune of Llama3 (X, HF, Try It)
* LLM 360 - K2 65B - fully transparent and reproducible (X, Paper, HF, WandB)
* Big CO LLMs + APIs
* Scale announces SEAL Leaderboards - with private Evals (X, leaderboard)
* SambaNova achieves >1000T/s on Llama-3 full precision
* Groq hits back with breaking 1200T/s on Llama-3
* Anthropic tool support in GA (X, Blogpost)
* OpenAI adds GPT4o, Web Search, Vision, Code Interpreter & more to free users (X)
* Google Gemini & Gemini Flash are topping the evals leaderboards, in GA(X)
* Gemini Flash finetuning coming soon
* This weeks Buzz (What I learned at WandB this week)
* Sponsored a Mistral hackathon in Paris
* We have an upcoming workshop in 2 parts - come learn with me
* Vision & Video
* LLama3-V - Sota OSS VLM (X,
Listen Now
Love PodBriefly?
If you like Podbriefly.com, please consider donating to support the ongoing development.
Support Us