← All creators
S
Sebastian Raschka
77 recommendations
Products they use or recommend
All recommendations
"Even back when I was a grad student, I was in a lab doing biophysical simulations, molecular dynamics, and we had a Tesla GPU back then just for the computations. It was about 15 years ago now."
"I try Claude Code on the web every three to six months, which is just prompting a model to make an update to some GitHub repository that I have"
"We had a Tesla GPU back then just for the computations. It was about 15 years ago now"
"Sometimes for pastime I play video games, like I like- Video games with puzzles, like Zelda and Metroid."
"Sometimes for pastime I play video games, like I like- Video games with puzzles, like Zelda and Metroid."
"So, I use the Codeium plugin for VS Code."
"So, I use the Codeium plugin for VS Code. You know, it's very convenient. It's just like a plugin, and then it's a chat interface that has access to your repository."
"What I would recommend doing, or what I also do, is if I want to understand, for example, how OLMo is implemented, I would look at the weights in the model hub, the config file, and then you can see, 'Oh, they used so many layers.'"
"I can give you also a hands-on example. I was training the Qwen 3 base model with RLVR on MATH-500. The base model had an accuracy of about 15%. Just 50 steps, like in a few minutes with RLVR, the model went from 15% to 50% accuracy."
"Sometimes it takes me a day. With OLMo 3, the challenge was RoPE for the position embeddings. They had a YaRN extension and there was some custom scaling there, and I couldn't quite match these things."
"I was training the Qwen 3 base model with RLVR on MATH-500. The base model had an accuracy of about 15%."
"I used that feature before, and I always feel bad because it does that every day, and I rarely check it out"
"Exa is my preferred search provider"
"I should say I use Composer a lot because one of the benefits it has is that it's fast"
"The Recursive Language Model paper, that is one of the papers that tries to kind of address the long context thing"
"So I suggested, 'Hey, let's try ChatGPT.' We copied the text into ChatGPT, and it fixed them. Instead of two hours going from link to link fixing that, it made that type of work much more seamless."
"even Transformers, the library, is not used in production. People use SGLang or vLLM, and it adds another layer of complexity."
"even Transformers, the library, is not used in production. People use SGLang or vLLM, and it adds another layer of complexity."
"And then you start, let's say, with your GPT-2 model and add these things."
"With OLMo 3, the challenge was RoPE for the position embeddings. They had a YaRN extension and there was some custom scaling there, and I couldn't quite match these things."
"They had a YaRN extension and there was some custom scaling there, and I couldn't quite match these things."
"For the character training thing, I think this research is built on fine-tuning about 7 billion parameter models with LoRA, which is essentially only fine-tuning a small subset of the weights of the model."
"And listeners may know diffusion models from image generation, like Stable Diffusion popularized it."
"There was a paper on generating images. Back then, people used GANs, Generative Adversarial Networks."
"It's kind of similar to the BERT models by Google. Like, when you go back to the original transformer, they were the encoder and the decoder."
"But there was an announcement by Google, a site where they said they are launching Gemini Diffusion, and they put it into context of their Gemini Nano 2 model, and they said basically: for the same quality on most benchmarks, we can generate things much faster."
"they put it into context of their Gemini Nano 2 model"
"Like what Apple tried to do with the Apple Foundation models, putting them on the phone, where they learn from experience."
"One thing people still use is LoRA adapters. These are basically, instead of updating the whole weight matrix, there are two smaller weight matrices"
"With Nemotron 3, they found a good ratio of how many attention layers do you need for the global information compared to having these compressed states"
"DeepSeek-V3.2, where they had a sparse attention mechanism where they have essentially a very efficient, small, lightweight indexer"
"There was a paper by Meta, a paper called World Models. So where they basically apply the concept of world models to LLMs again"
"There is a competition called CASP, I think, where they do protein structure prediction"
"AlphaFold, when it came out, it crushed this benchmark"
"There's some work in this area like RTX, I think it was a few years ago, where people are starting to do that"
"I think when I was at Hugging Face, I was trying to get this to happen, but it was too early. It's like these open robotic models on Hugging Face"
"I don't know if you like the originally titled AI27 report. They focus more on code and research taste, so the target there is the superhuman coder"
"I think there are startups—maybe Harmonic is one—where they're going all in on language models plus Lean for math"
"language models plus Lean for math"
"We talked about Memory, which saves across chats. Its first implementation is kind of odd, where it'll mention my dog's name or something in a chat"
"You want to add a new tab in Slack that you want to use, and I think AI will be able to do that pretty well"
"take something like Slack or Microsoft Word. I think if organizations allow it, AI could very easily implement features end-to-end"
"The word 'transformer' could still be known. I would guess that deep learning is definitely still known, but the transformer might be evolved away from in 100 years with AGI researchers everywhere."
"ChatGPT has a memory feature, right? And so you may have a subscription and you use it for personal stuff, but I don't know if you want to use that same thing at work."
"Let's throw in Mistral AI, Gemma..."
"Let's throw in Mistral AI, Gemma..."
"gpt-oss, the open weight model by OpenAI... gpt-oss-120b is actually a very strong model and does some things that other models don't do very well."
"Actually, NVIDIA had a really cool one, Nemotron 3."
"For example, if you read a Substack article, I could maybe ask an LLM to give me opinions on that, but I wouldn't even know what to ask."
"I would love to have tried Bing Sydney. Did that have more voice? Because it would so often go off the rails, which is historically obviously a scary way—like telling a reporter to leave his wife—is a crazy model to potentially put in general adoption."
"There was a lot of backlash last year with GPT-4o getting removed, and I've personally never used the model, but I've talked to people at OpenAI where they get emails from users that might be detecting subtle differences in the deployments in the middle of the night."
"We see this with TikTok. You open it... I don't use TikTok, but supposedly in five minutes the algorithm gets you. It's locked in."
"A lot of researchers at these companies are so well-motivated, and definitely Anthropic and OpenAI culturally want to do good for the world."
"A lot of researchers at these companies are so well-motivated, and definitely Anthropic and OpenAI culturally want to do good for the world."
"my wife the other day—she has a podcast for book discussions, a book club, and she was transferring the show notes from Spotify to YouTube, and then the links somehow broke."
"my wife the other day—she has a podcast for book discussions, a book club, and she was transferring the show notes from Spotify to YouTube, and then the links somehow broke."
"even something simpler like MMLU, which is a multiple-choice benchmark. If you just change the format slightly, like, I don't know, if you use a dot instead of a parenthesis or something like that, the model accuracy will vastly differ."
"When you code these from scratch, you can take an existing model from the Hugging Face Transformers library. The library is great, but if you want to learn about LLMs, it's not the best place to start because the code is so complex to fit so many use cases."
"But I do still think that eventually, something like ChatGPT would have happened and a build-out like this would have happened, but it probably would not have been as fast."
"I think it only happened because you could purchase those GPUs."
"Amazon is making Trainium."
"We hear about Reflection AI, where they say their two billion dollar fundraise is dedicated to building US open models"
"They're signing licensing deals with Black Forest Labs, which is an image generation company"
"signing licensing deals with Black Forest Labs, which is an image generation company, or Midjourney"
"We are starting to see some types of consolidation with Groq for $20 billion"
"Scale AI for almost $30 billion and countless other deals like this"
"I think there will be some other multi-billion dollar acquisitions, like Perplexity"
"That's why part of what Vera Rubin is- where they have a new chip with no high-bandwidth memory, which is one of the most expensive pieces"
"Like, Google obviously can make TPUs"
"Amazon is making Trainium"
"The moat of NVIDIA is probably not just the GPU. It's more like the CUDA ecosystem, and that has evolved over two decades"
"We should say that the AI27 report kinda predicts one of the things it does from a narrative perspective is that there will be a lot of centralization."
"That's supposed to be the point of the Groq acquisition."
"Like, Google obviously can make TPUs."
"The ADAM Project started as me calling it the American DeepSeek Project"
"The goal with the book is basically to understand how the LLM works. It's not going to be a production-level LLM, but once you have that, you can understand the production-level LLM."
"I have a repository called MLxtend that I developed as a student around 10 years ago, and it is a reasonably popular library still for certain algorithms, I think especially like frequent data mining stuff."