@JohnThilen @dwarkesh_sp @ilyasut In addition, and that’s the important point, I think GPT-5 is smaller than GPT-4.5.
@JohnThilen @dwarkesh_sp @ilyasut I am speculating that all GPT-5.1 models (instant, thinking, Pro) are the same model but with different inference scaling budgets. Same for GPT-5 Codex. And Gemini 3 Pro and Gemini 3 Deep Think are probably also the same...
How do devs feel about this job change? Example below in Codex. Pros: you can kick off tasks from anywhere (just talked to one dev who started multiple codex tasks while getting into a cab), you get multiple versions to pull...
Any talented Product Managers with experience in MarTech/SalesTech want to work together? I'm building a new product (currently in stealth mode) and looking to make our first hire on the product side after the holidays. This person: • Loves the...
RT @VisionEdgeMktg 🤝 Trust is a strategic asset. It fuels customer loyalty, advocacy, and long-term value creation—and it should be embedded in your growth strategy, not left to chance. https://t.co/WCiTg9B4Va #Trust #CustomerValue #GrowthStrategy
@GiorgioMantova @dwarkesh_sp @ilyasut I’d say this is the jump from last gen to current gen, but I think the argument is that further improvements will fizzle out in the next gen if we keep scaling pre-training. Ie it won’t give...
Reachy mini is my new podcast assistant! Coming soon with @ti_morse... https://t.co/VfUQn1Cgz6
@_The_Prophet__ TPUs had low availability for ages and also low memory relatively on the v6e especially versus the hoppers working pretty much out of the box similar to a100s Grace Blackwell is the next thing that needs reworking so there is...
I was spending 30+ hours a week on stuff that has zero impact on growth. Scheduling meetings. Updating spreadsheets. Missing my kids' bedtime because I was answering emails until midnight... That’s when it hit me: I didn’t need more hustle....
@_The_Prophet__ TPUs have been more stable for training than CUDA equivalents for a couple of years now, especially on large batch sizes XLA is pretty good now! For inference it makes even less of a difference (We previously trained sota models on thousands...
I agree. Incredible interview by @dwarkesh_sp of @ilyasut. I could listen to both for months and not get bored. It's like being at a great university and hearing the best professor. I love X. This just LIT UP the AI community....
One AI use case that is only getting more popular: LLM as a judge. Everyone still talks about AI generating more content, but not enough people are talking about: 1) the horrific deluge of noise we're going to have to deal...
So thankful to Ann Handley and the entire MarketingProfs team for allowing me main stage time for a very important concept I've been sharing. The TLDR: Someday soon, our glasses, cars, phones - everything - will create synthetic content on...
B2B marketers are prioritizing #AI tools for 2026 https://t.co/txvQbppcAJ
it is wild how far we’ve come since the hot image gen summer of 2022 image cred: @bfl_ml https://t.co/i5KbszlFDM
Just spoke to a friend who founded an early-stage AI startup (yes, this was an actual conversation - not one of those LinkedIn “I was just chatting with…” cold opens 😭). We were talking about a real dilemma: enterprise buyers...
the community’s favorite image creation and editing model just got better: welcome, FLUX.2 by @bfl_ml 🤩 https://t.co/iLrbYYK4bd
I think it is somewhat true though that scaling helps with benchmark performance but not necessarily with with new model capabilities. Like the example he mentioned > U: "Please code xyz." > M: "Ok here is xyz." > U: "You have a bug." >...

So you are a CV engineer, what do you know about Computer Vision? I have used YOLO for... https://t.co/lX4OrFFYqE
PING! You ship that super interesting, well-written blog post. People love it. Traffic spikes, comments roll in… then things start to settle down. PING! You fire off a hot take on LinkedIn. Debate flares up in the comments, your audience...
Here's our paper: https://t.co/RmNft3zU5Z
Excited to present our new AI paper as a @NeurIPSConf spotlight next week: we find that the problem of controlling artificial superintelligence remains unsolved. With simulations and scaling laws, we find that an implementation of the least unpromising...
OpenAI is very deliberate about how they talk about Codex. It's not positioned as an operating system. It's heavily positioned as a teammate. Their site says: "Your new coding partner", "accelerates your team" Their job postings say: "we're building an AI software...
@dwarkesh_sp @ilyasut “The Age of Scaling is over.” I agree with that. Basically, since GPT 4.5 a lot of the perceived real-world progress was driven by clever engineering wrappers (context filtering, inference scaling, multi-turn tricks, retrieval, tool use, etc).
Excited about the Genesis mission - congrats to @POTUS @SecretaryWright @ScienceUnderSec @mkratsios47 @sriramk! We've experienced first-hand how more openness and collaboration in the US can massively accelerate progress. In my opinion, that's what led to the current AI boom and US...
Just shared this brilliant mind map on the 15 key architectural characteristics of AI agents — absolutely packed with insights! Modularity, evolvability, context awareness, security compliance… everything you need to design robust agents. Huge thanks to @Python_Dv for creating this gem
@jeffkmains @Target @AskTarget Explainer videos right around 90 seconds tend to do very well.
@ronzilla @_lhermann Thanks! Also did a talk at MicroConf that covers all of these questions: https://t.co/nWzaXhZLyI The SaaS Playbook section on pricing is an adaptation of this talk, with a bit more info included.
What key things should founders look for when hiring an investment bank to help sell their business or raise capital? Most software founders don't know exactly who needs to be involved in order to pull off a successful sale or...
Ok, so what Ilya saw was extreme benchmaxxing, which in turn prompted him to create his own company to do LLM development the proper way?! Makes sense, I sympathize with that.
@jasonlk Taste is hard (hello feature bloat) Maintaining is hard Legit integrations are hard
@giffmana @dileeplearning the "correct-unintended" rules were just that -- correct on the demonstrations but using "shortcuts" (e.g., the numerical value of a color). We also saw a small percentage of "correct-unintended" rules that humans generated, but much less...
📢 Image-GS: Content-Adaptive Image Reconstruction using 2D Gaussians In this week’s deep dive, we explore Image-GS, a groundbreaking framework that reimagines how images can be represented, compressed, restored, and upsampled using adaptive 2D Gaussian splats. Unlike traditional codecs or neural...
@giffmana @dileeplearning There was a big difference between "not classified" rules generated by humans and "correct-unintended" rules generated by machines. For humans, the "not classified" rules were generally humans writing nonsensical things like ⬇️
My top takeaways from executive coach Rachel Lockett : 1. The biggest skill gap in new leaders is knowing when to coach vs. when to tell people what to do. When you constantly provide answers, you train your team to...
Have you ever wondered why so many companies struggle to align their sales and marketing teams? The misalignment often stems from the different mindsets that each team brings to the table. Marketing typically zeroes in on the top of the...
#SaaS Video #Marketing Trends: What Actually Works in 2025 - B2B Marketing Blog | Webbiquity - https://t.co/JXmCYFiQ67 Also, I think I'm just going to start adding #TargetSucks to every tweet. Obviously, @Target and @AskTarget don't care how badly they abuse their...
One more comment is that giving this image to an AI and asking about it is not sufficient to show the diff because it's all over the training data by now. You'd have to use a new, very recent image,...
@matejhladky_dev AI has crushed it since this post way beyond expectation. I made the same category of mistake all of AI was making, of thinking we have to discover and write the algorithm. You don't. You pretrain and then finetune...
I've had medium success asking LLMs if a thing exists, it works out of the box for some of the more well-known things (e.g. both GPT 5.1 and Gemini 3 know about this function if you describe the tensor transformation...
The CMO-CFO relationship: Here’s what the numbers say https://t.co/R5srOt2Ewj
@UmmayHabiba0 @SchneiderNA Certainly. We’re witnessing a major shift in real time. AI and energy tech are finally converging in ways that will reshape how industries operate and how infrastructure is built. Here’s the video if you’d like to take a look: 📺...
@jasonlk Agreed. Do you think tho that the threat of vibe coding plus ability for more competitors to emerge will create more pricing and gtm pressure?
@the_AI_girl @SchneiderNA Absolutely, the momentum building across AI, energy, and infrastructure is setting the stage for a major transformation in the U.S. economy. I just shared more of my insights here on @LinkedIn : https://t.co/WwaOkGdcNm Big shifts ahead.
Always a slightly mixed feeling to write pretty good first-principles code to do some tensor rearrangement, only to find that PyTorch has a built in function that does it faster. I had made a point of at least skimming the docs...
Check out our new B2B Thought Leadership research that uncovers the 'Answer Engine': How top performers achieve nearly 4x higher ROI https://t.co/98EPovULaW via @toprank & @Ascend2research https://t.co/5a1Xt1LTtZ
85% of organizations believe responsible AI is a top management issue. Yet only 25% have governance mechanisms in place to address it. This trust gap is costing companies dearly. In Europe alone, 68% of companies don't understand their EU AI...
1.2 million samples. BM25, Embeddings and Hybrid search. Tutorial and code comes tomorrow! Stay tuned! https://t.co/FlmaDlpASR