What I've Been Reading (August 2025)
Here are my thoughts on a few things I read lately.
Books
- Algospeak by Adam Aleksic
In an attempt to try to be a little bit more offline, I picked up a physical copy of Adam Aleksic’s Algospeak from a local bookstore. Aleksic makes some fairly astute observations, such as how creators “cross-post… and as such try to work within the most severe algorithmic constraints” and how middle-schoolers “aren’t saying ‘unalive’ to avoid algorithmic censorship, they do so because they see other people using it.”
Aleksic uses etymology to really dissect how slang and new words move from memes and trends into the offline world. He says they need to start in a community (where they spread through memes), be unobtrusive (not too tied to a meme or too random), and fill a “semantic gap” (that is, express an idea that doesn’t have an existing word, so that its use is sticky).
He also discusses techniques that video creators use to hold your attention – like how videos often start instantly with someone speaking (unless they have a “millennial pause”) and the use of new phrases like “no because”.
Aleksic then shows how these two ideas, combined with the growing proliferation of communities, creates a cycle that causes English to evolve faster and faster.
Finally, Aleksic closes by showing how corporations use microlabels and hyper-compartmentalization of identity to commodify them and sell you more stuff (clothes, music, and so forth).
After reading the book, I was surprised to find that a dozen or so friends also follow Aleksic on Instagram and a few even had it on their Goodreads To-Read lists. IMO, Algospeak is a good and quick read for anyone who works as a developer on a social network or ecommerce.
Blogs and other internet stuff
I’ve heard a few anecdotal reports from dev friends that AI doesn’t make them 2x more productive, let alone 10x more productive. Vibe Code is Legacy Code offers a simple, compelling argument why this could be the case.
fast reiterates that if your software feels instant, that it changes how users use it.
There are a lot of fun new performance problems to solve with AI / LLMs. The 0.1 / 1.0 / 10 second guidelines that Catherine alludes to date back to 1968 (Response Times: The 3 Important Limits) and they still apply today.
Last year we saw techniques such as speculative decoding (Lex Fridman podcast #447) and streaming responses.
Key word being open (well, weights-available).
I’d probably still default to trying a closed model first. Of course you need a strategy for dealing with HTTP 429 Too Many Requests
if you happen to get a sudden surge of traffic, a way to track if quality (“evals”) regress, and to periodically re-evaluate with new models and providers.
It has been cool seeing some of the throughput numbers from Cerebras and Groq though, if you can find a subproblem in the grander architecture where it makes sense to slot them in.
I first came across the Bitter Lesson from a former colleague. If it does hold, then in the limit you get results like “we don’t even need OSes any more; just emulate the OS with a generative model”. Some folks at the University of Waterloo tried this (NeuralOS, (HN)) and the results were simultaneously shocking (“wow I can actually click an icon and open a window”) and underwhelming (“we’re using the full power of an H100 and this is in many ways a worse ‘computer’ than I had in 1995”).
Power, compute, and financial constraints (that is, “the laws of physics” and “budgets”) suggest to me that there’s still plenty of economic value to be delivered by “building knowledge in” in the short term with classical methods.
I think my favorite version of this was Timefold’s How I built an AI company to save my open source project. Timefold is “just” a rebranded constraint solver and has nothing to do with LLMs or generative AI, but brands itself as AI because that’s what VCs, boards, and investors demand of businesses.