Bits and Bobs 7/7/25
1LLMs are a mirror.
2ChatGPT's answers feel to me like being served up a personalized Axios article.
- ChatGPT's answers feel to me like being served up a personalized Axios article.
- More formatting than substance.
- Punchy and "simple" while obscuring nuance.
- Presents the answer as an objective, simple truth, as opposed to a nuanced observation as a point of departure and follow up.[fp]
3Your AI assistant isn't your friend.
- Your AI assistant isn't your friend. It's a double agent.
- This comes from Bruce Schneier's excellent AI and Trust.
4Chat is a forgiving interface for quality.
- Chat is a forgiving interface for quality.
- If it gets the answer wrong or doesn't do what you wanted, you can easily ask follow up questions.
- One-shot UIs that have to give a perfect answer every time with no second tries are very hard to get to the bar of viability and then improve.
- Compare Google Search and (spoken) Google Assistant interactions.
- With Google Search, as long as the answer is in the top 10, it's fine.
- Formulating a query is fast, and text results are a broad channel, easy to skim through quickly.
- Then, if users are consistently clicking on the third result, the system can twiddle its ranking to be to the top.
- This gives a clear gradient to hill climb.
- With Google Assistant, formulating the query and getting the response over audio are a narrow channel.
- It's impossible to speed up the readout, so you get one answer, not several.
- If the answer isn't right, users give up.
5I loved this socratic dialogue from Geoffrey Litt about why chat is not the final UI for LLMs.
- I loved this socratic dialogue from Geoffrey Litt[fq] about why chat is not the final UI for LLMs.
6Every form of UX around chatbots today has tried to grapple with the model's limitations.
- Every form of UX around chatbots today has tried to grapple with the model's limitations.
- Then the next model gets better and obviates the UX.
- What are the downsides fundamental to chat itself?
- Those will help discover the ideal UX for LLMs.
7Chat is ephemeral and squishy.
- Chat is ephemeral and squishy.
- That's what makes it good for starting open-ended tasks without it feeling like a burden.[fr]
- That's also what makes tasks as they continue feel like wading through molasses.
8If chat isn't a good UI for LLMs, why is it winning?
- If chat isn't a good UI for LLMs, why is it winning?
9Chat is the poor man's open-ended UI.
10With LLM UX it feels like we have all the ingredients but we don't have the cake yet.
- With LLM UX it feels like we have all the ingredients but we don't have the cake yet.
- Chatbots seem like a local maxima to me.
- Chatbots are not the ideal UX.
- I think the ideal UX will include the ability to chat, but that will not be the primary interaction that everything orbits around.[fw]
11A coactive relationship is one that empowers both participants.
- A coactive relationship is one that empowers both participants.
12"Apps that adapt to you" could be powerful.
- "Apps that adapt to you" could be powerful.
- Today apps are static.
- Over time you'll not even reach for apps at all because everything you do will be adapted so much to your situation that the notion of an app fades away.
13I want Living Apps.
- I want Living Apps[fz].
- Apps that are alive, that adapt themselves to me, pulling in context or code to make themselves do what I want to do.
- I could tweak them to my heart's desire.
- Apps for living.
- They could also be called "coactive apps".
- Coapps are apps that adapt themselves to my needs.
- Alternatively, call them tools. Living Tools, or Coactive Tools.
- 'Tools' doesn't have the expectation of "these are just like apps of today."[ga]
14The ideal software in the era of infinite software is pre-assembled lego sets.
- The ideal software in the era of infinite software is pre-assembled lego sets.
- You get a full, useful thing out of the box, that an expert designed to be holistically useful.
- But it's made of legos, so you can replace any block… or add whole new things to it.
- So then creators can create both new lego blocks... but also new pre-assembled lego sets that all fit together nicely and coherently and are useful right out of the box.[gb]
15How much you trust a suggestion is partially due to what context the system is drawing on.
- How much you trust a suggestion is partially due to what context the system is drawing on.
- Is it drawing on relevant facts about you?
- Is it missing important ways that you differ from the general population?
- Is it including irrelevant things that will distract it?
- The quality of the context is a big determinant in how good the results are.
- This is one of the reasons the ChatGPT dossier memory feels off to me.
- You can't inspect the context to say, "include this, not that".
- The chat is append only, not coactive.
- You can imagine a UX where there's a coactive context drawer at the top of the interaction.
- Things in the interaction and that context drawer are all that is given to the LLM, nothing else.
- When the drawer is collapsed it shows a short summary.
- When you expand it you see trees of context that you've pulled in explicitly.
- You can add in trees of context on demand easily.
- E.g. "Include information about my nuclear family."
- The tree pulls in all of the sub-items that hang off of it.
- You can choose to pull in something high up in the tree or low down.
- You can delete any tree of context that's not relevant.
- There's also a list of auto-included context that the system guesses is useful.
- Those are included by default, but can be explicitly added to the included items, just like if you'd added it yourself, or deleted.
- The ranking function is: how well does the system predict which trees of context to include, (predicting whether the user would accept or delete a suggestion?)
- The user choosing to include or excluding bits from the suggested context is an extremely powerful ranking function.
- The user wouldn't even realize that they're training the system for themselves and others by gardening their context.
- Only a small number of users would need to do it to help tune it for a whole population.
16Formulating context as a memory makes it sound like it's for the LLM.
- Formulating context as a memory makes it sound like it's for the LLM.
- Ideally you'd organize it for yourself, which would be useful on its own.
- As a bonus it makes the LLM significantly better at doing things for you.
17A good executive business partner doesn't "remember" you, they know you.
- A good executive business partner doesn't "remember" you, they know you.
- There's distillation.
- How can you get a good enough model of yourself for the LLM?
18Imagine a productivity system that does all the grunt work for you.
- Imagine a productivity system that does all the grunt work for you.
- That gets better as you use it more, not because some PM added a feature or some fully automatic LLM-based insights.
- But because it draws on the collective wisdom of everyone using it.
19Dan Petrolito: I built something that changed my friend group's social fabric.
- Dan Petrolito: I built something that changed my friend group's social fabric.
- An extremely trivial script at the right moment can have a massive impact.
- Situated software catalyzes cozy potential.
20LLMs have lots of last-mile problems.
- LLMs have lots of last-mile problems.
- Largely due to gardening the right context and the right coactive UI.
21Claude Code feels like a ride-on mower.
- Claude Code feels like a ride-on mower.
- At first you go "Whoa, this is so easy to use, I can do 10x more than I could before."
- But as you use it for more things you realize it's too coarse a tool to do detail work.
- An extremely useful tool, but doesn't replace all of your tasks in the garden.
22Claude Code feels like a "choose your own adventure" style of developing software.
- Claude Code feels like a "choose your own adventure" style of developing software.
23I liked Grant Slatton's summary of techniques for LLM memory.
- I liked Grant Slatton's summary of techniques for LLM memory.
24Chris Joel pointed out to me that DNS origins are like Neal Stephenson's burbclaves.
- Chris Joel pointed out to me that DNS origins are like Neal Stephenson's burbclaves.
- "Now a Burbclave, that's the place to live. A city-state with its own constitution, a border, laws, cops, everything."
25If a task was 1000x too hard and is now 10x easier, it's still 100x too hard!
- If a task was 1000x too hard and is now 10x easier, it's still 100x too hard!
26I think people want a common-sense vision for optimistic human centered computing in the age of AI that is not:
- I think people want a common-sense vision for optimistic human centered computing in the age of AI that is not:
- 1) Cynical engagement-maxing tech products of today.
- 2) Crypto.
- 3) E/Acc / Successionism.
- People presume that if you're optimistic about tech and are in the industry you just want centralization.
- Or that if you're optimistic about tech and don't want centralization then you must like crypto.
- But it's possible to be optimistic about tech and push for neither centralization nor crypto.
- Such a third way is more important than ever before in the era of AI.
27A HackerNews comment that stuck with me: "From the very beginning Facebook has been an AI wearing your friends as a skinsuit."
- A HackerNews comment that stuck with me: "From the very beginning Facebook has been an AI wearing your friends as a skinsuit."
28When naming something novel, the slingshot maneuver can be helpful.
- When naming something novel, the slingshot maneuver can be helpful.
- Name it based on what people know they want.
- Then slingshot them to the thing they didn't know they needed.
- That latter part only becomes clear once they've used it.
- Meet them where they are to take them to where they should go.
29There's a GitHub project with simple little LLM based "gremllms".
- There's a GitHub project with simple little LLM based "gremllms".
- When you access a method, the LLM generates code JIT.
- I think the mental model of gremlins fits well: small, not too powerful, but mischievous and a bunch of them together can make an impact.
30The word "context" is a good one for "relevant data to give as background to the LLM."
- The word "context" is a good one for "relevant data to give as background to the LLM."
- But in the deep philosophical sense, your real context is outside your control.
- Your context is a gravity field.
- The water you swim in.
31If you've gone through the effort of having high quality programmatic thinking LLMs can write infinite Op Eds for you, on demand.
- If you've gone through the effort of having high quality programmatic thinking LLMs can write infinite Op Eds for you, on demand.
- The backlog of Bits and Bobs feels exceptionally valuable to me in the age of AI.
32Teaching forces you to abduct your intuition.
- Teaching forces you to abduct your intuition.
- That's why one of the best ways to understand something is to teach it.
33A slippery slope is an example of an emergent phenomena of noisy signal with consistent bias.
- A slippery slope is an example of an emergent phenomena of noisy signal with consistent bias.
- No individual step is that bad, obscured by noise.
- The bias is in one direction: the gravity of incentives.
- So the emergent global effect is clear and powerful.
34Bruce Schneier points out that LLMs will bring mass spying.
- Bruce Schneier points out that LLMs will bring mass spying.
- Before, we had mass surveillance, but a human sifting through the collected data happened rarely.
- That limited the oversight to things that could be done at quantitative scale, or the most egregious tails of behavior.
- A panopticon kind of game theoretic dynamic.
- But LLMs give qualitative insights at quantitative scale.[ge]
- Society now has the technology to get qualitative insights at scale from that surveillance.
- What could possibly go wrong??
35LLMs are infinitely patient so good enough ACLs aren't good enough any more.
- LLMs are infinitely patient so good enough ACLs aren't good enough any more.
- Before your data was protected a bit by security through obscurity.
- But LLMs are infinitely patient to sift through data that was accidentally left open.
- Another outgrowth of the "qualitative insights at quantitative scale".
36Some examples of sycosocial relationships with LLMs:
- Some examples of sycosocial relationships with LLMs:
- Examples in this Ezra Klein interview.
37Garry Tan on Twitter:
- "New social networks are going to appear that will be LLMs creating a cozy web customized for us and our real friends, and their friends and so on
- There will be a new social network built on mutual trust, all curated by machines of loving grace
- Personal Cozyweb is inevitable"
- I originally interpreted this tweet as "Use LLMs to make a psychosocial bubble of fake friends" which I think would be terrible for society.
- But I think he meant it more as "make a garden of possibility for you and you friends," which could be good in some circumstances.
38Doug Shapiro: Trust is the new oil.
- Doug Shapiro: Trust is the new oil.
- Trust will have to be rooted in in-person interactions that are known to be authentic.
- Bruce Schneier's focus on integrity is important here.
39Culture emerges.
- Culture emerges.
- Someone tries something that other people find viable and then the others reshare it, build on it, and remix it.
- The things people like are what get built on, emergently.
- For things like architectural styles there's some geographic pockets which helped give distinctive styles in different parts of the country.
- As everything gets more fluid and lower friction, we'll likely see less and less distinctive architectural styles.
40Building for scale and building for viability are different.
- Building for scale and building for viability are different.
- Imagine working at a company where on day 1 of a product you can expect 50M users.
- You have to think about every little detail before building it.
- In that environment product discovery is about talking and planning.
- Writing down plans so they can be critiqued and collaborated on is a critical step in the process.
- But in most contexts, the 0-1 phase has very little usage.
- Product discovery there is about experimenting and surfing through the problem domain fluidly.
- It's all about getting something concrete into contact with real people as quickly as possible to iterate.
- In that environment, writing down things slows down the discovery process significantly.
41The demoware mindset is "does it work?"
- The demoware mindset is "does it work?"
- The product mindset is "do I want to use it?"
- Very different bars!
42People who have a high need for novelty won't focus on polish.
- People who have a high need for novelty won't focus on polish.
- Polish isn't novel.
43Even if you build the right surfboard, will you catch the wave?
- Even if you build the right surfboard, will you catch the wave?
44Hank Green on the Cacophonous Age: "You're not addicted to content, you're starving for Information."
- Hank Green on the Cacophonous Age: "You're not addicted to content, you're starving for Information."
45In the Cacophonous Age, the new privilege is patience.
- In the Cacophonous Age, the new privilege is patience.
- What kind of thinking are you not outsourcing?
46How can you have a secure attachment to reality instead of trying to leave it?
- How can you have a secure attachment to reality instead of trying to leave it?
- To get it you need to sit with the discomfort of uncertainty.
- The discomfort is the point.
- Reframe the discomfort as excitement!
47When you have your own authentic clarity, you stick with your intuition when everybody else would have given up.
- When you have your own authentic clarity, you stick with your intuition when everybody else would have given up.
48Terms have an "inferred definition."
- Terms have an "inferred definition."
- That is, what a term means is what the majority thinks it means after the first time they've heard it.
- People will bring their own preexisting priors to any given term, and that bias will lead to what the term means, especially if it's a consistent bias many first-time hearers will share.
- Terms like "context engineering" are useful because they mean the thing that most experts hearing it for the first time would think it means.
- "Inferred definition" is itself a term coined by Simon Willison.
49Your taste creates an architecture for your thoughts.
- Your taste creates an architecture for your thoughts.
50If a CEO could direct a swarm of LLM clones of themselves, we'll expect more volatility in company performance.
- If a CEO could direct a swarm of LLM clones of themselves, we'll expect more volatility in company performance.
- Founder led companies are more volatile.
- If the founder says to go in a given direction, even if it's to avoid an obstacle the employees can't see, they go along with it.
- But if the founder missteers, there's no one and no thing to countersteer.
- Founder-led companies have greater returns than average, but also higher likelihood of death.
- Non-founder led companies are harder for the CEO to steer.
- But even founder-led companies are hard to steer at scale.
- Before, the swarm of employees trying to implement the CEO's vision imperfectly gave some insulation.
- For good, when the CEO's idea was disastrous.
- For bad, when the company had inertia that counteracted a good idea.
- But if every employee is just a clone with minimal principal agent problem, it's like the Wreck it Ralph 2 swarm of poorly rendered copies creating an unruly emergent leviathan.
51Which does the organization prioritize: loyalty or competence?
- Which does the organization prioritize: loyalty or competence?
- The former is a zombie organization.
- Alignment at all costs, even if it kills the host.
- No adaptive capacity.
- Top-down alignment comes at the cost of local adaptivity and the possibility of emergence.
52Whiteboard scribblings after a meeting are often completely meaningless to anyone else.
- Whiteboard scribblings after a meeting are often completely meaningless to anyone else.
- But they're extremely meaningful to the people who were there.
- Having the experience of that meeting in that space gives you the key to understand it.
53Novelty is risk.
- Novelty is risk.
- It's noise.
- Some novelty will turn out to be innovation.
- Most will simply not work.
- Invest your novelty budget on the things that are differentiators, but nowhere else.
54The optimal strategy for a swarm and for an individual are different.
- The optimal strategy for a swarm and for an individual are different.
- Multi-headed search vs single-headed search.
- If you can only have a single search head on a problem domain, you want it to be the best it can be.
- K-selected.
- If you can have a multi-headed search on a problem domain, you want as many heads as you can get to flood-fill the problem.
- R-selected.
- The swarm needs breadth, the individual needs depth.[gf]
56Some things are convex and some are concave.
- Some things are convex and some are concave.
- Some tend towards a center point.
- Auto-stabilizing.
- Concave.
- Some tend away from a center point.
- Auto-destabilizing.
- Convex.
- Two systems that look superficially the same but are convex vs concave have infinitely different outcomes.
- What determines if a system is convex or concave?
- If it leads to convergent outcomes or tears itself apart via entropy and diffusion?
- I think it's whether the locally good behaviors lead to emergently good outcomes at the macro level.
57Some things, the more useful they are over time, get cleaner.
- Some things, the more useful they are over time, get cleaner.
- In some cases the more useful they are, the more fractally complex they get.
- Convex vs concave.
58Useful things tend to snowball.
- Useful things tend to snowball.
- In proportion to:
- 1) the amount of people who find it useful.
- 2) how useful they find it compared to available alternatives.
- As it gets bigger it gets more useful to more people because more people are exposed to it.
59Emergent systems need to have all logic decided at the local level, but have global-level outcomes emerge.
- Emergent systems need to have all logic decided at the local level, but have global-level outcomes emerge.
- In emergent systems, all decisions are local but they have emergent global consequences.
- You can't get a bird's eye view to coordinate, which is necessary for top-down convergence in a large system.
- Many problems can't be framed this way, with a successful macro level outcome without a birds eye view, but some subset can.
- These are typically grown, budded off of other systems that are working.
- But a true bird's eye perspective is an impossibility anyway.
- The farther you get away from the details, the more fuzzy they become.
- The constraint of "local information only" feels overly restrictive, but it's close to the real constraint anyway.
60Feedback is generated when you take an action and the world reacts.
- Feedback is generated when you take an action and the world reacts.[gg]
- You can't simulate what the world will do without taking the action.
- The world is a multi-headed environment of execution.
- Your head is a single-headed environment.
61Debugging tools only give their highest leverage if they're at the proper layer of abstraction.
- Debugging tools only give their highest leverage if they're at the proper layer of abstraction.
- An inode-level debugging tool for a userland storage system is not helpful.
- Good debugging tools give good leverage to features / bugs at that layer.
62One reason Rust's Borrow Checker is palatable is Rust's amazing error messages.
- One reason Rust's Borrow Checker is palatable is Rust's amazing error messages.
63"What's your dirty little secret?"
- "What's your dirty little secret?"
- Related to the carrot problem.
- Every team or product has one.
64Data used to be sent on physical media, and moving the physical media had significant friction.
- Data used to be sent on physical media, and moving the physical media had significant friction.
- You'd print words in a book and then ship the book.
- When data moved to the plane of entirely bits it got orders of magnitude faster.
- We divorced information from atoms.
- We saw much more aggregation much faster.
- The same dynamic as before, just playing out orders of magnitude faster.
65An important tactic: lashing yourself to the mast.
- An important tactic: lashing yourself to the mast.
- To make it so you can't do the things you fear you'll want to do.
- Past you can constrain future you, to make sure your reptile mind doesn't override your aspirational mind.
- A game theory solution that's similar to playing chicken with yourself.
66Sometimes what you think will be a shortcut is actually a detour.
- Sometimes what you think will be a shortcut is actually a detour.