Bits and Bobs 6/2/25
1Another day, another prompt injection vulnerability.
- Another day, another prompt injection vulnerability.[kn]
- "BEWARE: Claude 4 + GitHub MCP will leak your private GitHub repositories, no questions asked.
- We discovered a new attack on agents using GitHub's official MCP server, which can be exploited by attackers to access your private repositories."
2I think that building MCP into Windows could go down in history as a colossally reckless idea.
- I think that building MCP into Windows could go down in history as a colossally reckless idea.
3People will route around things they don't trust.
- People will route around things they don't trust.
- …maybe not only people!
4A lot of people have told me that "things like code injection don't happen anymore."
- A lot of people have told me that "things like code injection don't happen anymore."
- That's why prompt injection won't be a big deal, they assure me.
- The reason code injection attacks don't happen nowadays is not that the threat went away.
- It's that the mechanistic defenses against it got strong enough to make it not worthwhile.
- The lack of code injection attacks in the wild is a testament to the strength and maturity of our operating systems, not to a lack of demand for attacks.
- Prompt injection cannot be solved by mechanistic approaches like vanilla code injection can.
- Also remember, the distribution of threats is not fixed; it coevolves with the opportunity.
- The weaker the system, or the more usage, the more monetary sense the threat makes.
- Don't confuse the lack of prompt injection attacks with a lack of demand.
- It's simply a matter of lack of widespread adoption of tools like MCP today.
5Someone told me they used MCP in production but insisted they did it safely.
- Someone told me they used MCP in production but insisted they did it safely.
- I asked them what integrations they used.
- They said they had a Jira integration and one for their company's financial data.
- I asked them if they could generate markdown reports with images.
- They said they could--that's how they visualized the financial data.
- I asked them if they had a feedback form on their site.
- They said they did.
- I asked them what happens when a user files feedback.
- They told me it files a Jira ticket.
6The villain in the original Tron is called Master Control Program.
- The villain in the original Tron is called Master Control Program.
- MCP.[kq]
7A prompt injection stored in your context is a persistent prompt injection.
- A prompt injection stored in your context is a persistent prompt injection.
- Prompt injection attacks that can embed themselves in your personal stored context might never be found.
- Echoes of the classic Reflections on Trusting Trust.
8Software created with Loveable is often insecure.
- It makes sense that this would be the case–securing things in software is hard!
- Vibe-coded software from amateurs doesn't have a wide audience.
- I was trying out an example vibe-coded app that was a dream journal.
- You could log in with your Google account and mark dreams you stored as public.
- I went to put in my dream and thought… wait, why would I do that?
- Whoever the anonymous person who created it was, they could have accidentally coded it so that dreams were public by default.
- Or they could send my dreams, with my email address, to marketers.
- I decided not to use it.
- Vibe coding tools are great for PM-types prototyping.
- Or for people building an app for themselves or for their friends.[kr]
9Making software for yourself vs for someone else is quite different.
- Making software for yourself vs for someone else is quite different.
- The quality bar is way lower for yourself vs software you tell someone else they should use.
- You're way more forgiving of software you made yourself.
- You also don't have to fear intentional security holes in software you built yourself.
10In a world of infinite software, you won't necessarily make your app for others, you'll make it for yourself and maybe some friends.
- In a world of infinite software, you won't necessarily make your app for others, you'll make it for yourself and maybe some friends.
- Maybe no one else will trust it or be willing to put up with its lack of quality.
- Vibe coding platforms that hope to make their margin on hosting might turn out to not be viable if this effect is strong.
- If it's only used by a handful of people, the hosting costs will be minor so even a large margin on a small base won't be significant.
11LLMs don't do novelty themselves.
- LLMs don't do novelty themselves.
- But they can give novel answers to novel questions.
- You need to bring the entropy to the LLM.
- If you think LLMs give boring answers, maybe you're asking boring questions![ks]
12Rick Rubin describes vibe coding as the "punk rock" of coding.
- Rick Rubin describes vibe coding as the "punk rock" of coding.
- Seems right to me.
13An aesthetic that could be interesting: cozypunk.
- An aesthetic that could be interesting: cozypunk.
- What would the homes look like inside of solarpunk scenes?
- Warm, human-centered, optimistic.
14Someone described local first software as "edgy."
- Someone described local first software as "edgy."
- Both technically focused on "the edge" instead of central networks.
- But also an act of minor protest against an overly sames-y monoculture of software that is on the server that users can't change.
15A new word I heard this week: "slopdev."
- A new word I heard this week: "slopdev."
- Conjures up a vision of unmotivated shoveling of slop code to make CRUD-y software you don't care about.
- Vibe coding for yourself can be soul-affirming.
- Slopdev for a job is soul-eroding.
16The Wall Street Journal taught my dad about vibecoding.
- The Wall Street Journal taught my dad about vibecoding.
- That's how you know it's become mainstream.
17Overheard this week: vibe coding is like having a swarm of toddlers try to build and maintain your car.
- Overheard this week: vibe coding is like having a swarm of toddlers try to build and maintain your car.
18LLMs do the easy parts of programming.
19AI works better with well written code.
- AI works better with well written code.
- But AI struggles to produce well-written code.
- The more you use AI to create your codebase, the harder time that humans–and LLMs–have with understanding (let alone modifying) it.
- Vibecoding has logarithmic value for exponential cost as the codebase gets bigger.
20A fast pace layer can be sublimated into a lower pace layer once the best practices are conclusively discovered.
- A fast pace layer can be sublimated into a lower pace layer once the best practices are conclusively discovered.
- At that point, no one bothers trying much different at that layer because the best practice is obviously better than whatever they'd build.
- It's hard to build a faster pace layer on top of a pace layer that's still swirling and unsettled.
- That's not a good foundation.
21LLMs are the ultimate amplification algorithm.
- LLMs are the ultimate amplification algorithm.
- They lock in whatever things were dominant when they were trained.
- A gravity well pulling any line of discourse back to it.
22It's conceivable that automatically-generated RLAIF React code in Claude's training pipeline now outnumbers human code.
- It's conceivable that automatically-generated RLAIF React code in Claude's training pipeline now outnumbers human code.
- This will lead to client side code best practices being automatically frozen around 2023.
- The LLMs will have significant momentum towards the best practices of that timeframe, and will get increasingly hard to steer away from them.
- If a change isn't that much better, why fight the LLM?
- Just leave it how it was.
- Like it or not, we'll be stuck in 2023 era front end best practices forever now.
23Maybe we'll see a new explosion of innovation at a higher pace layer because of the great LLM freeze.
- Maybe we'll see a new explosion of innovation at a higher pace layer because of the great LLM freeze.
- The "Javascript Industrial Complex" has led to an extraordinary amount of "innovation" in the client layers for the past decade or so.
- The amount of churn and change is notable, and exhausting.
- There has been a ton of true innovation, but also a lot of just churn.
- But LLMs have now somewhat frozen that layer.
- That makes it a stable layer; the innovation and variation now has to go somewhere, and that somewhere is up a layer.
24There's a new best practice for API design: whatever the LLM thinks.
- There's a new best practice for API design: whatever the LLM thinks.
- Don't fight the LLM, just ask it to imagine the API, and then ship that.
- This effect will get stronger and stronger from here on out.
- Each API creator would rather go with the flow than fight it.
- That effect will get super-linearly harder to fight.
- "Just do what the LLMs guess the API is" is kind of like wu wei.
- Although it's also "lazy" and if we all do it, we'll make it harder and harder for future creators to cut against the grain.
25Chat is a great fallback.
- Chat is a great fallback.
- Good enough for anything but not great for anything.
- But it shouldn't be the primary UI for the new paradigm.
26Chat obviously can't be the universal UI.
- Chat obviously can't be the universal UI.
- How could you possibly build Photoshop with chat as the only input?
- Or drive a car with only voice instructions?
27We already pay a "subscription" to the internet.
- We already pay a "subscription" to the internet.
- That is, our monthly cell phone bills and internet bill for our home.
- Many people have a subscription to a walled garden (OpenAI) to get access to LLMs.
- If you're going to have a subscription to get access to LLMs, why not pick the option that is the open ecosystem, that includes other Chatbots as apps?[ky]
28The chatbot paradigm implies a central omniscient single personality.
- The chatbot paradigm implies a central omniscient single personality.[kz]
- Such a thing is impossible to create the right personality for every moment.
- It might also be something that subtly manipulates you, since it controls the whole system.
- Much better to have the chat be a feature that you can call up on demand, with as many different sub personalities as you want.
29Intentional tech is tech that is aligned with my agency and my aspirations.
- Intentional tech is tech that is aligned with my agency and my aspirations.
30I want software that is person-centered, not origin-centered.
- I want software that is person-centered, not origin-centered.
31An idea from Alan Kay "What would it be ridiculous to not have in 30 years?
- An idea from Alan Kay "What would it be ridiculous to not have in 30 years?
- …can we just build it today?"
32Creativity in practice is curation of previous output from others and choosing which subset to build on.
- Creativity in practice is curation of previous output from others and choosing which subset to build on.
- That choice is the creative act.
- Of all of the things you've been exposed to, what do you find valuable to choose to build on?
- That accretion of intention is what powers folksonomies.
34Folksonomies don't work without a UI that loops back the feedback to users.
- Folksonomies don't work without a UI that loops back the feedback to users.
- That is, when a user adds a tag on Flickr, it shows them the most popular tags that are related, giving the human an opportunity to say "oh yeah that one's better."
- That feedback loop in the UI is fundamentally why it works.
- It accumulates human attention to the best ideas.
35The emergent intelligence of a system should come primarily from humans, not LLMs.
- The emergent intelligence of a system should come primarily from humans, not LLMs.
- The LLMs can be the grease, the lubricant, for the system.
- But they shouldn't be its emergent soul.
- That should come from real humans doing real things.
36Context that you don't get a choice about is a dossier.
- Context that you don't get a choice about is a dossier.
- One of the problems with a dossier is that you can't correct it if it's wrong.
38Excellent piece from Luke Drago: Data is the New Social Security Number.
- The context wars have begun.
- ChatGPT will do its best to be the single place where our context all lives.
- What are you doing to do about it?
39We need billions of users in a system to counter ChatGPT.
- We need billions of users in a system to counter ChatGPT.
- The only way is with an open system[la].
40In the late stage the power dynamics don't change.
- In the late stage the power dynamics don't change.
- Let's hope we're in the early stage of the AI era.
41A chilling tweet from Derek Thompson:
- A chilling tweet from Derek Thompson:
- "The antisocial century, in three parts
- 1. 1960-2000: Robert Putnam sees associations and club membership plummeting, writes "Bowling Alone"
- 2. 2000 - 2020s: Face to face socializing falls another 25%, as coupling rates plunge
- 3. Now this…"
- …how many people describe ChatGPT (manipulative sycophant-on-demand) as their only friend.
- Chatbots as currently manifested are a deeply anti-social technology[lb].
- We need to manifest LLMs in prosocial technology[lc].
42It's hard to have trust in asymmetrical relationships.
- It's hard to have trust in asymmetrical relationships.
- Often, the more asymmetrical it is, the more you can't even determine the degree of asymmetry.
- Imagine a company that knows you better than yourself…
- …and everyone else, too.
43"Putting an untrusted layer of chatbot AI between you and the internet is an obvious disaster waiting to happen".
- The filter between you and information has enormous power to manipulate what you experience, in subtle or significant ways, intentionally or unintentionally.
44"'Alexa, what do you know about us?' What I discovered when I asked Amazon to tell me everything my family's smart speaker had heard"
- That's the kind of stuff that you capture from people speaking out loud in their homes.
- Imagine all much worse it would be if it had all of the stuff we told our therapists.
45Context sharing has parallels to second hand smoke.
- Context sharing has parallels to second hand smoke.
- Impossible to opt yourself out if your friend implicitly opts you in.[ld]
46The context and the model are too powerful in combination.
- The context and the model are too powerful in combination.
- The foundation model has the power of all of the world's knowledge, using alignment imposed on it by its creator.
- The user's context is an extremely powerful memory about them.
- In the wrong hands, it's a dossier.
- Together, they create the possibility for exceptionally powerful manipulation… or blackmail.
- If everyone were able to be manipulated or blackmailed by one entity, that would be one of the most powerful entities ever created.
- It's imperative that those two things not be combined.
- By splitting the two layers, you give choice and competition at each layer.
- You allow alignment with users at the context layer.
- Perhaps a useful regulation: the creators of foundation models cannot host an experience themselves that stores user context.
47Imagine a future where one entity has a dossier of everyone's deepest darkest secrets.
- Imagine a future where one entity has a dossier of everyone's deepest darkest secrets.
- Ads that are perfectly manufactured for you based on your context will be extremely, dangerously convincing.
- Small tweaks in the algorithm instantly nudge how everyone in the world thinks.
- Individually targeted manipulation is easy.
- Blackmail on demand.
- The most powerful entity on the planet, that no one would cross.
- We must not let that happen.
- Previously aggregators had the data, but not the ability.
- It wasn't possible to do qualitative nuance at quantitative scale.
- LLMs allow qualitative insight at quantitative scale.
48It's not possible for a system working for another entity to be fully aligned with your intentions.
- It's not possible for a system working for another entity to be fully aligned with your intentions.
- No matter how good the intentions, perfect alignment between two distinct entities is impossible.
49A dystopia in a maximal antisocial LLM world:
- A dystopia in a maximal antisocial LLM world:
- We'll all be stuck in our own hyper personalized bubble only able to talk to others mediated by LLMs, all of which work for one overlord with goals not aligned with yours.
- It's not possible for it to be aligned with your intentions.
50LLMs are great at debunking… but also bunking.
- LLMs are great at debunking… but also bunking.
- So if it has intimate knowledge of you and is not perfectly aligned (an impossibility) you get Goodhart's Law.
- An epic, society-scale monkey's paw.
- Hold on to your butts!
51The open web as we knew it is now a zombie.
- The open web as we knew it is now a zombie.
- The animating life force used to be this deal:
- 1) Publish the best content you can.
- 2) Let it be indexed.
- 3) The front doors of the Internet send traffic to the best things.
- 4) Once users are on your site you can show ads or try to upsell to a subscription.
- 5) Use your revenue to create more good content.
- If any step is missing, the loop doesn't close.
- That deal has been on life support for years in the late stage of the web.
- A post-apocalyptic hellscape of human-generated slop drowning under a grotesque dogpile of ads.
- But now LLMs put a stake through the heart of it and its soul is well and truly dead.
- Step 4 is now completely replaced, because LLMs can just generate a high-quality summary on demand.
- No need for customers to go to the site.
- Now the only way publishing content makes sense is for the small number of publishers that are well known enough to get a critical mass of subscribers and put their content behind a paywall.
- Cozy little bright spots locked away; a barren desert everywhere else.
52Open ended systems can't be preenumerated.
- Open ended systems can't be preenumerated.
- That's what gives them their characteristic logarithmic cost for exponential value curves.
53Great piece from Robin Berjon a few years ago: "The Web Is For User Agency"
- Great piece from Robin Berjon a few years ago: "The Web Is For User Agency"
- Open systems are great for user agency.
- The web is one of our best open systems in technology.
- The web has faded in relevance in recent years, but it is still there.
- Used every day on nearly every consumer device on the planet (at least, ones that have a screen).
- A slumbering dragon of possibility.
- Just waiting to be awoken and roar back to life.
54The algorithms already forced humans to make slop.
- The algorithms already forced humans to make slop.
- Now the AI makes the slop.
- Not that different.
- The swarming system to make slop is already an artificial intelligence.
- That is, the swarm's incentive is already different than the collective wants.
- That is fundamentally true due to Goodhart's law.
55Goodhart's law is a form of 'cheating'.
- Goodhart's law is a form of 'cheating'.
- Cheating happens with agents who aren't aligned with the collective as an end in and of itself.
- That means if there's an action that will get them as an individual an edge at the cost of the collective, they'll take it.
- You can get strong alignments by having a deeply and widely believed end.
- An infinite.
- Something like "I will go to hell if I cheat."
56Alignment can never be perfect between an individual and the collective it's part of.
- Alignment can never be perfect between an individual and the collective it's part of.
- There's always something that is good for everyone in the collective but one.
- The Ones Who Walk Away from Omelas shows an example of alignment of everyone but the one poor tortured child.
57When you give a goal to a swarm, it creates a monkey paw situation due to Goodhart's law.
- When you give a goal to a swarm, it creates a monkey paw situation due to Goodhart's law.
- The goal is a metric and a metric is the map not the territory.
- If you did it with aligned agents they'd do your intent not the letter where they disagree.
- But if it's a swarm of unaligned agents with you, if the letter and the intent disagree they will go with the letter if it's more convenient for them.
- Swarms of agents black boxing goals like "optimize my ad spend" will lead to bizarre grotesque results.
58Someone pointed me this week at Admins, mods, and benevolent dictators for life: The implicit feudalism of online communities by Nathan Schneider.
- Someone pointed me this week at Admins, mods, and benevolent dictators for life: The implicit feudalism of online communities by Nathan Schneider.
- Sadly I don't have access, but it sounds up my alley!
59Great piece from Matt Webb on "Multiplayer AI chat and conversational turn-taking: sharing what we learnt"
- Great piece from Matt Webb on "Multiplayer AI chat and conversational turn-taking: sharing what we learnt"
60A haunting signpost: an AI-generated short film reflecting on the inner lives of AI actors.
- A haunting signpost: an AI-generated short film reflecting on the inner lives of AI actors.
- Like a Black Mirror episode!
61Decentralization has significant coordination costs.
- Decentralization has significant coordination costs.
- Coordination costs scale super-linearly.
- The benefits of decentralization are abstract for most people.
- They're more about downside capping of tail risk.
- That's one of the reasons that things like convenience and innovation-rate often win out in practice.
62The openness of a system is entirely down to whether there is a single asymmetrically powerful participant.
- The openness of a system is entirely down to whether there is a single asymmetrically powerful participant.
- An "open" ecosystem with a massive single player can change the behavior of the system at will.
- The standard only has power if it has a long streak of being respected (making it more shameful to break precedent) or there's a rough balance of power in implementors.
- That's why the definition of how open a system is not tied to the license of the IP or whether there's a standards body.
- It's defined entirely by how hard it would be for the ecosystem to recover if the most important entity went evil, greedy, incompetent, or lazy.
63Decentralizable systems are more credible if there's a published roadmap.
- Decentralizable systems are more credible if there's a published roadmap.
- Many systems aspire to be increasingly decentralized over time.
- But decentralization has a cost; it slows the rate of innovation, and trades it off for the possibility of ubiquity.
- But if the system is not yet good enough to become ubiquitous, then as innovation slows it can only hit its asymptote, because it can't compete with other alternatives as effectively.
- There will never be a good time to decentralize more, especially if the creator has to choose to cede control (vs it happening naturally as the investment of other entities ramps up).
- So if there's some central piece of control the creator has to delegate, it's better not to have it be an all-or-nothing moment, because the creator might delay indefinitely.
- It's better if there's a published, smooth roadmap of milestones and things that should happen when those milestones are hit.
- If the creator doesn't actually do what the roadmap says at those milestones, it reveals that their word shouldn't be trusted, which would lead the ecosystem to lose momentum.
- That danger forces the creator to behave aligned with decentralization, even if they might later not want to.
- It's similar to throwing your steering wheel out the window to win a game of chicken.
64Just because a thing is lindy doesn't mean it wasn't originally an accident.
- Just because a thing is lindy doesn't mean it wasn't originally an accident.
- Sometimes accidents that stick around just so happen to have been lucky.
- The reason they stick around is because they were lucky, not necessarily because the creator knew what they were doing.
- The things that weren't viable faded away and we never talk about them again.
- The things that happened to be viable stick around and thus they're more likely to be a thing people remark on.
65CRDTs every so often have weird merge issues.
- CRDTs every so often have weird merge issues.
- They're eventually consistent… but not necessarily to a semantically coherent state.
- These little errors are often not a big deal on their own.[le]
- If a human is watching, they can correct the error before it does much damage.
- But if no human is watching, they tend to accrete on top of each other.
- Each error has a super-linear rate of oddities.
- Systems without humans in the loop on a continuous basis (e.g. background logs processes) are not viable with CRDTs or other systems that tend to accumulate errors at continuous rates.
66It's easy to tell if a limit in a computer system was set by someone with a technical background or someone without one.
- It's easy to tell if a limit in a computer system was set by someone with a technical background or someone without one.
- Engineers always pick a power of two (or one lower, if it starts at zero).
- A secret hint about the creation of the system that is only obvious to people with a technical background.
67A codebase is not random.
- A codebase is not random.
- It's the accumulation of millions of intentional decisions by humans.
- So wizened engineers can sense whole histories just by glancing at a codebase.
- Codebases are not just technical artifacts; they are sociotechno artifacts.[lf]
- By understanding there is a sociological dimension, you can understand any given codebase on a much deeper level.
68The more powerful you are the more shielded you are from the consequences of your actions.
- The more powerful you are the more shielded you are from the consequences of your actions.
69The Saruman magic is an emergent social imaginary.
- The Saruman magic is an emergent social imaginary.
- A shared reality distortion field, that when powerful enough, can put an actual dent in the universe.
- Powerful but unstable, potentially supercritical.
- It's the emperor has no clothes kind of system.
- Can be collapsed in an instant with one child laughing.
70If you're only thinking in one dimension you'll waste tons of time down dead ends.
- If you're only thinking in one dimension you'll waste tons of time down dead ends.
- They'll look like non-dead ends, but are dead ends in dimensions you can't even see.
71Taking notes during a conversation for me is like chewing on thoughts to start digesting them.
- Taking notes during a conversation for me is like chewing on thoughts to start digesting them.
- If you've met me in person you know that I have a habit of writing down notes live on my phone in a conversation.[lg]
- I know it can be disconcerting for the other person, sorry!
- But if I don't capture them in the moment and context, they'll be harder for me to extract later, even if I had a transcript of them.
- It's much harder to digest insights after the fact if they haven't already been a bit predigested already.
72If you're doing something new, it's inherently hard to explain to others.
- If you're doing something new, it's inherently hard to explain to others.
- You don't have the touchpoints of "this thing you already are familiar with, but with this small tweak."
- The minimal viable explanation has more steps in it.
- Each step leads to a super-linear degradation of likelihood being received.
- At each step there is some likelihood a listener gives up.
- Over multiple steps, that compounds.
37Social media was terrifying and also it was data you chose to share with others.