my most commonly repeated prompt; would be nice if the baked it into the tool itself:
"No emojis. be concise. no suggestions unless I explicitly ask for them. answer questions like the machine you are. Don't try and add personality or humour; remember you're a robot."
It really does make you wonder why all the models seem to require that. In principle, it shouldn't be a property of LLMs, and lol no it's not an "emergent property".
Post-training and "human preference" according to "data". Don't know a single developer who use these tools for work who prefer that though, but also don't know anyone who use LLMs a lot just "for fun" either, might just be vastly different preferences between the two userbases.
These days announcements like this just make me want to put on my tinfoil hat - what's in it for Google, though? Why make it more convenient for people to submit webpages to you?
My prompt collection lives in three different places right now — Raycast snippets, Apple Notes, and a Notion page that keeps growing. I know I wrote a good one for my git commit/push flow somewhere, but finding it when I need it usually takes longer than just rewriting it.
The browser approach makes sense for Claude code and ChatGPT. I wonder how well it holds up once you have 50+ prompts though — finding the right one fast is the real problem for me.
So much of the web has no API anymore and is hostile to robots.
The script to turn the coffee maker on when dad posts on Facebook for the first time each morning that worked in 2014 won't work anymore in 2026.
Having this sort of thing built into a mainstream browser will open up a new avenue for automation, which I think will be a good thing for breaking down data silos and being good for the world overall.
I hate that. I understand that it might be useful, and tbh, on personnal PC, i'm not even concerned. But it is going towards people pushing to replace XQL or other query languages with prompting in natural languages, for no good reasons. Generate your query and copy paste if you don't want to read the documentation man, but please, please keep an intermediary between the LLM and the real world data. The last time your fucking prompt gave me a "log overview" i lost 2 hours understanding what the fuck i was reading, when a query would have taken me at most 20 minutes.
Convert my AI prompt into the code for a one-click tool, let me read and share it, that would be _great_.
I can imagine a moderator, or a marketing person, wanting such a tool. "Respond to this post in a polite and friendly manner, thank the user for choosing our company, discovering a problem, and taking the time to report it. Promise to sort this out quickly. If the user is really angry and threatens legal action, promise an immediate refund, and shoot me an email with the summary of the issue, and all the details."
If instead of a copy-pasting spree, or setting up a whateverClaw, the user might just click a button in Chrome, it could be actually useful. (Consider a dozen such buttons.)
This sounds to me like yet another way to automate filling out forms. I had been thinking about vibe-coding a Chrome extension for one form I fill in regularly, but perhaps this is easier.
There is something comforting about seeing that the SV stopped having ideas and now just recycles and recombines the same tropes over and over again.
It's still all terrible, but it's a devil you know. You can live with that. You can skip the broken stair and duck, knowing exactly when they're trying to punch you in the face again.
Now here's hoping that eventually, they get bored and just stop entirely.
I would be more excited by this if there was a better permissions model for these things. For example I can think of a skill that would need access to a certain corpus of documents that I host on Google Drive, but, as far as I have been able to determine using Google's other AI products, there is no way for me to grant read-only access to that corpus without granting read-write access to all of my data on Google, which is simply too much access for my taste. There has to be something less binary than Personalization:on/off?