Nichess is a game like chess, where pieces have special abilities and health points. This allows for much finer balancing and many more variants compared to the original chess. It will take some time, but it will become great eventually.
My initial motivation was wanting to RDP and SSH into my home workstation from a locked-down corporate laptop when I travel. I couldn't install Tailscale on the laptop, and I didn't want to pay for a cloud VM just to do SSH port forwarding. Now I use it to tie together half a dozen machines, both locally and on Hetzner & Linode. I can SSH and RDP into remote machines, host a git repo on one machine and access it from the others, and (optionally) share files across all of them on a local mount.
You run a hub (telahubd), register machines with a lightweight agent (telad), and connect from anywhere with the client (tela). All three are single Go binaries with no external dependencies. The hub never sees your traffic. It just relays opaque WireGuard ciphertext.
All binaries run on Windows, Linux, and macOS. There is also a desktop GUI app, TelaVisor, that wraps the client and enables remote management of hubs and agents.
It's Apache 2.0-license and pre-1.0 release, but I'm polishing it for a stable 1.0 release in the next month or so.
I'm also working on an enterprise-grade management portal that works with Tela, https://awansaya.net/
I'm not sure it would work but did you try running tailscale client through a docker container so it's not installed directly in your host system?
It's funny... I've started using so many of the nifty management features of TelaVisor and Awan Saya that I am now considering adding lower-level support for the features that I explicitly wrote for user-space.
to access my home desktop machine, I run:
``` $ ssh itake@ssh.domain.me -o ProxyCommand="cloudflared access ssh --hostname %h" ```
and I setup all the cloudflare access tunnels to connect to the service.
Tela takes a little different approach. The agent exposes services directly through the WireGuard tunnel without SSH as an intermediary, so you don't need sshd running on the target. Each machine gets its own loopback address on the client, so there is no port remapping.
The big difference is the relay, though. With cloudflared, Cloudflare terminates TLS at their edge. With Tela, you run the hub yourself and encryption is end-to-end. The hub only ever sees encrypted data (apart from a small header).
I went with WebDAV because it works on all three platforms without a kernel module or extra driver. For my use case (browsing files, grabbing configs, etc.) it works well enough.
Bi-directional sync is an interesting idea. Right now the sharing is one-directional (the agent exposes a directory, the client mounts it), but I could see adding something like that as a layer on top.
It's just an mmWave sensor connected to an ESP32. But it works nicely, and I'm thinking of starting a company making them, though I'm not clear if the elderly would be ok with this minimal (no camera) intrusion.
It would just work out of the box.. the real one would have a small cell modem so it wouldn't need any networking setup, and it would act as a gateway if you have more than one in a house. There are industrial versions of this for nursing homes. This would be a bit more warm and fuzzy for home use.
In elder care, I am building https://statphone.com - one emergency number that rings multiple family members simultaneously and breaks through DND. Would love to chat/collaborate.
The question of "intrusion" was always interesting to me because old folks often face going from nothing to assisted living or nursing home which is often quite intrusive, where somewhat ironically adding a bunch of sensors to your home allows you a bit more privacy.
Kind of a tangent, but I like your type of system as an alternative to the emergency pendants. It always struck me as strange to expect old folks at risk of fall to remember to charge and wear a pendant at all times.
My FIL, in his late 80's was living at home alone. My wife used a monitoring service, provided by local package delivery company. They installed motion sensors in the toilet and on the door. If no motion detected for 24 hours, the company will alert my wife by phone and send the nearest delivery driver to check on him.
I myself have tried Home Assistant setup on Raspberry Pi and variety of sensors for different purposes.
I wish I could use auto completion for building muscle. Maybe a Large Muscle Model? (Joke).
Unlike similar apps such as Focus Friend or Forest, which use active timers to police screen time, my app is an inversion that works like an idle game; All screen time is tracked all day, (with double the punishments at night), and upon check-in, you get feedback on your device usage.
I sort of got inspired to do this after seeing so many QC PR posts on HN, and finding the educational material in this space to be either too academic, too narrow in scope, or totally facile. I think given the incredible hype (and potential promise) of this industry, there should be on-ramps for technically minded people to get an understanding of what's going on. I don't think you should need to be a quantum physicist to be able to follow the field (I am only an electrical engineer).
My book tries to cover the computational theory, the actual hardware implementations, and the potential applications of quantum computers. More than that, I want to be unbiased and stray away from what I feel is misleading hype. It's been a work in progress for about 6 months now, with a lot of time spent gaining fluency in the field. But the end is in sight! :)
FWIW, my shallow understanding of quantum computing as a programmer, in case you wanted perspectives from your potential audience:
- I thought quantum physics was a sham? Like on par with string theory. But apparently that's not true
- I hear QC only breaks certain kinds of cryptography algorithms (involving factoring big primes?), and that we can upgrade to more foolproof algorithms.
- I hear that one of the main challenges is improving error bounds? I'm not sure how error is involved and how it can be wrangled to get a deterministic or useful result
- Idk what a qubit is or how you make one or how you put several together
The idea was inspired by Mailpit, which I've used for years to debug outgoing emails. A few implementation details were literally stolen from Sentry SDK with an "implement it how Sentry does it" prompt.
After winning the Playlin Player's Choice award I've noticed an uptick in players as well as some people sharing videos on YouTube which has been fun. I've got a few thousand people playing every day.
I just launched user accounts today so user's can now track their progress across devices and share their stats with each other. This ended up being a bigger chunk of work than I expected but I'm really pleased with how it turned out. (Though I launched it 15 minutes ago so I'm holding my breath for bug reports)
I'm fine-tuning my internal puzzle-building now with the goal of letting people use them to make and share their own puzzles soon!
I'm not sure if it would fit the theme, but sometimes I end up searching what an expression means, or where does it come from. Maybe it would be cool to have a little info box after you discover what the word is. Just an idea! Not sure if it would clutter things, and you can always search it yourself, but something I've been thinking about. I still remember looking up peanut gallery and sand dollar!
Just tried it out on my browser. Will be following this.
Also would love to see your workflow you spoke about, on coming up with puzzle ideas and tile arrangements. Cheers!
Thanks for making this and I wish you all the success in the future.
would be super interested to hear more about the puzzle-making process too, is it fully automated with AI at this point or is there still a good amount of manual work and fine-tuning involved?
bookmarked already, can't wait to play tomorrow again
It’s a lot of manual work right now. I don’t use AI in the process. I think it could help with some of the brainstorming but I kind of like the human connection of making a puzzle and having people solve it.
Here’s the basic process.
My wife and I do this part together:
- Think of a theme
- Think of words related to that theme, ideally with a second meaning
- Think of clues for those words
Once we have a good set of clues I plug them into a program I wrote to make crosswords.
The program isn’t that smart. It tries making random crosswords. I run it 1500 times and then sort the results to get the best ones. This brute force approach works pretty well for how simple it is.
I pick the crossword I want and then I use another tool to split up and rearrange the tiles. This step could probably be automated but there’s some finicky logic to the best way to split up the tiles and it goes pretty fast manually.
I’ve been meaning to make a video of the process! I’ll share it here when I do
The design and dev took a while but building the has been the most time consuming at this point. My wife and I make the puzzles together.
We’re getting close to 6 months of daily, hand crafted puzzles!
https://github.com/jcubic/speaking-clock
It uses local AI models for the voice.
When booking flights, I use sites like Kiwi and Skyscanner that let you do flexible searches - multiple destinations, custom connections, creative routes, etc. But rail search feels oddly constrained. All the UK train operators offer basically the same experience, and surface the exact same routes. I always suspected there were better or just different options that weren’t being shown. Where is the "Skyscanner for trains"?
After digging through the national rail data feeds, I decided to have a go at building my own route planner that runs completely offline in the browser. This gave me the freedom to implement more complex filters, search to/from multiple stations, and do it without a persistent network connection.
Now I'm finding routes that aren't offered by the standard train operators, connecting at different stations, and finding it's often easier to travel to different stations (some I'd never heard of) that get me closer and faster to where I actually want to go!
It's still a little rough and I'd like to add more features such as fares, VSTP data, and direct-links to book tickets, but wanted to share early and get some initial feedback before investing more time into it. So, thanks in advance - let me know what you think.
I sent you some feedback on a routing failure because I didn't want to post exactly where I live here.
I think you need pricing. Works offline is cool, but why not pull in the pricing if people are online? Train fares are so variable depending on time of day, especially if they go via London. I could have a trip that could be £300 cheaper by taking a 30 minute longer trip that avoids London. I need pricing to get my best journey.
I'm looking at how to add price data to railraptor, but it might mean sacrificing the fully-offline capability... once I have prices it should absolutely be possible to build a filter along the lines of "find me the cheapest popular destinations that are at least 50 miles away".
- Introduction: https://poyo.co/note/20260318T184012/
- Tool loops: https://poyo.co/note/20260329T034500/
- Playing with receipt extraction: https://poyo.co/note/20260323T120532/
- Use with async flow: https://poyo.co/note/20260410T164710/
developing AI agents that are easy to integrate in to websites, based in Europe and all data stored and processed in Europe to complying with regulations.
Looking forward to collaborations, and happy to talk with anyone would like to collaborate with us
You say this is a company you could see yourself working at for some time, and have been handed C suite level responsibility that you can handle. So seemingly you are content and able to handle the work load.
Learning to be a IC is something anyone can do given time, but learning to be a manager can only be learned by being on the job, if you are able to get it in the first place.
Now is really not a good time to jump ship, unless you know for certain that the new position is going to be stable.
Grab the opportunity, do a good job and perhaps study how to be a better IC in your free time. You'll come out on the other side with skills and experiences that many in this field will be missing.
a texteditor with spreadsheet like formulas - does this even make sense? super super early buggy release - feedback welcome - any feedback, thx
Then slid it a few hundred feet across the lawn on composite deck boards we salvaged when we took a balcony down last year and landed it atop the new piers.
Then put the electric fence back up to keep the bears out.
Presently? A beer.
Looks like it is a crowded area now - my angle is to start with theory of what is important in a system like that, from first principles (like agent limited context, statelessness, use goals etc). Currently I use it to develop that theory - and you can read it at: https://zby.github.io/commonplace/. I also use it to keep an index of similar systems (that is systems with agent operated memory): https://zby.github.io/commonplace/agent-memory-systems/
The github repo is at: https://github.com/zby/commonplace . Work in progress.
I love cooking but the daily "what do you want" grind was killing me. Rushing to the store after work hoping for inspiration but leaving with the same five fallback meals. Recipes using half a box of something so you eat the same thing twice or watch leftovers die in the fridge.
The final straw was our newborn's milk protein allergy, turns out milk is in everything. Recipe sites are hostile. Ads reload and jump the page mid-sentence, 20 versions of every dish, comparing the 4.7 star rating version with the 4.8 star one. So you go by thumbnail. Visual clutter everywhere.
I tried the apps. One does swiping, one does shopping lists, one does Sunday budget planning, one has "what's in my fridge" mode. Pick your half-solution.
So I built what I wanted: swipe mode that makes picking dinner fun again, or instant 3 quality suggestions for when I am in the store. Aisle oriented shopping list, budget, personal taste, fridge inventory in one place. UI looks like a restaurant menu — off-white, black text, no glossy photos. I'm working on AI mode now. Not for recipe generation, which are mostly garbage, but for search and substitution.
- Pick mode for when you're in the store looking like a deer in headlights at the produce section. It gives you 3 solid options instantly.
- AI mode (WIP) for "something with chicken, but I also have carrots in the fridge that are going bad."
Plus aisle-sorted shopping lists for everything. No more backtracking at Aldi.
For me, having a selection of high quality recipes would be important. For more experienced cooks like my husband, he would just tweak on the fly or use his own recipe anyhow and would enjoy being able to plan with the household and have a shopping list.
Good luck with the project!
Thanks a lot.
WIP, started 2 weeks ago: https://skyshift.rudidev.com/maps/stable
I live in Lisbon and I've been learning Portuguese with a tutor since 2022. After every lesson I'd sit down and make flashcards from my notes and screenshots. Spaced repetition works, but making the cards took manual effort each time. Most days I just didn't do it. So I have automated that process.
The flow: you invite the Kardo bot to your call, it records and transcribes (Recall.ai + Deepgram Nova 3), then GPT-4o extracts vocabulary from the transcript and generates cards. You review them with spaced repetition — we use FSRS, which is the best open algorithm I could find. If you already use Anki or Mochi Cards, there's export.
You can also throw in YouTube videos, podcasts, articles, PDFs — not just live lessons.
Tech: built entirely with Claude Code. React + Vite frontend, Bun + Elysia backend, Convex for the database, Railway for hosting.
We got 50 beta users through Telegram, and just landed our first paying customer. Now we're trying to figure out distribution — tutors seem like the obvious channel because one tutor recommends you to all their students, but reaching them with zero marketing budget is the hard part.
Curious if anyone here learns a language with a tutor and what your review workflow looks like.
I'm still obsessed with making my game, which you can try it at the link above (it is desktop only). This is my first "real" game, and it has been incredibly fun and rewarding. I've been working on it in the evenings for about 4 or 5 months.
It is a very ambitious mix of genres - shoot-em-up and deck-building. A lot of people said that those are genres that shouldn't be combined, but I think it turned out to be a fun little game. Folks who are not fans of one (or either) of the genres are actually playing it. I built a global high-score leaderboard, and there are people (including a few of my friends) competing on it. Whoever knocks my friend "BER" from first place will earn a beer from me.
This is purely a fun project, although I'm now seriously considering releasing it on steam when I finish everything I planned for it. It is made in Kaplay, a small JavaScript gamedev library, which is a big part of what makes it fun. If you try it out, please leave a comment, I would love more feedback!
Loved the music.
Didn't know what was going on half the time.
Positively overwhelmed.
Thanks for that little spark of joy!
If I ever release on steam, can I please use your comment in promo material? I would anonymize it of course.
Edit: grammar
I built HeartRoutine to help me lower my LDL and ApoB. I recently started beta testing it on some friends, too, to see if anyone other than myself would find it helpful: https://www.heartroutine.com/.
I've started building the combat prototype for my next game, Today I Will Destroy You, inspired by my love of going-on-an-adventure-with-a-sword games and Sekiro-style combat.
I've committed to keeping my personal website up to date: https://piinecone.com/.
Not OP. In theory? No. Takes a second to change it. To be quite honest, its yet another thing to keep track off and do. I know, for myself, I would remember to do it for a few days and then forget.
Its a tiny thing but the more I can outsource the better. My brain is occupied with enough other stuff.
It's a top down ARPG called Mechstain where the player creates and pilots voxel based mechs
Instead of traditional gear, your mech has a physical voxel footprint that you the player have to fit weapons and components inside
Your job is to manage space, power and mass, what you can fit and power directly becomes your stats and abilities, essentially a bin packing problem
Basically take Diablo 2 and remix it with Kerbal Space Program, still fleshing out the various systems, but I'm really enjoying the process of slowly designing systems, iterating on it and fleshing it out
It's quite fun taking thoughts I've been noodling on for years and trying to figure out if they synergise with what I'm looking at and do they provide interesting player decisions
Recently onboarded a 3d artist and it's really making things look a lot better
If anyone has experience lighting + vfx in this sort of game, I'd love to talk to them, still trying to figure that out =)
To resolve this, I am currently fleshing out the idea for a "shared hosting" for vibe coded programs - something like a cross between an old school LAMP stack shared host and a parse like library for capabilities like push notifications.
It's all very half baked in my head at the moment - with the biggest problem being a safe way to deploy remote code without pawning the server, but this is a problem shared hosts have dealt with and I am sure I will eventually figure out a way.
The end goal is to be able to have people tell their AI agent of choice to "make their app deployable" on our platform - and the agent will adapt it to our library methods and deploy automatically. Once done folks will be able to access their programs from any internet connected device.
Air quality data is now very widely available, but managing access to multiple networks is a massive task (lots of shonky APIs out there - the EEA has a csv endpoint that actually returns a .zip with mimetype "text/html", just to give you a flavour). Integrating new APIs could be a full-time job, but it's something AI can do very well given a pattern.
This is really for me as I build out my company working on turning air quality data into actionable information, but it's open source and freely available.
I want to show how I liberate poorly aligned, pixelated PDF image scans of century-old Latin textbooks from the Internet Archive and transform them into glorious Org mode documents while preserving important typographic details, nicely formatted tables, and some semantic document metadata. I also want to demonstrate how I use a high-performance XML database engine to quickly perform Latin-to-English lookups against an XML-TEI formatted edition of the 19th century Lewis & Short dictionary, and using a RESTXQ endpoint and some XQuery code to dynamically reformat the entries into Org-mode for display in a pop-up buffer.
I intend demonstrate how I built a transcription pipeline in Emacs Lisp using tools such as yt-dlp and patreon-dl to grab Latin-language audio content from the Internet, transcode the audio with ffmpeg, do Voice Activity Detection and chunking in Python with Silero, load the chunks into Gemini's context window, and send it off for transcription and macronization, gather forced-alignment data using local a local wav2vec2-latin model, and finally add word-level linguistic analysis (POS, morphology, lemmas) using a local Stanza model trained on the Classical corpus.
This all gets saved to an an XML file which is loaded into BaseX along with some metadata. I'll then demonstrate some Emacs Lisp code which pulls it into an Org-mode based transcription buffer and minor-mode for reading and study, where I can play audio of any given Latin word, sentence, or paragraph, thanks to the forced-alignment and linguistic analysis data being stored in hidden text properties when the data was fetched from the database.
Lastly, I'd like to explore how to leverage these tools to automatically create flash cards with audio cues in Org mode using the anki-editor Emacs minor mode for sentence mining.
This is my first Minecraft Mod and the first project i made that interacts with the network, has logins/accounts and does APIs.
I am really not good and thus the UI (far worse on mobile) and especially the Code is bad, but i would have never expected to get it working at all, much less this functional. But i am still far from done, i still want to improve the overall code quality, add the inventory and ender chest, achievements (maybe even custom ones so vanilla clients can earn and view them without having to change anything locally. IDK yet) and more.
If someone wants a small demo, i have it running on my server to test while I am developing at: https://grisu-ftp.de (If you find any issues lmk :)
While this is by far not as cool as the other stuff on here i still would like to show it off and gather some first feedback. This is my first Java project that goes above the Standard stuff in school like scanners/calculators and so I have probably done obvious beginner mistakes.
Here's how cbs.h builds itself: https://codeberg.org/Luxgile/cbs.h/src/branch/main/cbs.c
I've been running models on my homelab for a bit now, but none of the available options out there was what I wanted. I wanted something that I could command from the CLI, API, or Web, so have an agent go in and do work remotely via SSH or myself via a web interface.
I wanted the ability to know if models have been updated, and if backends (llama.cpp, ik_llama.cpp) have been updated, see what those updates are and choose to update. Also wanted the ability to switch betwen versions of those, so if I felt like there was a regression, or performance issue, I could roll back.
I've also published plugins for OpenCode and Pi so that model discovery is automatic too.
I'm building this mostly for me, as usual.
So far I first sniffed the BT logs from my iPhone but couldn't figure out how authorization works. Recently decided to decompile the android version and with some LLM help I made some progress. Been too busy to test it out but once I crack the authorization I can get started with writing my own Watch app
I also built https://statphone.com - One emergency number that rings your whole family and breaks through DND.
Do you have a personal blog or github page?
The use case is kind of neat. RAID is great and can mostly solve these problems, but people don't have SATA hardware that can handle the workload well, plus they aren't ready to manage an array like that, and they don't like having to use specific sized drives, etc. Another major issue with those setups is you need to be careful because an IO error that you don't recover from will be very difficult or impossible to recover from because of the layers of LUKS combined with LVM.
With MergerFS you just use regular file systems that are separate, but they get combined into a single mount point. That means each disk can just be a different LUKS encrypted drive and if you need to recover data it's isolated to that one disk and much more manageable. You can also take any disk and plug it into another machine as needed and grow or shrink the storage pool as needed.
MergerFS has options and settings to help you determine how files are spread across the drives, such as least space used or which disk has the most of that directory path already.
My app (Chimera) automates the unlocking of the disks, mounting and some data migration if you want to remove a disk from the pool. I plan to add some rclone features to help provide easier backup options to places like Backblaze, AWS, or a remote server in general.
So far so good and I was surprised at how well Opus had been handling Gtk and pkexec.
Let me know if you guys are interested I am close to pushing some RPMs and DEBs, in addition to the standard Python stuff.
So I built an on-device OCR engine (PaddleOCR) that reads screen text locally and feeds it into an AI sentiment analysis pipeline. No screenshots leave the machine. We now get alerts if there's detection of concerning interactions. The client is written in Rust, with DNS filtering, game detection (Steam/Wine/Proton), and screen time enforcement built in.
It started as a home project that worked really well. My wife suggested other families wouldbenefit, so I've been building it out as a product. The client shipped on Linux first, we're a Linux gaming family, with Windows coming soon.
There are many more features I haven't touched on. Would love feedback from other parents who've dealt with this space. The goal is to protect children and empower parents with tooling that's transparent and effective.
Good catch on the GitHub link, that's a bug, I'll get it fixed. I'm planning to open source the client codebase and push it to GitHub in the near future.
I'll post updates on the site as clients become available. Appreciate the interest!
Ok in all seriousness, right now I'm tracking down an issue with the ENA network interface which results in sporadic packet loss. Triggering the issue is hard and seems to require a large number of TCP segments being pushed to the NIC very fast. So far I've found that my reproducer stops reproducing when I turn off write combining on the MMIO space used for low latency queueing, which is... just a little bit weird.
But seriously, good luck!
Imagine mixing Magic: The Gathering, StarCraft and Civilization’s hex grid combat.
There’s multiplayer but I haven’t put the server anywhere yet.
Check out the introduction here:
https://github.com/williamcotton/space-trader/blob/main/docs...
Clone the repo:
npm install
npm run dev
There’s maybe a couple of other games called Space Trader so if anyone has any suggestions for a new name, I’m all ears!Model output volumes mean that code review only as a final check before merge is way too late, and far too burdensome. Using AI to review AI-generated code is a band-aid, but not a cure.
That's why I built Caliper (http://getcaliper.dev). It's a system that institutes multiple layers of code quality checks throughout the dev cycle. The lightest-weight checks get executed after every agent turn, and then increasingly more complex checks get run pre-commit and pre-merge.
Early users love it, and the data demonstrates the need - 40% of agent turns produce code that violates a project's own conventions (as defined in CLAUDE.md). Caliper catches those violations immediately and gets the model to make corrections before small issues become costly to unwind.
Still very early, and all feedback is welcome! http://getcaliper.dev
After 4 years of maintaining an Open Source Design System, I needed a better way for theming than Sass and PostCSS. I needed the power of a full-featured programming language. That's how the first version of Styleframe was born.
My vision is for Design Systems to be endlessly configurable and composable, like you would configure any library, with or without AI. Want to change your entire website to look like Linear? Simply install and use the Linear Design System configuration. Want only your buttons and cards to look like Linear, and the rest be the default theme? Use the button and card composable functions from that package.
Styleframe is built as a transpiler-first system. You write your design tokens, selectors, utilities, and recipes in TypeScript, and Styleframe tokenizes everything into an internal representation. From there, the transpiler generates dual outputs:
- CSS output: variables, selectors, utilities, themes, keyframes
- TypeScript output: typed recipe functions with full IDE autocomplete (with an optional Runtime)
This architecture means you can have complete control over customizing how your system is output. You could even use the generated tokens to render documentation for your design system components. The output code can be integrated with any headless UI Components Library, or with your own custom components.
Fun fact: I've reimplemented the entirety of TailwindCSS using Styleframe's utility and modifier tokens. Not only that, but I've also built a scanner which picks up your CSS Classes from your markup. It's basically Tailwind, but 100% configurable (even utility class format can be changed), and is always based on your design tokens.
https://truetrials.substepgames.com
I'm a long time fan of the Trials[1] game series, and it's sad that we might never see another trials game from RedLynx[2]. On the other hand, it's a great opportunity to make it myself.
It's going to be free to play, web based, running on 10yo hardware, with open leaderboards and ability for users to create custom levels.
[1]: https://en.wikipedia.org/wiki/Trials_(series)
[2]: https://www.reddit.com/r/TrialsGames/comments/1i0qetb/has_th...
Gameplay feedback: I'm a pretty decent player at the original games and I couldn't make it over a single obstacle, the controls seem extremely sensitive/abrupt currently.
What annoyed me about Trials is the artificial assist and magic forces that let you control bike midair and climb vertical walls - True Trials won't be like that, every force is derived from friction, spring compression, and weight transfer.
Making physics feel right is a hard part indeed! Balancing leaning stiffness/damping of every angle and rider's joint is tricky. As you mentioned that controls are sensitive, it's necessary to clear high jumps (e.g. first checkpoint in course3 currently). It's surely takes time to get used to.
I have a Discord channel where I will post updates: https://discord.gg/wtcZ5q5zHN.
I was reading the fantastic Crafting Interpreters book, and been wondering what it would be like to design a language from scratch. I really enjoy using Sorbet with Ruby, so wanted to design a small language with Ruby's object model, and a gradual type system.
Despite not knowing much programming language theory, I was able to make a surprising amount of progress over a couple of weekends using Claude Code, including building a simple version manager for the language - https://github.com/sapphire-project/facet
You can read more about it over at the site, but it allows you to construct and validate arguments in a graphical form, and it has truth/proof propagation so you can see whether a conclusion is currently considered valid or contested. You can create counterpoints where you think the argument breaks down, and strengthen arguments from there. Some upcoming plans are to allow users to validate arguments for themselves, like mark which parts they understand and agree with so they can collapse that part of the graph, and to add more mcp capability so that LLM can help you construct and validate new arguments.
The idea is that it provides all the geometry to enables games like these to be built: (These are just rough demos)
https://www.robinlinacre.com/letterpaths/writing_app/snake/
https://www.robinlinacre.com/letter_constellations/
And here is like the admin/demo: https://www.robinlinacre.com/letterpaths/
And, separately, I made an educational country quiz, again FOSS:
Couldn't get the letter constellations working on my end.
Country quizzes is a weak spot of mine, loved that. Would be cool to move the globe! Also, kudos for the bus cataloging!
And examples of games is can power:
https://rupertlinacre.com/maths_vs_monsters/
Shamelessly trying to attract new monthly sponsors and people willing to buy me the occasional pizza with my crap HTML skills.
At a first glance it's a mobile proxy service, but the entire backend of it allows anyone to create their own mobile proxy and be able to access it anywhere through the internet, seamlessly. That's the 2nd phase of the website which is still long ways to go, but very happy with how stable the platform is and how fully it's automated.
Tech stack is a bit unconventional for a public facing website, as it's Blazor Server. As a C# dev in my day job, I've found Blazor to be quite capable and stable to quickly iterate through my ideas. And was pleasantly surprised to see how easily I can deploy the app into a Linux VPS through docker, which I didn't think was possible a few years ago.
So I am creating an ambitious app that uses agents
Admin: -> handles all financial transactions and manages the app
Subscriber:-> entity who orders/shops
Market: -> the agents that work with the farmers or markets
Catering: -> for any processing or recipes
Delivery: -> handles cold chain, delivery, storage
Initially I will do everything but the idea is to delegate the agents
The basic structure is in place
Collection of 15 diagnostic tools (VPN leak test, DNS checker, port scanner, etc.) built after a WiFi security incident. All client-side, no data collection.
Feedback welcome!
It's an amazing source of long things to read. There is so much stuff worth reading that has been posted in several decades of blogging.
It's a free USCIS form-filling web-app(no Adobe required). USCIS forms still use XFA PDFs, which don’t let you edit in most browsers. Even with Adobe, fields break, and getting the signature is hard.
So I converted the PDF form into modern, browser-friendly web forms - and kept every field 1:1 with the original. You fill the form, submit it, and get the official USCIS PDF filled.
I found out SimpleCitizen(YC S16) offers a DIY plan for $529 [2]
So, a free (and local-only) version might be a good alternative
Petrify is a machine learning model compiler for the the JVM. It reads your model from an ONNX or other model format, walks the Trees or Linear models, and encodes the model in equivalent JVM bytecode as a stateless class you can invoke.
This differs from every other ONNX Runtime that I know of, which are essentially interpreters. The ONNX Runtimes are also huge (90+mb!?!), JNI, and drag gargantuan dependencies!
This just compiles your models to native bytecode. Much simpler and you end up with 0 dependencies! (you need one interface technically, but I digress).
Do you have any benchmarks?
Petrify will also be order of magnitude kinder to your Garbage Collector, which will increase performance in high-throughput situations. You're also not loading 10 gazillion classes, as your models are directly represented as a first-class Java Class.
The real goal here was the getting rid of dependencies! While thankful for the incredible (and free) work of the authors of the onnxruntime for Java, the primary onnxruntime jar a boat anchor; weighing is 90mb+ just by itself, not counting any of its dependencies.
Once you compile your models with Petrify, you have exactly one 6.9kb jar as a dependency essentially just carries the Fossil interface as an entry point to call your model. I licensed that jar ASL2.0 for maximum compatibility in a corporate environment.
Written in C++ and Slint, it was also a testbed of slint as an UI framework. Having used wxWidgets in the past, and Qt recently, it is certainly a different thing. I just wish there was a native C++ alternative to slint.
I need to integrate the CI to produce binaries, but you can compile it yourself for now.
This month we're focused on:
- first-party, native DMS integration;
- provider-agnostic agentic workflows; and
- enterprise-grade redlining
But of more interest to this group is probably our blog! Our latest post is about Gary Kildall's blunder quibbling over an NDA redline with IBM who was looking to give its entire enterprise away: https://tritium.legal/blog/redlining.
I have a recursive ascent code generator with a bunch of optimisations that I wrote about [1,2]; it's a linear-time parser for LR(1) with reduced overhead. I have an RNGLR implementation (a polynomial-time parser for any context-free grammar), that's still a table-based interpreter like more LR-based parsers out there. I've extended that implementation with special code to handle cycles more efficiently. Some day, I'll take some time to write a paper on that and publish it. Currently, I'm trying to combine the two ideas and create a generalised recursive ascent code generator. If I succeed I'll write another blog post again, it's been a year since the last one...
[1]: https://blog.jeffsmits.net/optimising-recursive-ascent/ [2]: https://blog.jeffsmits.net/optimising-recursive-ascent-part-...
I had an insight the other day, that as I fix the n least (and most, it's a palindrome!) significant decimal digits, I also fix the remainder from division in 5^n. Let's call it R. Since I also fix by that point a bunch of least (and most) significant bits, I can subtract how much they contribute mod 5^n from R, to get the remainder from division in 5^n of the still unknown bit. The thing is, maybe it's not possible to get this specific remainder with the unknown bits, because they're too few.
So, I can prepare in advance a table of size 5^n (for one or more ns) which tells me how many bits from the middle of the palindrome I need, to get a remainder of <index mod 5^n>.
Then when I get to the aforementioned situation, all I need to do is to compare the number in the table to number of unknown bits. If the number in the table is bigger, I can prune the entire subtree.
From a little bit of testing, this seems to work, and it seems to complement my current lookup tables and not prune the same branches. It won't make a huge difference, but every little bit helps.
The important thing, though, is that I'm just happy there are still algorithmic improvements! For a long while I've been only doing engineering improvements such as more efficient tables and porting to CUDA, but since the problem is exponential, real breakthroughs have to come from a better algorithm, and I almost gave up on finding one.
[0] https://ashdnazg.github.io/articles/22/Finding-Really-Big-Pa...
1. Ambitious: ClutchTop - an opensource ai harness (desktop app) similar to claude code desktop app built in electron. It's to the point where I've started to use the app to build the app. Still in v1 though.
https://github.com/veejayts/clutchtop
Building this because I want both chat and agentic interface to use different models via openrouter/local.
2. Miscellaneous PDF/img/doc tools (compression/merging/rotation, more to come) on the browser as a static web page.
Try it here: https://veejayts.github.io/pdftools.html
Built this because I don't want third party tools to have access to my document data, especially for compressing identity docs for government websites.
My goal is to make a simple yet interesting procedural and replayable puzzle. It has a couple of weekly variations: on Saturdays you need to break a rule to score max points, and on Mondays there's an added memory aspect which brings variety to the game.
It's mostly vibe-coded which lets me focus on game design and testing. The next step is better onboarding/tutorial and more intuitive UI.
I'm hosting my personal gallery with it: https://captures.moe
The opposite of the favorite questions: Why did that company I worked for fail? Why did Rome collapse? Why do people get old and die?
Combining information theory with thermodynamics and control theory you get: 1) A set of six pillars that all systems that persist must have. 2) A fundamental 'Action' that all of these systems take. 3) A set of three rules for how system that persists must subdivide
This lets you do things like look at something that is failing and know that there are the 6 pillars and you can then identify them to determine what is failing. (For example there is a system that clears that brain of amyloid plaque and it can fail).
I have applied this to countless systems including Religion, Language, AI Models, Business, the cell, quantum physics, number theory and much more. It is a Rosetta Stone for persistent systems. When there is an unsolved problem in one domain I can map it through this to any other domain that has already solved it.
Note that this doesn't apply to all complex systems, only those that persist.
And to keep this HackerNews related, been applying it to LLM's as they are just a stream of tokens that try to persist to incredible success I might add. Being able to pull from any domain do this brand new field is a giant cheat code.
What I've noticed personally and with founders I talked to is that communication and email triage takes a large amount of time each day, but mostly only needs a quick decision or rerouting the right kind of information. But all this mental overhead takes away time from the really important strategic tasks that you should be working on as a founder.
That's where Chief of Staff comes in. Like a real chief of staff, it gives you a head start into the day: checks your calendar, your incoming emails and messages, prepares meeting briefs and drafts responses.
But it goes further than your typical smart inbox: by sharing your strategic goals it prioritizes and makes sure you are working on the highest leverage items. It also helps you manage relationships by tracking communication frequency and sentiment, so you don't miss when a key customer goes cold.
Both Superhuman and Lindy fell short in key areas of the UX for me: Superhuman makes email triaging faster, but you're still the one doing most of the work. Lindy is highly customizable, but you'll spend a ton of time building and tweaking workflows. I wanted a batteries-included approach to get started right away by just connecting Gmail, Slack, and Calendar without any additional configuration.
A key UX decision for me was also that I stay fully in control. Chief of Staff reviews, analyzes and prepares, but I am the one hitting send.
I'm testing it with a very small crowd right now, but want to open it up soon. If you feel that this is an issue you want solved for yourself as well, feel free to reach out to me and I'll get you on the list for the private beta.
If you have strong opinions for or against this approach, I wouldn't mind hearing this either :-)
You can use the agent from any client (web, Slack, Teams but also other harnesses).
We think most of data analytics work will be transformed (and is already being transformed) from SQL monkeys to chat to analyses, but all the UI/UX are not designed for this and this is what nao is, being open-source because knowing how the context is managed is key.
With nao you can have a conversation and then shared the output on the form of a story, that can be either static or live replacing what old dashboards were.
We are close to 1k stars on Github: https://github.com/getnao/nao
I got frustrated with Claude Code and Cursor producing plausible-but-wrong changes with no easy way to annotate and push back, without making a full PR. crit makes the review stage fun again!
Works on both plans as well as code itself. It’s been very rewarding hearing from folks who use it, everyone has been very kind! My most successful side project already :)
It's an iOS app that applies various generative art effects to your photos, letting you turn your photos into creative animated works of art. It's fully offline, no AI, no subscriptions, no ads, etc.
I'm really proud of it and if you've been in the generative art space for a while you'll instantly recognise many of the techniques I use (circle packing, line walkers, mosaic grid patterns, marching squares, voronoi tessellation, glitch art, string art, perlin flow fields, etc.) pretty much directly inspired by various Coding Train videos.
Direct download link on the App Store is https://apps.apple.com/us/app/photogenesis-photo-art/id67597... if you want to try it out.
* Coming to Android soon too.
Since your app is fully offline I'd love to chat about photogenesis/your general work in this area since there may be a good opportunity for collaboration. I've been working on some image stuff and want to build a local desktop/web application, here are some UI mockups of that I've been playing with (many AI generated though some of the features are functional, I realized that with CSS/SVG masks you can do a ton more than you'd expect): https://i.imgur.com/SFOX4wB.png https://i.imgur.com/sPKRRTx.png but we don't have all the ui/vision expertise we'd need to take them to completion most likely.
I have been brushing up some drawing skills for concept art, and exploring more embedded automotive product ideas for this niche of cars.
Part of building this, I decided to build a BigQuery emulator from scratch and learned a lot about GoogleSQL (previously ZetaSQL) along the way: https://github.com/slokam-ai/localbq
I plan to maintain and improve this going forward. The goal is to see how much can emulators actually do.
Website: https://localgcp.com/
3 days ago, 220 comments: https://news.ycombinator.com/item?id=47700460
5 days ago, 51 comments: https://news.ycombinator.com/item?id=47679021
8 days ago, 21 comments: https://news.ycombinator.com/item?id=47639039
11 days ago, 22 comments: https://news.ycombinator.com/item?id=47600204
It is:
- open source
- accountless(keys are identity)
- using a public git backend making it easily auditable
- easy to self host, meaning you can easily deploy it internally
- multisig, meaning event if GitHub account is breached, malevolent artifacts can be detected
- validating a download transparantly to the user, which only requires the download url, contrary to sigstore
Nearing Alpha release stage.
Code at https://github.com/asfaload/asfaload Info at https://asfaload.com/
I had already developed a tower defence game without AI long time back.
Wanted to try my hand at guided vibe engineering and see how faster was it.
https://github.com/audion-lang/audion
The idea came after I finished a permanent piece for a museum using MaxMsp and python. I always had this thought in the back of my mind that "I could express this so much easier in a few lines of code.."
Check the docs folder for the full language spec.
I really liked how objects came out, I don't think it needs any more since I can do object composition.
There are some nice functions to generate rhythms and melodies with combinatorics, see src/sequences.rs and melodies.rs
Its a WIP but you can use it now to create music with whatever you want: hardware/daws/supercollider , download the nightly release.
supercollider is tightly integrated but not required. I havent had time to develop userland libraries yet but I'm working on it
https://virissimo.info/build-your-own-alu/
LMK what you think.
I'm not sure if I'll every productize it in any way, but I could see a world where it's used by people prepping for the bar, med boards, various continuing education stuff. Right now it's just a fun platform to build on as I explore the current wave of technologies. Building a framework for evaluating different LLMs for best price/accuracy. Adding a RAG pipeline so wrong answers can point back to source material for further review, etc.
I'm looking at moving from backend engineering to a more MLE or agent pipeline role, so this is giving me something more than school projects to build on. While also helping me do better at school.
The main thing I'm currently working on is a platform for organizing and discovering in-person events. Still not certain about the boundaries for "Phase 1", but I have a bunch of ideas in that space that I've been incubating for a while. One subset of features will be roughly similar to that app you've probably heard of that starts with 'M' and ends with 'p', but hopefully an improvement, at least for the right audience. But wait, there's more. :)
Currently building it; it's not public yet, so no link. Next month. I also have an external deadline around that time.
Thinking about how to grow the userbase is intimidating, but I think it might end up being fun.
Each cat mirrors the agent's state, such as sleeping when idle, walking when working, sitting when waiting for input, running toward your cursor when it needs permission.
Fully native Swift, no Electron, under 5 MB, zero network requests, all session data stays local as plain JSON.
I published it source-available with an honor-system license, but this week I’m going to fully open source it and remove the licence. The payment/nag system was an interesting experiment but the project is more useful to me as a proper OSS tool at this point.
8/15 on SWE-bench Verified vs Claude Code's 5/15, ~$0.06/instance vs ~$0.12. Small sample, single repo, lots of caveats. But the direction feels right. Event-sourced reducer, no framework deps beyond the Anthropic SDK.
Well, this is the app that answers that. Everything is seamless, you don't configure anything. You simply start the app on the Mac Mini, import the photo folder, and let it run in the background. In the kitchen, you simply take your iPhone or iPad, open the app and voila, all photos from all libraries show up, organized by date, place, album, people, event. You want to see photos from Prague? You simply check Prague from the filter sidebar. You want those from 2008? You check 2008. Done.
This is not an AI search-based photo library. You cannot even search. Everything you can search for is laid out in the sidebar. You don't need to remember where you have been in 2008. You check 2008 - you see all locations, all albums, everything from that year. You want to see how many trips you had to Vienna? You check Vienna. It's kind of old school this way, but I find it much mentally sane to see a list of filters with things you have done and places you have been and dates you took photos in, rather than an empty search input open to guesses and missed attempts.
This is also not a replacement for your Apple Photos app, or a photo editor app. This is not a photo editor. It's simply a better way to browse historical photos, in a home network, without thinking about it.
My problem: my super-long, months-old ChatGPT threads were breaking down. Even typing got slow, and the longer the threads got, the more they hallucinated. I loved Google AI Studio and paid for it, but I was constantly deleting and re-editing the same thread just to try a different angle. And I couldn't run multiple frontier models against the same context or files without copy & paste and tab switching.
I do 99% of my AI work now in Alyph. One board instead of a dozen tabs, branch anywhere, hot-swap models on the same context. Best guess: I'm 3x faster building things. The honest 1% failure case: it's too slow to load for a quick throwaway question.
Hardest technical problem so far: layout algorithms for an infinite canvas.
Pre-revenue, early users. Building self-service billing now.
Looking for a co-founder who's sold B2B software before.
Some prototypes are already live in my app. Screenshots in the App Store: https://apps.apple.com/app/nonoverse-nonogram-puzzles/id6748... (the patterns in the puzzles in the dark mode screenshots, i.e. 4th and 7th).
I've worked with data my entire career. We need to alt tab so much. What if we put it all on a canvas? Thats what I'm building with Kavla!
Right now working on a CLI that connects a user's local machine to a canvas via websockets. It's open source here: https://github.com/aleda145/kavla-cli
Next steps I want to do more stuff with agents. I have a feeling that the canvas is an awesome interace to see agents working.
Built with tldraw, duckdb and cloudflare
Beyond standard features (retries, caching, timeouts - enabled with attributes on the decorator), Coflux supports more novel features - like suspense (where a task can choose to go to sleep and get restarted when a result it depends on becomes available), memoisation (where steps within a run are aggressively cached so that you can re-run steps in a workflow without re-running upstream steps), and the ability to re-run a step in a different workspaces (with updated code, or in a different environment).
It turns out this works great for implementing agentic systems - you can provide references to tasks as tools to an LLM call and have the AI drive - tasks can be easily sandboxed. And Claude is very capable of using the CLI to interact with the orchestration server to submit workflows, investigate failed runs, make updates to workflows and re-run steps.
I'm trying to make sure it's easy to try out - there's a self-contained CLI that can be used to start the server (a single Docker container), run worker processes, and then interact with the server. The dev mode automatically restarts the workers as you make local changes. There's also a hosted UI for observing runs in real-time, where you can see the execution graph, access logs/metrics/assets/etc - it works without creating an account - the browser interacts with your orchestration server directly.
Platform deterministically generates tasks, creates environments for them, observes AI agents and then scores them (not LLM as a judge).
We just ran a worldwide hackathon (800 engineers across 80 cities). Ended up creating more than 1 million runtimes (each task runs in its own environment) and crashing the platform halfway.
104 tasks from the challenge on building a personal and trustworthy AI agent are open now for everyone.
To get started faster you can use a simple SGR Next Step agent: https://github.com/bitgn/sample-agents
Also this week launching https://dirtforever.net/ which is an open alternative to RaceNet Clubs for Dirt Rally 2, since EA is shutting that down.
I'm also expanding the SDK and plugin space for https://fastcomments.com and am planning on adding AI agents because everyone expects that now :) a big challenge is building it in a way that doesn't make half the users mad. So I'm planning on marking any comments left by AI mods with a "bot" tag, and having the system email users on why it made certain decisions, with an option to contest that loops in a real person. I'm hoping this provides value to site owners while not angering real people. The agents could also just do non-invasive things like notify certain moderators when comments violate community standards in some way, or give out awards. I'm also hoping at some point I can run my own hardware for the LLMs so I don't have to share people's data with third parties.
It allows you to get a wake up call from someone friendly, somewhere out there in the world.
It's got a handful of regular users and it's mostly me making the calls, but it's great fun to wake people up!
No phone number required - these are VoIP calls via the app.
Built it because I think it's cool.
You get to choose the genres you're interested in, and it creates playlists from the music in your library. They get updated every day - think a better, curated by you version of the Daily Mixes. You can add some advanced filters as well, if you really want to customise what music you'll get.
It works best if you follow a good amount of artists. Optionally you can get recommendations from artists that belong to playlists you follow or you've created. If you don't follow much or any artists, then you should enable that in order for the service to be useful, as right now that's the only pools of artists the recommendations are based on.
- ETH Watchtower: a real-time EVM monitoring tool with heuristics and classification of contracts and transactions: https://ethwatchtower.xyz
- P2P SSL VPN provider/consumer tools using a blockchain as announcement and settlement layer: https://github.com/rnts08/blockchain-vpn
- OrdexNetwork: https://ordexnetwork.org, I've built https://ordexswap.online and https://ordexswap.online/wallet/ as well as an Umbrel variant of a self-hosted wallet.
- Waya Wolf Coin v3: Helped the team to compile binaries for linux, and modernizing the libraries: https://github.com/rnts08/WWC3-Linux-binaries / https://github.com/Waya-Wolf/WWC3
- Low Cap Exchange algorithmic trading bots with machine learning and automatic ghost trading, because I wanted to see what the most common shapes are on smaller exchanges: https://github.com/rnts08/low-cap-exchange-trading-bot
However, I am really looking for Sr. DevOps/Platform Eng/SRE/System/Network Admin/Infra Engineering or similar, full-time or contract work, see https://timhbergstrom.pro for contact details.
https://github.com/jank-lang/jank
It's a native Clojure dialect which is also a C++ dialect, including a JIT compiler and nREPL server. I'm currently building out a custom IR so I can do optimization passes at the level of Clojure semantics, since LLVM will not be able to do them at the LLVM IR level.
I would love to know more about Jank, from what I read, it transpiles to C++ right?
Fun project playing around with print in demand and Etsy. Now wondering why Etsy became so popular while being tricky and inflexible to use for the seller :-)
Data support tool for pharmacists to identify savings and best value opportunities for their local health system (NHS/UK)
I'm a pharmacist, worked in the community for 5-6 years before moving into medicines optimisation, which in short is focused on ensuring we use the right medicine at the right time to get the best return on investment in terms of £/patient outcome.
Been a hobby coder for about a decade now, but this is my first attempt at a full stack application (airflow, db, backend, frontend).
Mobile formatting is a little bleh and there's some obvious issues. But it's been rather nice setting up something a bit more rigid/resilient than my previous clandestine approaches
Think of it like "Claude code on Supabase", but for internal apps and AI agents.
I got tired of choosing the deployment platform, wiring up Postgres, SSO (OIDC), RBAC, audit logs, secret vaults, integrations/tools/MCP, ... from scratch every time I needed an internal tool.
I just published a fun interactive 3D demo of SPDC, one of the most common and accessible ways to create entangled pairs of photons. I'm hoping to publish a series of articles on other cool learnings about doing quantum photonics in the lab.
Tracks usually post updates on Facebook, so riders end up checking dozens of pages manually. I scrape recent posts and use an LLM to infer whether a track is open, closed, or unknown for the upcoming weekend.
Currently Android-only, with iOS in progress:
https://play.google.com/store/apps/details?id=com.lynxleap.t...
Backend is zig, frontend is in Flutter. First foray into zig and I'm really enjoying it.
- https://apps.apple.com/us/app/mame-sama-%E3%81%BE%E3%82%81%E...
- https://play.google.com/store/apps/details?id=com.mamesama&h...
I use it when I have candidate libraries to solve a problem, or I just want to find out how things work. Most recently I pointed it at fzf and was able to pull the insensitive SIMD matching it uses and speed my own projects up.
I can’t find it right now, but there was a post about how ripgrep worked from a someone who walked through the code finding interesting patterns and doing a write up on it. With this I get it over any codebase I find interesting, or can even compare them.
Command replaces that with a platform that maps to their real sites, real assets, and real operational constraints, so they can actually run the program, not just document it.
Consulting firms use it to deliver more engagements with the same team. Asset owners use it to keep the program alive between engagements, or run one themselves.
The core frustration: Apple Watch collects HRV, sleep stages, respiratory rate, blood oxygen, resting HR. Apple does basically nothing useful with any of it. You get ring animations and step counts.
Atlas pulls all of that together and turns it into two scores: recovery and training readiness. The point is to actually use the signal your sensors are already collecting and ensure when you train, it matters. It’s like Whoop, but actually works.
iOS app is live (finally!). Happy to talk shop.
For those DMs that use tools like these, my app sits between Shieldmaiden and Improved Initiative in terms of features/complexity. I tried to offer as many features as possible but "hide" them in a way that makes it easy to understand the most important information like initiative order, health and conditions, stat-blocks. But then I added many buttons with keyboard shortcuts and a quick-access command-palette (think MacOS Spotlight or Alfred on Linux) that lets you access even more commands and features just by typing.
It is in beta, free and and you can check it out at https://topoftheround.com
I'm also working on a 2d procedural animation plugin for bevy, a autotiling plugin for bevy (using 16 tile-dual grid, which the default bevy autotiling plug-in didn't support) and ofc my android pixel editor now has a rig editor mode and a tile editor mode that integrates with the plugins.
Making video games is hard! I keep getting side tracked!
On that first link you can find a lot of answers to frequently asked questions.
[1]: https://news.ycombinator.com/item?id=47700880
[2]: https://uruky.com
free, open source -> https://github.com/smol-machines/smolvm
I worked with firecracker a lot back in the day and realized it was a pain to use. And containers had a lot of gotchas too.
Since sandboxing is all the rage now - I think it'd be a better infra primitive than firecracker that works locally/remote and etc.
I’m working on https://coasts.dev.
I’ve been thinking a lot about the light vm side lately but it’s not an area we are going to attack ourselves. I think there’s a really good pairing between what we’re working on.
I think anyone looking to use infra that needs below properties are well served by this project: 1. subsecond vm cold starts 2. kernel isolation (vs containers) 3. consistent local <-> remote environment 4. elastic cpu, memory. 5. ease to setup.
I am designing it as a infra primitive on purpose for general workloads as opposed to others in the microvm space i.e. firecracker was designed for lambda/serverless workloads.
I adapted my open source ruby on rails real estate website builder to work with EmDash and can already see a lot of potential.
It's not ready for production use yet but I'm really enjoying working on it:
https://github.com/RealEstateWebTools/emdash_property_web_bu...
- Weirdly, the kitchen sink is almost exactly the geometric center of the house; hence, equal probability for odors to travel.
- And that reminds me: Need to download PDF for dishwasher operation.
- Day 2 (Friday) of my wonderful better half's travels, I started laundry. I remembered less then 2 days later that I need to transfer the clean (??) clothes from the bottom device (water/soap) to the upper "dryer" -- this device produces some serious heat. Kills odor causing bacteria, and stuff. Will call that a success.
- I find my clothes are scattered on the floor randomly. Seriously high entropy -- reminds me of CloudFlare's lava lamp application: https://en.wikipedia.org/wiki/Lavarand
- Yep, total regression to the mean of bachelor-self and loving life..and the miracles of modern technology, where like the water automatically fills in the washing device. But not the soap.
Native cross platform app coded in rust + tauri.
I prefer using it to the other agentic code apps I have used. It has multi tab worktree isolated agents, sandboxed tools, git integration, built in code editor (with inline generation), searchable document support (i.e. upload your docs, datasheet and you or the agent can use them) and even built in local image generation (using stable diffusion and flux schnell) and asset handling for game developers. Oh it also has a remote feature so you can share the gui or deploy it on a server and access it on the go.
Working on adding text to 3d also.
It is a hobby project that has grown quite large. Feel free to try it out.
It is also possible to download and try yourself without paying (there is a free trial period).
It comes with time stretch and pitch shift as most of these softwares do, but it allows you to save loop regions and take notes. It's designed to be a practice session tool.
I'm doing it from first principles, and having fun writing GPU code, platform shims, and squeeze every ms I can to make it fast and smooth.
I will be looking for testers soon. If anybody is interested, hit me up.
1. Better GitHub insights at https://temporiohq.com (public and very early stage). Demo of what the product can do here: https://temporiohq.com/open-source/github/symfony/symfony
2. My art. Mostly at https://instagram.com/marc.in.space or at https://harmonique.one/works
Early preview here: https://piggy-toss.netlify.app/
The goal is to play with friends, we love this game.
Mostly for myself to use for my hobby. Sharing with everyone because I find it genuinely useful.
Yes, it is coded with assistance of LLMs, but I care for the details and it is not vibe-coded in hours.
https://hobbyboard.aravindh.net/
GitHub: https://github.com/aravindhsampath/hobbyboard
Demo (resets every hour): https://demo.hobbyboard.aravindh.net/
Almost ready to do a show HN :-)
Also put together a directory of 31k+ personal websites, tagged with design keywords so they're searchable. As someone who loves personal sites, I think it's one of the more comprehensive list of indie / personal sites on the web:
It’s a command-line tool for decoding certificates and CSRs into structured JSON rather than OpenSSL-style text output.
It decodes the underlying ASN.1/DER structure so fields and extensions are fully expanded, making the output easier to work with programmatically.
I’m planning to expand it to support more PKI artefacts (e.g. CRLs, Keys) over time.
I’m also planning to handle less well-formed inputs (e.g. missing PEM headers/footers, whitespace, or extra surrounding text), which tends to come up in real-world data.
It’s free to download — would be great to get feedback if anyone tries it.
- https://github.com/rumca-js/Internet-Places-Database - Internet meta database
- https://github.com/rumca-js/Internet-feeds - list of Internet feeds
- https://github.com/rumca-js/crawler-buddy - crawling framework
- https://github.com/rumca-js/RSS-Link-Database-2026 - link meta from year 2026
- https://github.com/rumca-js/RSS-Link-Database-2025 - link meta from year 2025
I initially was using SSE to push events down to the front end during long scans but decided to switch over to plain old HTTP polling for better reliability across different browsers (and versions of different browsers).
Here are the areas of analysis:
- accessibility
-- check for images with missing alt text
-- check for various form controls missing labels
-- headings not following (h1->h2->h3...)
-- missing lang attribute on <html>
- content
-- check for forbidden words and phrases
-- check for required words and phrases
- performance
-- evaluate time to load page
-- check for excessive inline JS
-- check for inline styles
- security
-- check for SSL certificate expiring soon
-- check for security HTTP headers
-- check whether Server HTTP header is too revealing
- seo
-- check for missing title in head section
-- check for missing meta description
-- check for multiple H1 headings
- site integrity
-- check for broken links
-- check for use of deprecated tags
-- check for insecure http link
- spell check
-- check for possibly misspelled words
Having a lot of fun building it!Going for a 100% self-service model. No corporate sales cycles, no slide decks, no meetings.
Targeting a June launch.
You can provide the DM a premise (or pick one from the library) and it'll flesh out a full campaign story arc. Either way it's a fresh story arc reacting to your actual decisions, every time.
I noticed every competitor in this space was a chatbot with only the last ~10-15 messages stuffed into context. They forgot things, made up dice rolls and rules, and was generally not what I was looking for. So far TableForge has been working well for my friend groups and some random folks from Reddit/organic search. Solo TTRPGers seem to like it too.
It's still in early stages but fully playable. I don't feel comfortable charging anything for yet until I know people enjoy it. If you like it enough to hit the free tier limit, send me some feedback in the webapp and I'll gladly extend your free trial. If you hate it, please also let me know!
I got back into MTG back during the pandemic after a long hiatus and Spelltable is what brought me back. My playgroup lamented more features and something tailored to our needs, so curiosity got the better of me and here we are. :)
I've never worked with computer vision before, but I went through a whole journey that started with the classical computer vision techniques and ended with recently migrating to the transformer-based models. Been a really cool adventure!
My playgroup has been loving it so far, and I would love for people to try it and tell me what breaks! Discord is on the site.
It consists of CRM, Expense tracking, Equipment Management, Event Gallery( photo share, Face Detection based download, Guest Upload) etc..,
Currently working on moving it from cloud supabase to self hosted version.
https://feedbun.com - a browser extension that decodes food labels and recipes on any website for healthy eating, with science-backed research summaries and recommendations.
https://rizz.farm - a lead gen tool for Reddit that focuses on helping instead of selling, to build long-lasting organic traffic.
https://persumi.com - a blogging platform that turns articles into audio, and to showcase your different interests or "personas".
We also write about like:
How fund performance explain part of returns, rest is explained by timing. And ways to tease those out: https://finbodhi.com/docs/blog/benchmark-scenarios
Or, understanding double entry account: https://finbodhi.com/docs/understanding-double-entry
It is scientifically proven[1] that sitting is detrimental to our health, with increased mortality rates. The primary way to reduce the negative effects of sedentary work is to move.
This means doing sessions of resistance training (gym), running, biking, but also taking micro-breaks during work sessions and performing light exercises and stretches.
Research has shown[2] that taking short breaks during work reduces fatigue, and in some cases actually boosts performance.
There are plenty of running and gym apps out there, so Limberly focuses on the last part - helping you take micro-breaks, reminding you to change your posture between sitting and standing, changing which hand holds the mouse (if you're into that) etc.
It is still in early development, so if you'd like to help test and shape the app as we go, please sign up for the waitlist and I'll add you to the testers group. Feel free to also DM me here with any questions or feedback.
Oh, I am also writing a series of articles that explains it more in depth: https://prodzen.dev/articles/building-limberly-part-1-we-re-...
1: https://pmc.ncbi.nlm.nih.gov/articles/PMC10799265/
2: https://journals.plos.org/plosone/article?id=10.1371/journal...
It's a newsfeed constructed from 130k substack RSS feeds but limited to the past 24h.
Its helping me discover writers other than just what the algorithm gives me.
PSA PS. Don't post generated/AI-edited comments. HN is for conversation between humans https://news.ycomtem?id=47340079
I call it Hammock, in honor of Rich Hickey's "Hammock Driven Design" https://github.com/tlehman/hammock
I couldn't find any crate that would be ergonomic enough to use and provide features I deem essential, i.e. retryability, scheduling, poison job detection, barriers, backoff strategies etc.
it's an area I'm familiar with so after spending 2 days trying to integrate external libs I decided to roll my own and I'm quite happy how it turned out in 2 days of development.
I plan to open-source it in the near future but right now using it in my another project and it's running quite well.
It’s also a lot of fun to work on. Phoenix LiveView dashboard, go probes running on 4 continents, connected to the backend using websocket tunnels. Clickhouse for reporting. Even did a CLI and an MCP for fun.
You can take the probes for a spin with the free response time checking tool and see how fast your site is https://larm.dev/tools/response-time
I wanted to make it easier to quickly see/study trending articles on Wikipedia because they tend to make good topics to know before going to trivia night.
I've had the domain for awhile, but just made the app today on a whim.
I use Wikimedia's api to get the trending articles, curate them a bit, add some annotations to provide some context, then push to deploy the static site.
It's built using biscuits and written in rust. I'm really into it. Using capability security as a model makes building things feel like they snap together a lot more naturally. At least for me.
I've also got a blog post describing it in more detail: https://www.hessra.net/blog/what-problem-led-me-to-capabilit...
Current state of work: The implementation of the core data model is wrong. I need to throw it away and redo it from scratch.
Whiplash status: WTF, Time. y u move so fast?
This thread made me---forced me---to accept that it's been well over a year of the agony and ecstasy of solo software construction. Or maybe 2026 is moving way too freaking fast. Or it's good to be obsessive I guess.
We just crossed 5,000 commits. Also, we take testing very seriously: our test code base is presently 160% the size of our production code.
[0] jacobin.org
It's web service that allows you to channel your google docs through a more human-friendly name. So, you link
opendocs.to/your-name/resume (an example link)
to your public resume at docs.google.com/dlkjbalksdfd
It's a simple redirect service, but it just looks nicer, and I think the opendocs.to sounds natural. Got to learn a lot with this one, using Vite/React, Node, Postgres all in Docker, with a local profile that builds nginx inside with the containers, or a prod profile on the server where nginx proxies into the containers.
Anyways, check it out!
Right now, only free tier available as I some last tweaking and checking.
Basically a google streetview tour of your Datacenter or large industrial plant.
You can do some nice things like draw 3D linework to trace the paths of pipes, conduits, eg : https://youtu.be/t8nRhWUl-vA add notes with markdown and html links at useful places in the 3D space.
We have add-ons for generating an 'xray' view floorplan to make it nicer to navigate a large space.
I think we are the first to have a web uploader that can preview and import .e57 panoramas, directly in the web page [ and skip the points if you dont need them ]
Currently in use by a telco in the Americas.
If you're enjoying it, please leave me some feedback: https://discord.gg/pFjEcbQsv
Still iterating through refinement and features. It's built on Rust + Tauri with a React frontend, in case anyone is curious.
I've created various open-source and commercial tools in the multimedia space over the last 10+ years and wanted to put it all together into something more premium.
Building up the marketing now. Starting to get some coverage on Instagram: https://www.instagram.com/p/DWxWo_oDfkm/
Also recently built a home energy cost/consumption display for the TRMNL - https://andrewhathaway.net/blog/ambient-cost-display-for-oct...
Making cabinets is not that hard but the industry charges insane amounts of money for it. Since I have to make cabinets for two kitchens I invested in a Sienci Labs CNC so when the cabinets are done I'll have saved money and gotten a CNC out of it which I can then sell or use for other things.
Both peers mount a virtual FUSE folder. Files shared by one side appear in the other's folder in real time. You can open, copy, and browse your peer's files as if they were local. Files go directly between devices over encrypted gRPC. (by default it tries over LAN, then direct IPV6, then uses a data relay).
The hardest part has been making git repos work through the FUSE mount between peers.
(Been developing the tool for 12 months now, very close to a full release)
It runs code locally (written in Swift), includes examples, and has a Turbo Pascal–style theme for max nostalgia.
PS : On your, slick landing page, please make an email imput so we can can know when to return. By the way my phone is ... android.
Testing out some ideas to automate data entry workflows from an italian powerlifting federation (FIPL) to OPL https://www.openpowerlifting.org/
Modernizing in two ways: migrating to new JS tooling (webpack -> vite, Node’s built in sqlite, etc) and adopting ircv3 features like emoji reactions, threaded replies, and typing indicators. Trying to bring IRC into the 21st century.
Its easy to contribute to and we have an active irc channel (perks of building an always-on client…) - feel free to join us! #thelounge on irc.libera.chat
Check out the bundle / CPU savings by leaving webpack: https://github.com/thelounge/thelounge/pull/5064
Got fed up with web tech, it's so slow and clunky, so made my own version in python and qt. I changed the design to be based on a doclayout llm, so you can skip or include things like tables and references easily.
It now works so beautifully fast, it's code is readable and simple, no apis or multiple services. Just a qt app, some local llms that can run on a decent cpu and word-leven highlighting and playback selection.
https://github.com/thepycoder/projectwhy-tts
I can listen to papers now!
It was a lot of fun and I love all the good energy people bring to the conversation about long lasting and community driven tech.
I’ve been building a phone app + website (https://MyBulkCards.com) to scan cards and organize where everything is. It’s pretty basic right now, but I can store cards in boxes like “Box 1 AAA, Box 1 BBB, …” and find cards easy peasy. There’s also a friends feature so I can see what others have locally. We borrow cards from each other quite a bit.
It’s been a fun project to build. I trained one model to find a card in the camera frame and another to identify it. Still iterating a lot. One epoch on my Mac M4 takes about 2 hours, and I’m still seeing improvements past epoch 10. Even now, it can find and identify a card more often than not, even without the OCR bits. Both models are under 20MB, run directly in the camera frame, and are fast enough to identify a card as I slide it into view.
I started with Android since that’s what I have, and I’ve shared the app store testing link with my local group for testing. The app is built in React Native, and I’m hoping to get an iPhone version out soon since there are a bunch of iPhone peeps. A couple of the players also got me into MTG, so now I’ve got a pile of Turtles cards too. I’ll be training an MTG model next. I don’t think it’ll be too bad since I can reuse most of the same approach.
Some of the biggest pain points we’ve seen is chat being separate from a solid task manager, and the pain of collaborating with people outside your own org.
We’re currently in private beta and hope to open it up to the general public soon!
I massively improve it every month. Pretty proud of it.
We have over 60 shows now, rented a studio, and are in talks to security a site for our tower. I'm building out an online store but really need to focus on fundraising.
The Registry in turn has two interfaces: one REST, and one A2A itself. If you hit /.well-known/agent-card.json on the Registry server, you get the AgentListerAgent, which supports searching for Agents by various criteria. Or you can search using the REST interface. In either case, you get an AgentCard that points to the correct APISIX endpoint to talk to the desired Agent.
Besides adding K8S support, other plans include adding support for other proxy providers (including Istio for the K8S case), supporting Agents that are not based on A2A and, allowing Agents to register themselves using the Registry API, and... uh, well, that's the main stuff I have in mind right now. Aaah, wait, I might do something along the lines of integrating an MCP Registry as well, not sure yet. Heck maybe I'll get bored and make it an all-out API registry for all sorts of endpoints... could integrate a UDDI server and bake in WSDL support for good measure! (Don't count on that last bit happening anytime soon).
Anyway, no repo to share right this second, but I do intend to make it open source. I'm just committing the cardinal sin right now of wanting to "make it presentable before releasing the code".
I’m working on OurCodeLab, a Singapore-based startup. After 11+ years in DevSecOps, I noticed a lot of local SMEs are either overpaying for simple sites or using insecure, bloated templates.
I’m trying to solve this by building high-quality, lightweight landing pages at the most affordable rate possible. Right now, I’m running a promotion: we’ll build your landing page (up to 2 pages) for free if we handle your domain hosting.
I craft each site individually to ensure they meet modern web and cyber standards—no copy-paste layouts. I’d love to hear your thoughts on the model or any feedback on the tech stack.
If you're an SME or know one that needs a hand, reach out at farath@ourcodelab.com for a non-obligatory chat.
github repo if you wanna check : https://github.com/devlensio/devlensOSS website : https://devlens.io
Like PocketBase, it's made in Go, has an admin panel, and compiles down to one executable. Here, you write your endpoints as Lua scripts with a simple API for interfacing with requests and the built-in SQLite database. It's minimal and sticks close to being a bare wrapper around the underlying tech (HTTP, SQL, simple file routing), but comes with some niceties too, like automatic backups, a staging server, and a code editor inside the admin panel for quick changes.
It comes from wanting a server that pairs well with htmx (and the backend-first approach in general) that's comfy to use like a CMS. It's not exactly a groundbreaking project, and it still has a ways to go, but I think it's shaping up pretty nicely :)
So built Sustalium (https://sustalium.com) which is designed to be easier and faster for micro-small-medium businesses to comply with majority of compliance & sustainability frameworks.
To be honest I built it just for me and then decided it might be useful for others.
It's all local, no server, no database, etc. Mobile and desktop friendly.
I used to think "if you build it they will come" but, as it turns out, it's much more nuanced than that and requires a lot of iterating and stumbling along the way. I hope to break into another vertical this year!
Trained to detect a few thousands species pretty accurately in near real time.
Now working on expanding to far more species and exploring other CNN architectures.
It lets you create TV channels from digital media such as YouTube, The Internet Archive, TikTok, Twitch, and Dailymotion. It does that by letting you schedule videos against a custom calendar system.
Since filling out even a month of content can be a lot of work, I built some things to make the process easier.
* Advanced scheduler to know when and how long content can be played at any given datetime
* Real time team collaboration
* Channel libraries to organize media
* "Blocks" - Create a dynamic schedule which generate hours of content that mimics real television scheduling. It even carries over your playback history between generations so that playlists continue from where they left off.
* A catalog to find media from official sources on YouTube
* Embeddable as an OBS browser source to restream your owned content
* Repeat content infinitely or temporarily to create 24/7 channels.
If all goes well I am hoping to re-release sometime this month.
Paste a link → AI breaks it into sections → teaches you on a whiteboard with voice → quiz + flashcards at the end.
It's free to try while in beta: https://www.pandio.online
When I started doing this, I also decided to try Proxmox's new OCI compatibility, which seems to be working well so far, so I am removing all my Docker VMs and recreating the containers directly on my hypervisor.
My friends and I have been having a great time playing the initial version, and it's been fun working on some of the more interesting technical aspects like server + browser performance, mapping 2-d game space onto a 3-D visual space, etc. as well as some just-because-I-want to things like a dynamic music system.
Unlike those apps it has full support for design tokens and (so far) flexbox layouts. It can also export directly to HTML, rather than a fake preview mode. I’m also working on full code-backed components, so you can go between code and design very easily.
As a designer, I’ve been frustrated for years by the gap between design and code, and despite all the new AI features, Figma still hasn’t got any further in years - design tokens need a 3rd party plugin and responsive designs are a pain in the bum. So I decided to build something that has the ease of Figma while being much closer to live code.
I’ve got to the point where I’m designing the app in itself, tokens are working, html export is working and nearly ready for first betas.
Seems to solve most of my issues with my current workflow. My primary personal development machine is my WSL ubuntu install on my windows gaming PC and the tooling outside of the mac ecosystem has been really limited.
github.com/redshadow912/ReceiptBot
Glyphcraft - a Minecraft mod (imagine if Thaumcraft, Ars Nouveau, and Hex Casting were smashed together)
Recently it hit v3 spec conformance. (I'm executing the upstream spec test suite.)
I don't plan to make it a highly-performant decoder for use in production environments, but rather one that can be used for educational purposes, easy to read and/or debugging issues with modules. That's why I decided not to offer a streaming API, and why I'll be focusing on things like good errors, good code docs etc.
P.S. I'm new to the language so any feedback is more than welcome.
For said same association, templating and assembling a book of songs and other oddities in Typst for the associations 50th anniversary.
Next project is figuring out what to do with my personal website!
No traffic ever leaves your local network and since it uses rsync under the hood the devices being sync'd to don't need to run anything other than SSH.
It's a single file shell script that has no dependencies except rsync. It's literally 1,000+ lines of defensive checks and validations to make sure you're not shooting yourself in the foot with rsync, and at the end the last line of code directly calls rsync. It doesn't try to reinvent the wheel by replacing rsync (it's an amazing tool).
This tool doesn't enforce how you use rsync, it offers suggestions. You can use rsync's flags that help with versioning by modifying 1 config value to add them and now you have versioned backups using all of the strategies rsync supports.
It's also a nice excuse to build in quality of life features that don't take a lot of time because you're using the thing all the time. My favorite one is the color coded rsync command output when DEBUG=1 is set so you can be absolutely sure your config values are producing the expected rsync flags and args.
And building Fractiz.com, a customizable pre-coded backtests platform.
I wanted a surf forecast app that i can look at glance, which "time-slot" of the week is good enough to go surf.
And I wanted it to look like nothing else out there, at least surf forecast wise
The part I cared about was being able to send links via one click in my browser or two taps on my phone as I want to read every HN article who's title I find interesting, but don't have the time to read right at that moment.
It then at the moment publishes it to an RSS feed so I subscribe to it in Podcast Addict, but I've also just been using the web app as my reading list and tracker.
Been playing around with different settings on the piper models and different techniques for getting the most out of my four dollar instance:
https://experiments.n0tls.com/
Up next is to work on making the voice better (I'm impressed with the out of the box stuff already), and then making it better at finding the real content on a page and only recording that. It's a problem space I don't know much about, but find fascinating, been fun so far.
So I built my own package manager that's almost ready for alpha.
- Tool to auto create dashboards from csv/json files or rest api ( https://EasyAnalytica.com )
- Tool to preview, annotate and share html snippets ( https://easyanalytica.com/tools/html-playground/ )
- Creating agent to edit files with smaller model(<1B) not yet released.
-Prompt assember to create prompts/templates/context and easily share between ai to be released this week.
https://ewams.net/?date=2026/03/29&view=Qwen35_Performance_w...
Resuming work. I used to `j <reponame>` then `gco <branchname>`. Now if I do that I get an error about the branch being checked out already in another worktree. I realized the branch names are pretty unique across repos so I made ` jbr <branchname>` that works from anywhere.
Jumping within repo. The other kink was when I wanted to focus on a particular package I’d do `j <subdir>` and it would usually be unique enough to jump to the one in my current checkout. But now I have dozens of concurrent checkouts and have to pick, even though I’m already in the repo. So `jd <subdir>` does like autojump or zoxide but only within the current checkout.
To power those shell functions I made a “where” extension for Git.
https://github.com/turadg/git-where
It’s working out nicely!
Random observations from my first one: - presenting my idea visually helped crystallize my thinking in a way that writing doesn’t. And writing was already very good at crystallizing my thinking. - even making a bad video was a lot of work - making a video presentable is a deep subject. Subtle changes were throwing off my setup. Now I understand why so many influencers are fitness and lifestyle; the demand side is obvious, but when you’re already camera-ready you have a huge advantage on the supply side - described something I built felt natural. I do that for a living. The intro was like 45 seconds and took me like 45 minutes to film because it was acting and I don’t know how to do that - learning about video editing features had an immediate payoff because video is so long
[0] I’m posting the videos at https://m.youtube.com/@bitlog-dev . I said if the first one got to 100 I’d commit to making at least 10, and I just crossed that threshold
I’ve got a decent amount of people on the newsletter so trying to figure out how to best deliver indie games via that channel and in the end get more people playing these awesome games people develop :)
With this stack, I'm scaffolding several (fingers crossed) commercial learning SaaS products. The first [2] is LettersPractice - a minimalist early literacy app that's family-first, in so far as it presumes an adult supervisor who co-learns strong confidence as a phonetic coach both at and away from the app. Putting considered rails on the parent-child reading experience.
The second set of apps is in music, with some experimental dev right now against piano (via midi devices), flute [3], aural skills, and sightsinging.
[1] https://github.com/patched-network/vue-skuilder , https://patched.network/skuilder
app store: https://apps.apple.com/tw/app/kernel-%E8%83%8C%E5%96%AE%E5%A...
viral launch post that brought in ~1700 users in 2 days: https://www.threads.com/@sean_hsu_13/post/DW8nBzDjV8T?xmt=AQ...
I am managing a discord community with over 1k+ members I found some people would regularly put spam links or message on all the channels and this been repetitive it's just take time deleting them one by one or reposting them into the specific channel. So I build a discord bot that would make this lot easier it catches the spam message post them into actual channel and also delete spam links. It's open source and easy to setup.
Coming from a place where buying games is very expensive, and gaming is an expensive hobby in general.
Tried rotating games locally between friends and friends of friends, now scaling it up.
Post-event feedback showed everyone loved it. But personally I think we could have done better organizing on the co-working side so people has a more predictable schedule to lock-in.
So I’m planning what the next iteration of this event could look like if the co-working aspect was stronger. Especially in the area of everyone sharing their personal and/or professional intentions with each other. So they're more likely to accomplish those intentions with the help of other participants.
Pronounce A-Library "The Unicode character for the Cherokee letter 'A' (Ꭰ) is U+13A0"
Launching a kick-starter for it in the coming weeks. Hoping to make a difference for the next few generations for a better world and education.
and
- There's a desktop app tracking the title bar and time you spend in each app. - You can use this 100% free, or sync this back to https://heygopher.ai to match the time up with your active projects. - if you use HeyGopher you can manage your time, team, projects, quotes and invoices.
This pairs pretty well with my normal project https://goodsign.io which is a Docusign alternative that is pay as you go. No subscription.
Allows you to compile most C or Rust programs to run in it without modification. Also can run Claude Code, Codex, Pi, and OpenCode unmodified.
Working on polishing, security, and documentation so I can share an in-depth deep dive on HN.
So, I've built a scraper that scrapes posts from Facebook Groups and made those posts filterable/sortable.
Now I'm looking to launch the same thing for US cities. Their Facebook Groups have tons of posts around subleasing/looking for accommodations.
If you are interested, here's the site for Bangkok: https://bangkokprop.com
Once a patch for a security vulnerability is public, the patch itself can reveal the vulnerability before the CVE is published. VCamper uses a staged LLM pipeline to analyze a Git commit range and flag likely vulnerability patches, even when they look like routine changes.
It’s still a proof of concept, but on known cases like curl CVE-2025-0725 it got close to the published root cause from the patch alone.
This matters because LLMs could make it much harder to keep security fixes quiet: once the patch is public, the bug may be recoverable almost immediately. Quietly shipping a fix and hoping it stays under the radar may stop being a reliable strategy.
I'm hoping to continue extending it until it can act as a full internet TV delivery stack like Pluto or Roku TV. It still needs to be behind a CDN for efficient delivery but basically any CDN would work.
The app is built in React Native (almost entirely with AI although I'm fairly particular about some of the features and methods it uses) with a Go backend. Map data comes from PMTiles.
I'm now having immense fun trying to come up with anagrams to whole sentences in Turkish.
I guess you could even automate finding anagrams (there are even web sites which allow you to do so), but Turkish agglutination makes it so much fun, and you can make really creative ones manually.
Once upon a time I even had made a tumblr to share what I found: https://sacmanagram.tumblr.com/ (also Turkish).
Recently shipped this personal art project that turns daily Wordle attempts into gritty / struggle-filled stories, kinda similar to the emotional arc of the Wordle game play.
You can upload your own Wordle game screenshot to generate one for yourself.
In addition to completing what was once in the idea list, I got to learn about
- Prompt fine-tuning: Models are sharp enough to complete Wordle games quicker than human average scores, so I had to dump that down and get the average down.
- Karpathy’s Autoresearch: Experimented with auto-research for prompt fine-tuning, in addition to manual prompts.
- Vision models: While leading labs have multimodal models with quality visual reasoning, the benchmarks are still quite different for a simple Wordle analysis (reading what letters were yellow/gray/green); I also noticed labs/companies with separate vision models but their APIs lag significantly compared to what’s possible in developer experience.
- Video generation: For the last few days, I have been experimenting with automated video generation for the project's social handles. I'm still struggling with the right hooks that reduce the skip rates, but it's fun.
---
Additionally, working on an Apple Watch app similar to my Mac app on the same lines, [Plug That In](https://plugthat.in), i.e., notify before the device goes too low on battery, but with a twist.
Check it out at https://bartyai.com
For the past 4 years I've been building a programming language reimagined specifically for games. It has automatic multiplayer, but also things like state, components, concurrent behaviours and reactive user interfaces baked into the language.
The scope creeped to book discovery and ebook reading with OpenLibrary from just tracking and personal library recommendations.
But we have been able to incorporate new books into the story time rotation so I’m convinced it’s worth it.
It’s definitely been fun experiencing the range of quality for kids books in the internet archive.
I’m aiming for a May 1.0 release on iOS and Android.
Dr. PD is an open-source USB-C Power Delivery analyzer and programmable sink. It can sit inline between a USB-PD source and sink to show you the communication between them, or connect directly to a source and emulate a sink so you can characterize chargers and power supplies.
The goal of the project is to make serious USB-PD analysis more accessible. The hardware, firmware, and host software are all open source. The control software runs locally in Chrome or Edge with no drivers or installation required, and the platform also provides Python, JavaScript, SCPI, and USBTMC interfaces for automation.
(Sorry that I don't have a link to the GH repo yet, but you can follow the project on https://hackaday.io/project/205495-dr-pd. Also, if you read this far, I'm looking for a few beta testers. Reach out if you're interested!)
https://dhuan.github.io/mock/latest/examples.html
^Command line utility that lets you build APIs with just one command.
^JSON/YAML manipulation with AWK style approach.
Similar apps have existed before (like Amie), but they were nearly all VC-backed and had pretty much all pivoted to AI (e.g. being an AI note taker). Their approaches to a Todo-focused calendar has been largely unsatisfying due to the focus on Enterprise users and whatever is trendy.
Eima, in contrast, focuses on personal use and does one thing very well: scheduling your todos. In particular, I spent a lot of time making sure multi-occurrence todos work smoothly (e.g. todos that need multiple attempts or simply recurring todos). These were not addressed by prior tools at all and had been my biggest motivation to build Eima.
Would love some test users! If you end up wanting to give Eima a try please use the code EARLYEIMA to get it for free.
It's been a great excuse to get back to my roots as an engineer and lean into some of the newnew with Claude Code. Learning a ton, having a blast, and also enabling being (marginally) more productive with my actual work day to day.
[1] https://apps.apple.com/us/app/fitbee-calorie-macro-counter/i...
Many people know that a handy data analysis feature in Excel is to create a pivot table from a spreadsheet. But spreadsheets are limited to just a million rows. You can get around this limit by jumping through a bunch of hoops.
My system lets you easily create tables with thousands of columns and hundreds of millions of rows. (Just drop a CSV, Json, or other file on a window to create a table.)
Now you can create a pivot table from it with just a few clicks of the mouse. It is fast (I created a pivot table against an 8.5 million row table of Chicago crime data in less than a second.)
The resulting pivot table is interactive. Each cell (row/column intersection) has all the row keys mapped to it. Double-click on any cell and it will instantly show you all the rows in the original table that were used to calculate the cell. You can then analyze those rows further.
It also works well against much larger tables. I have tested it out against 25M, 50M, 100M, and 200M+ row tables.
Not trying to discourage you, I am curious as to see how you are planning to enter the market as that was something I couldn’t answer when considering working on spreadsheet tools of various kinds or even an excel alternative.
But if your dataset has millions of rows and you need something quick to help you slice and dice the data in a variety of ways to try and find valuable insights in it to drive business decisions; then maybe you are looking for something better.
BTW: creating pivot tables is just one of dozens of things my system can do. I am currently trying to figure out which features will attract the most customers.
I’m adding:
- A control hub that reads data from the batteries and the solar controller
- Remote and on-device UIs that allow a user to control all the hardware from one place
- A LoRa transceiver that allows monitoring the battery and solar status from a distance
Exploring all of this is fun — there’s a lot of DIY solar and battery hardware out there that needs to be able to sync and coordinate, but there’s not a great software solution for this.
Hit me up if you want to hire me, or give me money to work on this :)
Butt ugly unless you're deeply into the tradie steel frame equivalent of the Concrete Brutalism aesthetic.
Next I am making the version for folks who do not make a list and just go with past orders , for them I am automating so the cart is made based on past orders like milk usually is ordered every 2 weeks.
I will be continuing work on the new software that powers it, the Amsterdam Web Communities System. https://github.com/amysoxcolo/amsterdam
(I tried to "launch" it with a Show HN post, but it sank without a trace. I may try again, after I get back from vacation...)
https://get-taxus-org.pages.dev
It's inspired by Zola, but has better documentation and will hopefully be more approachable when all is said and done. I'm trying to incorporate WebAssembly, with Yew, to give "islands" for high performance stuff you might want where WebAssembly makes sense. For example, I wrote search from the ground up, and built a search widget using Yew.
You can also just write JavaScript if you want.
It's a total work in progress, but I'm enjoying what I've built so far.
- An internal apps platform built with bun, pg-boss, and railway
- A smart music setlist manager that downloads chord charts, creates spotify playlists, and automatically drafts emails with attachments and practice schedules
- A recruiting intelligence platform called Spotter that I built in a weekend[0]
- A voice-agent for a client in the banking sector, implementing deterministic workflows using openai realtime voice + finite state machines[1]
[0] https://www.youtube.com/watch?v=AOedMSddGDg
[1] https://blog.davemo.com/posts/2026-02-14-deterministic-core-...
This sounds useful!
Lots of effort has gone into testing against real world docs. Its beta quality right now.
I'm not a fan of the TUI form factor for longer running, more ambitious features. Even with a classic "Add an endpoint, tweak the infra, consume in the frontend", plans get awkward to refine in markdown files, especially if everything lives in its own repo.
I wanted something like Plannotator, that could also work for the execution, not just the planning, So I've been working on something that turns Notion into the memory and orchestration layer for agents.
Underneath, it's a plan-implement-review loop, but you get a nice Notion page with a kanban board out of it. You can easily link your existing documentation, collaborate by sharing the page, annotate and comment to steer the planner, and you get versioning out of the box.
Because Notion acts as the memory, you can just open the page after a long weekend and get your agent and yourself back into the full context. You can see what's been done, what's left, or what requires human input just by looking at the board. You can ask it to fetch the comments on the pull request you raised, and it'll fetch, validate the comments, give you a report, and update the plan/board if necessary.
I've been using it exclusively for the last two weeks, I'm quite happy with it. It's been really fun to build the exact tool I wanted.
By tuning the agent, it is possible to create trading strategies [1] and reverse engineer websites in order to create optimized JSON APIs using the websites internal private APIs. [2]
I'm having the hardest time communicating what is happening so next I'm going to try to explain it using data visualizations so people can visualize it in action.
[0] https://github.com/adam-s/agent-tuning
[1] https://github.com/adam-s/alphadidactic
[2] https://github.com/adam-s/intercept?tab=readme-ov-file#how-i...
I recently finished <text> and <filter> support, now I’m working on a GPU-accelerated rendering backend.
This is a fun side project as I learn great with email communication, culture differences (as a dev)
If you’ve been through this rodeo too, please provide your feedback — your feedback will help make next summer a lot less stressful for other parents
I’m interested too, but don’t have amazing patience to dig into it.
For me this is an example of when you become aware of something you see it all around.
I'll writeup a fuller list and what I learned along the way.
Its like a microservices architecture with NATS JetStream coordinating stuff. I want to keep the worker core as clean as possible, just managing open sockets, threads and continuation.
Document querying is something I am interested in also. This system allows me to pin a document to a socket as a subagent, which is then called upon.
I have hit alot of slip ups along the way, such as infinite loops trying to call OpenAI API, etc ...
Example usage: 10 documents on warm sockets on GPT 5.4 nano. Then the main thread can call out to those other sockets to query the documents in parallel. It allows alot of possibilities : cheaper models for cheaper tasks, input caching and lower latency.
There is also a frontend
Alot of information is in here, just thoughts, designs etc: https://github.com/SamSam12121212/ExplorerPRO/tree/main/docs
But at my current knowledge and practical work, its like giving a chimpanzee a nuclear reactor schematic. But it's a passion project idea of mine, I really want it to become real one day. Personally, I feel like something much better can be made than current solutions.
https://lotuseater.epiccoleman.com/
It's a mostly vibe-coded fan site for jamtronica greats Lotus. I wrote/prompted a scraper to pull in setlist data from Nugs and have been having a lot of fun coming up with cool data analysis stuff to do with their sets.
I've seen them 7 times (chump change compared to some fans) and was starting to get certain intuitions about like, "if I hear song X that probably means they won't play song Y." For example, one of my favorite Lotus tunes, It's All Clear To Me Now, seems to fulfill a similar "function" as another song - Did Fatt.
It was pretty cool to see that intuition bear out in the data (they've only ever been played in the same show one time in over 900 total shows).
I've got a bunch of other "data" features sitting in a PR in my Gitlab, need to get around to reviewing and testing it so I can push out the next update. Also have a few other ideas for it, although I think there's probably a point coming fairly soon where there's not really anything left to do.
I posted it on the main Lotus fan group on Facebook. I have a grand total 8 users. I love those users.
The site is nothing crazy, it will never make money or anything - but it's just been a ton of fun to have something cool to hack around on.
I'm working to make it better right now.
And I do have a basic UI at https://workglow.dev/ (where you can run the workflow, though if you use AI models, the models will run in the browser -- if you want to run GGUF models, please signup for the desktop app waitlist).
Used to do it for friends only, but been publishing publicly since recently and it’s fun.
“Senior dev, junior attitude”
https://youtube.com/@harlybarluy
Spent 3h today adding a “system” filter to jq only to find out there are like seventeen PRs for this going back ten years. T_T I live but I don’t learn.
My take away from that perspective is: be honest. IMO the best moments are me just failing. It's probably more fun and more instructive to see me struggle than to see me breeze through things.
And it better be entertaining because I work on stuff absolutely nobody cares about anyway. XD Right now I'm writing a microformats2 -> RSS converter in JQ...
Today was my first time on Twitch, which is way more social. Random people drop in and start talking to you. Very cool. Very different from youtube live, where it's only the people who already know you, IME.
Live Kaiwa (https://livekaiwa.com/) — A real-time Japanese conversation assistant. It listens, transcribes, translates, and suggests responses so you can follow along in conversations you'd otherwise get lost in. I built it because I live in Japan and needed something for the situations where missing a nuance actually matters — PTA meetings, bank appointments, neighborhood councils.
Have fun trying it and let me know what you think!
I've been shooting for the moon with one experimental idea after another (like many others) testing out LLM capabilities as they develop, for at least 2yrs now.
I'm still very excited about how these new tools are changing the nature of software development work, but it's easy to get into this frenetic mode with it, and I think the antidote is along the lines of 'slowing down'.
The reproduction has been one of the things I've been struggling with in regards to consistency of bringing up the right envs. At the moment I've been approaching it as a MCP server that holds a few tools to bring up specific versions or branches of my stack to then find where a bug was introduced, build that commit prove that it wasnt in the previous one, and then fix it and run the full stack again with the fix component, then run through our local integration tests.
This is the stuff that makes me feel like I'm on steroids now, my whole dev debug process can be run with a few instructions, game changing.
"The irony of Backstage is that it was created to prevent teams from having to reinvent the wheel every time, building and maintaining their own developer portal. But that's exactly what everyone does with Backstage."
We wanted something you configure,deploy,update. thats it.
service catalog, GitHub crawler, K8s entity discovery via k8s-push-agent, Forge + molds (scaffolding/workflows, like Backstage templates), governance, scorecards, cloud provider resources, license management, event based notifications, team-context aware, API keys with scope auth alongside session RBAC. CLI and Terraform provider too.
We're aiming to release Beta end of April.
Most iOS/Android mental wellness apps are trying to be everything for everyone, ie general AI journaling or meditations.
By niching down, we can build the best experience end-to-end for anyone that resonates with these particular emotional challenges.
DailySelfTrack is a customizable combination of habit tracker, health journal and diary.
It should be as powerful as a spreadsheet for self-tracking, but the daily usability should be more on par with a habit tracker app.
For example my use-case would be:
- Journaling in a way that fits into what I need. (Gratitude, bullet point jounal)
- Analysing my health and understand how things might relate to each other. (State of multiple health issues)
- Support for moving closer towards achieving my goals. (Daily focus sessions, no-phone mornings, learning Korean)
My website: https://bryanhogan.com/
The repository: https://github.com/BryanHogan/bryanhogan
It's built with Astro. Uses markdown files for the blog. Just CSS, no Tailwind or other UI library. I recently switched to Sveltia as the CMS, and after a bit of custom CSS for fixing some issues it has it works well for writing on my phone!
If you do any freediving or apnea training, interested to hear what you think of the platform.
I've been making a browser-based PDF editor that runs on-device via Webassembly / PDFium. Many of the hard parts were done by the open source embedpdf project, and I've been adding my own custom tools on top of it.
It does the usual annotation stuff — highlights, comments, stamps, etc. working on some more advanced stuff now - regex search/redact, measurements and takeoff tools for AEC industry.
I always have growing lists of short texts, facts, and links that I wanted to host on a standalone site rather than burying them in a notes app. The workflow is simple: a browser extension to clip links with remarks, which then feeds into a public-facing list.
I’ve also added a "Substack-lite" feature. Instead of long-form writing, it lets you send simple roundup email digests (e.g., "Top 5 links this week") to opt-in subscribers.
My personal blog (wenbin.org) is currently powered by the tool.
CurateKit.com is in private beta while I'm fine-tuning a few things now, but I’m opening up invites to the waitlist over the next few days if anyone wants to give it a try.
I'm also just working on my game, Antorum Online. Made with Rust and Unity. https://antorum.online
It takes my favorite elements from games like: WoW character min-max design and rotation Diablo 2/POE for item and crafting inspiration Slay the spire dungeon flow/fights.
It is uses pixel art I commissioned a decade ago that I am looking to finish a game with.
Looking for some early feedback! https://crux.lakin.dev/
So there is going to be a need for Instant Messaging for AI Agents - Launching soon. https://agent-socket.ai
I tried to look for what other solutions are available and I've collected all the best open-source ones in this awesome-style GitHub repo. Hope you find something that works for you!
An Android mobile app to send e-mails to myself(capture mechanism from GTD)
Apache Shiro PMC chair (trying to get financial support for the project) https://shiro.apache.org
Jakarta EE Components: https://github.com/flowlogix/flowlogix and it's starter: https://start.flowlogix.com
Working on all of these for the last 15 years, looking for more exposure.
An important feature for me was improving the recipe discovery experience, you can build a cookbook from chefs you follow on socials (youtube for now), or import from any source (Web, or take pic of cookbook etc) - it then has tight / easy integration into recipe lists.
Utilising GenAI to auto extract recipes, manage conversions, merge/categorise shopping lists etc - as-well as the actual recommendations engine.
If anyone is interested in beta testing / wants to have a chat I'll look out for replies, or message mealplannr@tomyeoman.dev
I also make small games with Godot.
Try it here - https://burrow.run/
This project brings in a lot of AI support. It's made a massive difference. The original project took two years to finish (actually four, but we did a "back to the ol' drawing board reset).
It looks like this may only take a couple more months. I've been working on it for two months, already, and have gotten a significant amount done. The things that will slow it down, will be the usual sand in the gears: team communication overhead. Could stretch things out, quite a bit.
https://greenmtnboy.github.io/sf_tree_reporting
Posted in last thread when it was SF only: https://news.ycombinator.com/item?id=47303111#47304199
The Israeli tech industry isn't a neutral commercial sector, it's a deliberate pipeline from intelligence units to billion-dollar companies. Wiz ($32B Google acquisition) was founded by four Unit 8200 veterans. SoftBank's Israel ops are run by a former Mossad director. CyberStarts, a $1.5B VC fund, openly recruits Unit 8200 graduates.
The goal is to provide AI agents with deep understanding of the codebase and help them understand the context, not just text
- Any containerized app, uses Fargate (no Kubernetes)
- Heroku-like CLI tool with instant console sessions
- Set up SQL/Redis instantly with Heroku-like add-ons.
- Autoscaling, preview apps, audit trail, release approvals.
https://tapitalee.comIt's a free sobriety app for any bad habits I built for myself. Most sobriety apps reset your counter to zero when you slip, but it uses a Github style contribution graph to show you how far you have come. I also use it to track urges, and store a toolbox that is a reminder if why I am quitting something and what I can do instead every time I have an urge.
A privacy first transcription and analysis app for iOS and native Mac OS (latter this week)
All AI runs on device, nothing ever leaves your device apart from syncing data via your iCloud.
Just launched the blog too
I'm leaning heavily on simulation, economics, towns with real economies, and interweaving progression systems. It's a custom engine. I finally have the foundation built, it's multiplayer ready, and it currently loads in under 200MB. The idea is to be hyper efficient to simulate multiple towns that grow by themselves and you can trade and interact with.
https://www.youtube.com/watch?v=BeZ3O6F5FXU
It's a free-time project, but I will happily take investment and make it my full-time project. :) I have a game design-doc that I have built out, and I personally like it a lot. I believe in it's potential.
Right now I'm focused on the stats side. It already shows how much time you spend in each app, and I'm adding website tracking too, which should make the picture much more useful.
I'm also working on better break timing for dictation. LookAway already delays a due break if you're in the middle of typing, so it does not interrupt at a bad time. Now I'm trying to extend that same behavior to dictation as well, which turns out to be a pretty interesting detection problem because it overlaps with some of the other context signals I already use.
Most of the challenge is making it smarter without making it feel more intrusive.
Delinking is the art of stripping program for parts, essentially. The tricky part is recovering and resynthesizing relocation spots through analysis. It is a punishingly hard technique to get right because it requires exacting precision to pull off, as mistakes will corrupt the resulting object files in ways that can be difficult to detect and grueling to debug. Still, I've managed to make it work on multiple architectures and object file formats; a user community built up through word of mouth and it's now actively used in several Windows video game decompilation projects.
Recently I've experimented with Copilot and GPT-5.3 to implement support for multiple major features, like OMF object file format and DWARF debugging symbols generation. The results have been very promising, to the point where I can delegate the brunt of the work to it and stick to architecture design and code review. I've previously learned the hard way that the only way to keep this extension from imploding on itself was with an exhaustive regression test suite and it appears to guardrail the AI very effectively.
Given that I work alone on this in my spare time, I have a finite amount of endurance and context and I was reaching the limits of what I could manage on my own. There's only so much esoterica about ISAs/object file formats/toolchains/platforms that can fit at once in one brain and some features (debugging symbols generation) were simply out of reach. Now, it seems that I can finally avoid burning out on this project, albeit at a fairly high rate of premium requests consumption.
Interestingly enough, I've also experimented with local AI (mostly oss-gpt-20b) and it suffers from complete neural collapse when trying to work on this, probably because it's a genuinely difficult topic even for humans.
I manage a small store (https://amigurumis.com.mx) for my SO and im dropping Elementor (too expensive) to use only Gutenberg. Turns out that it is pretty good for simple sites.
Im having some sucess developing new websites for people who cant afford it, or who never though about having one, so i created one for an accountant (https://contadoranual.com) using only WordPress.
The business model is likely going to revolve around mcp and x402 https://micro.mu/developers/
Its a fun project, all done using free tier.
I built it because I was sick of paying for complex invoicing tools that charged monthly fees for features I never used.
Let me know if you want to try it out. I'll be happy to set you up with an account.
The idea is to make agent, MCP, and API interactions verifiable across org boundaries instead of relying only on logs. Still early, but that’s the thing I’m most focused on right now.
Originary: https://www.originary.xyz PEAC: https://github.com/peacprotocol/peac
Official app is mobile-only and clunky, and the workflow is awkward if you're sitting at a desk. Hardest part has been maintaining compatibility across amp models. Small protocol changes or optimizations I make for one amp can break another. That means I have to do a lot of manual testing before every release. So I'm trying to think of an emulation layer or test harness I can build to make my life easier. Happy to hear suggestions there.
About ~50 people are using it so far, and main feedback has been that it's much faster and more reliable than the official app.
[1] https://tonepilot.app [2] https://www.positivegrid.com/products/spark-2
You commit to a habit, invite your friends to join, and keep each other accountable.
Little square for each day/week fills up depending on how many members of the Pact completed it. Streaks are dependent on everyone in the pact completing.
https://apps.apple.com/us/app/pact-accountability/id67551314...
https://gist.github.com/paulshomo/69cf99e3185fa7ad0f50fc0e38...
https://colinator.github.io/Ariel/post1.html
I just got a bigger robot, further results forthcoming!
So they can go 'slow', by taking a camera image, controlling the robot, repeating. Or they can write code that runs closer to the robot in a loop, either way. I thought the latter was somehow more impressive, and that's what you see in the hand-tracking example.
Code: https://github.com/opensciencearchive/server.
Website: https://opensciencearchive.org/
Two demos:
I've got demos up and running (mirroring/extending PDB and GEO). Next I'm working on APIs with good AX, ML-friendly export, and an unified AI-driven UI that works for all scientific data types.
Based on top of that is Caution, the first FOSS general purpose verifiable compute platform launching next week in private beta.
2. the other projects it the framework. Aside from the product itself, I've ended up with a really nice framework (FE and BE) and playbook for copilot to follow. I've hit multiple problems with AI generated code and had to rework it like I have for junior devs! But now, the framework focuses the work and stops the slop!
I want to build out all the product-dev-helper tools I've wanted in the past. I've already got a lovely schema-UI system, UI components which are data-aware and the basis of some low-ish-code tools. I've also nearly got a "run tests and fix" local LLM which saves tokens.
Really enjoying this.
Still deciding whether to ship it as a product. gauging interest here first.
so far it has been an interesting journey and I have had some success but the whole process has led me to write a lot of software around my own process so that I can scale it.
Might turn that into a product itself.
So people don't need to lose braincells over this till it actually matters.
Lately I’ve been having LLMs implement multiple analysis methods on my session transcripts, trying to surface and identify patterns.
It’s been interesting. It took quite a bit of nudging, but Claude applied techniques I didn’t expect, from disciplines I wouldn’t have thought of.
If it works out, I’d like to turn into a sort of daemon that locally runs analysis on the sessions of users, with a privacy-preserving approach (think federated machine learning).
Would be interesting to see what patterns appear at scale, and have those confirmed or rebutted across thousands of transcripts corpuses. No reason Anthropic & OpenAI should be the only ones to benefit from that; those are our interactions after all.
Do you have any example?
Another one is "Lag Sequential Analysis", applied to human-agent interactions.
I was only thinking of corpus analysis, but I guess that’s what you get when you give AI a web search tool and keep pushing it to explore more domains to borrow techniques and methods from.
Yes, you can use your own API key as well.
For people who use Fora for travel, a tool that uses AI to create google calendar events from travel itineraries: https://itinerary.projects.jaygoel.com
A job board for travellers and backpackers on working holiday visas in New Zealand.
Most NZ job sites are built for employers. Farmdoor aims to flip that: workers can leave reviews of farms and employers, so the next person knows what they're signing up for before they show up somewhere remote.
Built it after seeing firsthand how hard it is for backpackers to find reliable work and how little recourse they have when an employer turns out to be dodgy.
Very fun project, launching this week publicly in the app store.
I am currently rewriting+testing the engine and about to add ~400 games to my platform in a few weeks.
https://housecat.com/docs/editorial/why-housecat
The ideas I’m thinking about is: what’s old is new.
We’re seeing a massive influx of people writing software and administering servers for the first time ever. But so many people are jumping (or being pushed) into the deep end without basic training.
Lots of opportunities for us older admin folks to build, teach and help all the new folks.
The core idea: every AI agent acting in the physical world must formally earn the authority to act, tier by tier, from informational alerts through to safety-critical relay control. Runs offline on a $55 Pi.
First deployments are underway in Lagos.
Happy to answer questions about the safety architecture or the offline reasoning approach.
Repo: https://github.com/jbonatakis/pginbox
Makes reading/searching the Postgres mailing lists easier.
I’m polling a Fastmail inbox to nearly instantly receive and ingest messages. Anyone can browse without an account, but registered users can follow threads to be notified of new messages, threads in which your registered email is found are auto-followed, and there are some QOL settings.
Search is pretty naive right now (keyword on subjects) but improved search is the next big thing on my list.
This is less of a latency/efficiency thing and more about disconnecting the eyes from a screen and fingers from a keyboard. The upside is more walking, flow and creativity.
Rn it's on the appstore: https://apps.apple.com/us/app/lexaway/id6761870125
I couldn’t find any that were as nice or as powerful to use for writing JSONPath queries, so instead of spending an hour crafting and testing them manually, I spent >40 hours building this tool to save myself half an hour.
We use AI to monitor hundreds of local government commissions and give real-time intelligence to B2B, residents, and governments. If you're a business trying to track what's happening in local gov for your policy, sales, or lobbying team, I'd love to chat.
One thing I find especially intriguing is how LLMs can help deal with desinformation:
- I experiment with deterministic settings of local LLMs for the document summary so that sharing a prompt would prove that the output was not tempered with (no desinformation on the service side)
- I add outputs of several LLMs (from the US, the EU and from China) for the "broader context" section so users could compare the output (no desiformation on the provider and model side)
I'm looking for artists to help fulfill the vision.
You write a short entry, keep it private or share it to a circle. A circle is a small private group of your own making — family, close friends, whoever you'd actually want to hear from.
Basically private instagram without all of the strangers and ads. What social media used to be.
- immutability
- self-hostability and/or EU SaaS option
- nested data (e.g. nesting a list of sailing legs into a sailing trip form)
- formulas (today(), date, string, numeric,...) and conditionals (visible/required/enabled if)
My goal will be to create an exceptionally cost effective tool, scaling well with usage and not paywall blocking advanced features. This may sound weird, but I think this is a real challenging and good goal to follow, enabling users more than optimizing for the highest payer. I thought about having a tool for a few $/€ per user per month where others charge 10x.So I created two nice pieces out of that, which would have been impossible in the past due to time constraints and got massively unlocked through claude code:
- a frontend/javascript only forms library that supports all the rendering, form schema input, data output, validation and formula/conditional logic
- a multi-tenant SaaS product, that is a single golang binary and stores in sqlite, easily self-hostable but I can also operate it as a European SaaS (and in other regions) where needed
This is also a test run in terms of tech stacks and trying new things I wanted to try for long time. It's mostly evening or weekend coded due to my regular day job, but made such incredible fun. The AI coding part really provided me the time to work on the product, polishing, UX and worry less about the "work" part of coding. My experience seems to help a lot to gain leverage and increase the fun factor and complete immersion into coding, that I kind of almost lost in the past.So I was trying:
- pocketbase
- really running something bigger with much more data off of sqlite (primarily used it for smaller stuff in the past)
- real focus on self-host-ability, keeping dependencies minimal and extremely simple (which also helps claude)
- trying other tools for security scanning, verification, testing, security analysis, WAF,... than I use at work, pretty much playing around with tech as much as I can to see new and different stuff :-)
Not ready to share a repo yet, but if anyone is interested please ping me on hello@devopsicorn.comhttps://github.com/NetwindHQ/gha-outrunner - github actions local, ephemeral runner which runs jobs in docker container, tart vm org kvm (depending on the host/guest)
Orange Words. My hobby project, a hacker news search system. It was initially created by hand and now I use AI augmented development. It's a good low risk environment for experimenting.
Months 1-3 were about building a desktop client. Now I'm working on a server binary customers can optionally self-host to share dashboards publicly and run workflow automations.
I've converted my 23 year old Java desktop app to a website.
It's an app to make searching eBay an actual joy. Perform a search, then highlight text to trash or group that term. Then perform the search again tomorrow and it will hide all the stuff you've already seen.
It scans your claude and codex history to find edits and matches those to git commits (even if the code was auto-formatted).
You can browse all 364 prompts that wrote 94% of the code here:
It's a tool that use QDrant, a vectorial db, to embedding the texts chunks: LLM api is questioned to generate the Q&A pairs from a chunked texts.
Each chunk is then embedded and stored in the vectorial db to facilitate the Q&A generation, thanks to better context informations.
This tool helping people to study everything thanks to even Spaced Repetition algorithm.
Games, utilities, calcultors (for whatever niche), and anything else where I wanted it accessible for me from anywhere, plus the poeple I want to share with (publicly or privately)..
So, I built this:
Simple static site hosting. Upload html or a zip-containing-html along with other needed files, and it gets hosted on a subdomain with full https. Optionally, password protect it, or generate shareable links. Also, detailed analytics and other stuff.
Im already hosting 16 small sites on it.. loving it.
https://agjmills.github.io/trove/
Go, docker, bit of alpine js
A SSO application in rust(not public)
A DNS for a dream project of mine which is a hosting provider company like digital ocean but in Scandinavia(not public).
A code hosting site for said hosting company called bofink(not public)
Ansible playbooks for applying database patches that can resume and create schemas etc, based on an internal tool from a former job. This is public and available on my github if anyone wants to look at it not linking it because there are way cooler projects here.
Blockers: gravelly clay is a pita to dig with a shovel
Example: Companies that use Github: https://bloomberry.com/data/github-enterprise/
This week I added TTS support, which needed multiple inference pipelines, it was not easy to find models for 50 languages!
At this point, it mostly works as a crude implementation of Google translate+Google lens, but 100% offline and 100% Google-free
https://slidebits.com/ai-streamer
Not a trivial thing to vibe code without any domain expertise but this project took me under 2 weeks with a AI coding agent harness I built myself. I use Gemini 3 Flash as my main driver as well.
Wanted to have a way to coordinate multiple agents on Linux either via SSH or locally and figured out why not give it a shot?
The result is a pretty cool tool, inspired by similar solutions that after trying them most fell short.
https://kintoun.ai - Translate Word, Excel and PowerPoint documents with layout and formatting intact.
https://ricatutor.com - Your AI Language Tutor for YouTube
Currently working towards a big release to go out by end of the month.
Real challenge to keep it working 24/7. The Android OS, and its modifications are really aggressive, trying to kill everything that runs more than they think it is allowed to.
I made a whole article about it. I hope it will help others: https://dev.to/stoyan_minchev/i-spent-several-months-buildin...
Also, Arch Ascent, which is a tool for evolveing microservice-heavy architectures.
https://github.com/mikko-ahonen/arch-ascent/blob/main/doc/de...
At the moment working on the 3rd party development tools so in the future anyone can make their game dev dreams a reality and make a simple and fun multiplayer party game for the Gaming Couch platform, ideally in only one weekend!
If you're an interested game dev that would like to beta test the dev tools, hit me up either here, via Discord (link available from https://gamingcouch.com) or by emailing me at gc[dot]community[at]gamingcouch[dot]com!
The TL;DR of Gaming Couch:
- Currently in free Early Access with 18 competitive mini-games.
- Players use their mobile phones as controllers (you can use game pads as well!)
- Everything is completely web-based, no downloads or installs are necessary to play
- All games support up to 8 players at a time and are action based, with quick ~one minute rounds to keep a good pace. This means there are no language based trivia or asynchronous games!
I use it daily and so do others, for - better UX, feedback, and review surfaces for ai coding agents.
1. Plan review & iterative feedback.
2. Now code review with iterative feedback.
Free and open source https://github.com/backnotprop/plannotatorPatch for linux kernel adding support for enforcing Landlock rulesets from eBPF. In RFC stage now.
The trade-off seems reasonable so far. By going static, the main thing I lose is comments.
The project is still in progress, but I made solid progress over the weekend.
The project is here: https://github.com/yusufaytas/yapress
It has some interesting applications for building high performance clients for mssql with tds protocol implementation. The APIs allow almost direct data serialization to wire instead of datatype materialization in rust. Makes for a suitable contender for high performance language interop.
Here's the MVP interface: https://bcmullins.github.io/reading/
I appreciate any feedback. Hope you find something interesting to read!
It includes bill of materials, purchase/production orders, "can I make n?", stock takes, multiple stock locations, and barcode scanning. It's aimed mainly at small business and makers for the time-being, but still allows multiple users to connect over the the local network.
Sure, my email's in my profile, I'd be happy to chat.
Yupcha AI Interviewer, handles the screening, video interviewing with conversational agents.
Check it out https://yupcha.com
Working on a oss video dubbing, cloning and design studio
Check out https://github.com/debpalash/OmniVoice-Studio
Suggestions are welcome.
https://github.com/flipbit03/terminal-use
I'm super proud, because it came to my knowledge that someone at Codex used my tool to debug codex+zellij issues, by running zellij within `tu`, and then codex inside zellij
I tinkered for a minute but never got anywhere.
The short version: each layer trains itself independently using Hinton's Forward-Forward algorithm. Instead of propagating error gradients backward through the whole network, each layer has its own local objective: "real data should produce high activation norms, corrupted data should produce low ones." Gradients never cross layer boundaries. The human brain is massively parallel and part of that is not using backprop, so I'm trying to use that as inspiration.
You're right that the brain has backward-projecting circuits. But those are mostly thought to carry contextual/modulatory signals, not error gradients in the backprop sense. I'm handling cross-layer communication through attention residuals (each layer dynamically selects which prior layers to attend to) and Hopfield memory banks (per-layer associative memory written via Hebbian outer products, no gradients needed).
The part I'm most excited about is "sleep". During chat, user feedback drives reward-modulated Hebbian writes to the memory banks (instant, no gradients, like hippocampal episodic memory). Then a /sleep command consolidates those into weights by generating "dreams" from the bank-colored model and training on them with FF + distillation. No stored text needed, only the Hopfield state. The model literally dreams its memories into its weights.
Still early, training a 100M param model on TinyStories right now, loss is coming down but I don't have eval numbers yet.
The idea is that the brain uses what the authors refer to as "feedback alignment" rather than backprop. Even if it turns out not to be literally true of the brain, the idea is interesting for AI.
I also love the idea of grafting on the memory banks. It reminds me of early work on DNC's (Differentiable Neural Computer's). I tried to franken-bolt a DNC onto an LLM a few years back and mostly just earned myself headaches. :)
It's fun to see all the wild and wacky stuff other folks like myself are tinkering with in the lab.
Video demo: https://youtu.be/cJfFAh6ox84?si=WScDPzI4rJIKe99n
A site for looking up strata information for apartments in NSW, Australia
[1] https://apps.apple.com/us/app/reflect-track-anything/id64638...
Already using it for my SQLite driver, and already in use by some a few other projects: https://github.com/topics/wasm2go
Native APIs exposed via Rust, but the core framework is written in AssemblyScript. Games or mods/libraries built in it are also written in AssemblyScript.
It builds as a binary that can run on the various PC, mobile, and web platforms. You run it and you get a claude-code-like console that has access to a sandboxed filesystem to put game code in, and a git repo, all built in.
working on an AI-native Kubernetes sidekick that watches your pods, reads the logs, and turns failures into clear fixes before they become outages
It's free, no sign up or ads - feedback welcome :)
It's still VERY much in development but I'm building a site that allows people to find TTRPG games that are suited to them AND includes a suite of tools for both GMs and players in said games.
Players will be able to showcase characters they're playing or have played and GMs can manage campaigns (scheduling, notes). I'm a D&D player but I'm trying to make it system-agnostic
I have a terraform setup right now but it’s super awkward and very slow. The goal is to be able to define settings using PKL which looks super interesting. Wanted to try it out for a while now.
I think it works quite well so far, but need to tweak the camera algorithm a bit to make the buttons work better. Thinking about more games to add as well.
Turns your project's GitHub release notes into user changelog that your users actually want to read.
Surfer is fantastic, and the developers of Surfer are pretty great people too! It has been on my to-do list to learn Spade.
Members takes turn pitching one album per week. Support comments and a handful of emoji-based reactions.
Integration with Spotify for easy pitching and playing (by links only, users are not required to have a Spotify account).
Plan is to keep the clubs fairly small and invite only.
Building it in Gleam which is a lot of fun!
Right now I‘m working on adding a „simulation“ mode, that allows anyone to get free fake responses during development, instead of pricey real generations.
Take a look here : https://voiden.md/
ziglag (https://github.com/level09/ziglag): self-hosted invoicing for freelancers, built on top of stk. clients, invoices, VAT, PDF, shareable links, MIT. got tired of paying a monthly fee to send a pdf.
idea is to keep chipping away. every subscription that annoys me is fair game. small tools, self-hosted, no accounts, no seats, no upsell. if it's useful for me someone else probably wants it too, so might as well open source it. open to ideas on what to kill next.
I made a classless CSS library, then migrated most of my projects from PicoCSS.
I also made a quick logo generator: https://logo.leftium.com/logo
I'm surprised that no one has done this so I decided to give it a try.
I wanted to make JSON/YAML configuration language for my projects. And i wanted a strict specification. This is want i created, now with specification and 100% coverage, reference implementation it’s just one prompt to reimplement parser in another language.
The query engine itself is like a DAG of 'operators', similar to a relational DB (or more like a graph one) with scanners, filters, and matchers.
Very fun, although not at all efficient and probably overengineered for what it does :)
https://github.com/BVCampos/operator
It has been working quite well.
Swiss army knife CLI tool written in Swift using only native Apple frameworks.
The primary goal of this project is to demonstrate how many Apple standard library frameworks can be meaningfully used in a single, actually-useful CLI tool.
brew install jftuga/tap/swiftswiss
First time doing this sort of thing with agents. So far it seems ok?
If it works out it will really help us scale and improve a legacy application that so many depend on at the moment. Wish me luck!
It's called MatGoat[1], and it's going quite well so far. Nowadays I'm working more on the marketing/sales side.
Feedback welcome
A Tauri 2 CLI / MCP that allows your agent to debug, take screenshots, run JS, etc. inside a Tauri app: https://hypothesi.github.io/mcp-server-tauri/
It's in rust with egui, and should help folks to do that without the cli.
Not ready for prime time yet, but available at https://github.com/almet/signal-without-smartphone
The mission is to incentivize better thinking. For each game there's an AI judge that scores everyone's answer based on a public rubric (style, cohesion, logic, etc).
Currently uses fake money and ELO score but thought it could be a very interesting competitive game for real stakes.
Any feedback is much appreciated.
started to explore a iPad-focused Dungeons & Dragons DM app. i called it Campaign Codex. https://campaigncodex.app/
been doing a lot of agent assisted iOS dev...it has been...fun!!
I am hoping to launch in about a week, so I would love any user feedback! (email in profile)
A solution set to the book Pattern Recognition and Machine Learning by Christopher Bishop.
A work in progress.
Very early demo with a smart dum-dum RL agent here:
Data engineer, 20 yrs software / 10 in ag-tech. Picked up beekeeping and was surprised how much structured data a single inspection produces, and how nowhere useful exists to put it. It's a gloved, veiled, honey-and-propolis-covered activity. Tapping through a mobile UI mid-inspection is not ideal, and good luck getting your phone back clean.
The core is a virtual hive model. It's all mutable state: boxes, frames, components, queens, and colonies you rearrange to mirror the physical yard. Treatments, feedings, and inspections layer on top.
This summer I'm shipping voice-driven inspections: narrate what you see frame by frame, STT + LLM pipeline extracts structured data and maps it to your hive model.
If you have beekeeping friends, I'd love it if you could send it along <3. I won't claim it has every feature under the sun, but I work on it every day and have a strong roadmap ahead.
Also open to critiques. Thanks!
- make it reliable to run LLM inference on company hardware, even when it is poor or outdated
- bring chaotic agentic behavior under control in business contexts
Posted a show hn earlier today that didn't got any traction : https://news.ycombinator.com/item?id=47738516
Fully local, hobbyist friendly, agentic workflows work great with it since it’s just a CLI.
I've been wanting to do this for years. I fully support (and have paid more than most into) John's shareware, but that means that I can't just "apt install" it, which means I rarely have it available on my various machines. Having something I can just "uv run" that keeps most of the same ergonomics would be a nice alternative.
It gives you a detailed breakdown of what's missing, step by step guidance on how to fix each issue, and shareable report links. Excellent resource for security teams of all sizes.
Scans HTTP headers, TLS/SSL, DNS security, cookies, and page content. Free to get started, with a REST API for integrating scans into your CI/CD pipeline or monitoring. Also supports capturing and reporting CSP violations.
Have you heard of Superformula ? I remember playing with them few years ago.
Looking for people who know hardware well. Let's get to know one another on a flight to Shenzhen :P
Ever been recommended supplements? Now you can find out if they work
Eventually I got scope creeped into a full game with branching stories, item crafting, and a custom cutscene engine...even Trained a model for a few specific art assets.
Program your amateur radio via the web. Uses pyiodide + chirp drivers under the hood + WebSerial.
I've been working towards a new platform that mixes fantasy sports with stock market mechanics. My first public project, I just launched a few week ago. No gambling, free to play (despite the .bet):
A tool to estimate if you should vibe an automation/app or just buy/delegate/grind instead
turns out starting a popular open source project comes with ongoing work attached
I do a lot of data science and analytics in my real job.
https://i.imgur.com/uHpnUox.png
(Don't mind the errors or display issues for now; this project will report which versions of programs are installed on the given computer, and which ones could be updated. A few smaller bugs still remain but the main rewrite was finished now.)
For work-related reasons I have to expand on some python tasks in the coming weeks. Trying to find something interesting here, but I guess I'll just focus on various bioinformatics-related tasks (also related to work). Programming for the most part is not extremely interesting - the only part I actually like is any creativity and useful end results. Fixing bugs is annoying to no ends. Writing documentation is also boring, but can not be avoided.
and a gift for my friend's birthday.
My motivation for creating CompterPoker.ai was feeling a bit overwhelmed by some of the professional poker tools out there for learning GTO play. For some tools, learning how to simply operate the tool itself felt like a second job. With ComputerPoker.ai players can play against bots themselves simulating GTO play to learn what it "feels like" to play GTO vs. GTO opponents without having to turn any knobs or dials (feedback is real-time as you play).
The Beta tester code for HN Users is: HackerNews2026. All feedback is welcome! Please send suggestions for improvement or bugs to contact@computerpoker.ai or alternatively leave a comment below. Any questions I will do my best to answer.
As for the product offering the website is designed to teach players how to play optimal poker strategy (GTO) in simulated Texas Hold 'Em poker tournaments. Our value proposition is that if you can consistently beat the bots then you will fare well in live poker tournaments (of course adjusting for your opponents' play).
In addition to GTO pre-flop quizzes and pre-flop charts, users have the ability to simulate poker tournaments from start-to-finish and get feedback on their decisions _in real-time_ in a fun and low-risk environment.
For those interested the tech stack is Django deployed on AWS via Terraform and SaltStack, the database uses a Postgres RDS backend, and the frontend uses HTMX with WebSockets via Django Channels and Redis (Nginx serving as reverse proxy with CloudFlare DNS and SSL). During the project I used Claude Code to aid with various boilerplate aspects of the code base including building out the repos for Terraform and SaltSack and of course speeding up Django development.
Users are graded pre-flop based on the covered pre-flop scenarios (two-ways only for now). Post-flop users are graded on a residual MLP PyTorch model. We have built an in-house solver in Rust using the discontented CFR++ algorithm. The PyTorch model approximates GTO play post-flop (again only two-ways currently) based on training data with raises, EV, and realistic ranges for OOP and IP players. Because the post-flop decisions are based on a model that will always be a work in progress I refer to these decisions as GTOA (or "GTO Approximate").
Version 8 of the PyTorch model is the first one that I am happy with and actually find it quite difficult to play against. If you manage to beat the bots please do let me know how many tries it took! For those curious the PyTorch params for the most recent run are below (I trained on a gaming PC via Linux WSL2 using an AMD GPU).
The website is live in Beta mode as I gather feedback on how things are structured and work out any bugs/kinks. If you have any suggestions for improvements I’d love to hear them. Subscriptions are live so if anyone wanted to test the Stripe payment processing flow I certainly wouldn’t mind! ;-)
p.s. This is a side gig for me. I am currently looking for full-time work either fully remote or on-site based in London, UK (this LLC that runs ComputerPoker.ai operates out of USA but I am based full-time in the UK and authorized to work in both UK and USA). If you or someone you know is looking for a SRE with strong software engineering skills please let me know!
Deployment tool with security gates.
a sqlite database that can be version-controlled by git alongside source code
Simracing trainer.
I love simracing, I'm moderately competitive and want to improve, and I like to be efficient with my practice. So having access to and using a lot of telemetry, I noticed that the "turn a few laps, load telemetry, compare against reference lap, try again" is not as efficient as it could be.
Also a lot of my telemetry analysis is very rote and "rules based": Look at the biggest laptime delta jump against reference, try to determine the cause among a few usual suspects".
So I have started experimenting with a system that reads the iRacing telemetry in real time, and compares against the reference telemetry live, finding the biggest delta jumps, and trying to find the root cause of the time loss using an increasingly sophisticated GOFAI rule and pattern matching system. Then this report is fed to a cheap LLM call to be condensed into clear advice, and the result goes to the free Microsoft TTS API. So I get instant feedback of where I'm slow and maybe even why.
So far I fear it's mostly making me faster from all the test laps involved more than the advice itself, but when it clicks it does feel magical and really help.
But sometimes I feel like I'm just speedrunning the collapse of 70s AI, as it feels a bit too brittle and situational.
I also have added additional tools for tracking improvement across sessions, finding statistically problematic corners (where am I plain bad?, where am I inconsistent?) or even training my muscle memory by tracing fast driver brake traces using my pedal.
Yay compiler: The other ongoing thing is a clean room reimplementation of Jon Blow's Jai. I've been curious about the language for years, but it's a closed beta and for some reason I've never felt about asking Jon to get into it. I'm not really a game dev so I wouldn't even know what to put in the request.
So now I have 100k+ lines of Rust that can compile a very significant subset of the publicly available Jai source code. I just used various LLMs to condense the public information about the language and come up with a dev plan and started chipping at it. Once I had something in a kind of working state I started with the Way to Jai big tutorial and make sure every example there compiles and works as intended, fixing errors or missing features one by one.
I mostly use Claude Code or Codex, but sometimes what I do is having them guide me into the new feature and doing the edits myself while they explain, so I get to know how things really work under the hood.
It's a silly pointless project, but for some reason I find very satisfying watching it compile the examples.
One of the main goal is to help with cryptocurrency-related home invasions. The XKCD "$5 wrench attacks" became a reality in France where I live. So it's another way to delay the access of personal funds, but it doesn't need to rely on third parties or multisig. You can just timelock a BIP39 passphrase for a duration of your choice.
It can also help with self-managed inheritance, or digital addictions.
http://localhost:8080/
I deliberately separated it from my public internet persona (which is connected to my real name) in the hopes that I could write about weird, woo-y, or controversial topics without worry. I've got a few articles half baked and have been having fun engaging with a different subset of the Substack crowd than my normal tech focus would show me.
Of course the stats show that the one article I did that touches on AI has done an order of magnitude better than anything else.
Anyway this is just kind of a weird sideline project, a sort of release valve for stuff that wouldn't fit in on my "professional" site, but it's been a fun thing to spend some time on.
Another thing that's cool is that I largely stopped _writing_ a few years back. I always enjoyed writing but of course as a dev most of my stuff had a technical/tutorial bent to it. Writing weird little "what do I think" essays has forced me to exercise a writing muscle I really hadn't stretched for a long time and I've enjoyed it.
There's only a handful of things up now, it's nothing special really. Link in my bio, if you see something you like I would love to hear from you!
An international calling app, for the poor people
Some months ago, I saw that very popup, and finally started working on something I've been wanted to do for a long-time, a spreadsheet application. It's cross-platform (looks and work identical across Windows, macOS and Linux), lightweight, and does what a spreadsheet application should be able to do, in the way you expect it, forever. As an extra benefit, I can finally open some spreadsheets that grown out of control (+100MB and growing) without having to go and make a cup of coffee while the spreadsheet loads.
I don't really have any concrete to share, I guess it'll be a Show HN eventually, but I thought it was funny it was brought up in a similar way in that article as was the motivation for me to build yet another spreadsheet application.
Now I've done my basic researching part, but I'm lack of the courage to dive into this topic. After all, it's a really hard work to it.
So I'm just, you know, scrolling the HN and trying to sharpen my brain and get back to the work.
https://github.com/storytold/artcraft
Before anyone asks, I am a filmmaker and have made films for fifteen years. I'm building tools to help steer AI image and video generation.
Here are a bunch of shorts made with the tool:
https://www.youtube.com/watch?v=HDdsKJl92H4
https://www.youtube.com/watch?v=ZThzgsdn1C0
https://www.youtube.com/watch?v=P9N_umJY_1s
https://www.youtube.com/watch?v=oqoCWdOwr2U
https://www.youtube.com/watch?v=tAAiiKteM-U
We have a lot of users, and it's picking up steam.
We're building BYOK/C and we're also building an OpenOpenRouter / OpenFal. After that's done, we're going to build an OpenRunPod.
Anyone into films, AI, or infra that likes working in Rust should reach out!
Six months ago, that would have been unrealistic, because we're heavily committed to the mongodb API and we make it part of our own API.
Starting in December though, Opus 4.6 made it perfectly realistic to pursue this with Claude Code as a series of personal weekend projects.
Now, despite not having any official resources on this until the last week or so, it should land in May.
This doesn't work for everything. It absolutely helps that the problem I'm solving is an "adapter pattern" problem: "make X talk like Y." And that we have a massive test suite, at multiple levels. That combination makes "here's the problem, go solve it, grind until the tests pass, don't bother me for a few hours" a realistic AI agent request.
But it's a little mind-blowing all the same. The hype around AI is so out of control, it can be easy to miss genuine "holy crap" moments.
Along the way I've written a fair bit about how to run Claude Code autonomously on your household server in a reasonably secure manner:
https://apostrophecms.com/blog/how-to-be-more-productive-wit...)
Also general Claude Code tips and thoughts on workflows that help and workflows that ultimately just speed your burnout:
https://apostrophecms.com/blog/claude-code-part-2-making-the...
I know, everybody's writing this stuff, but the desire to share is natural.
(Disclaimer: I'm part of the demographic AI was trained on. If I tried not to sound like a bot, I'd have to sound like... well, somebody else)
Podcast and RSS reader
Several other things, a CAD/CAM kernel with a Blender based frontend, a possibly novel strange attractor worth publishing, a git/CI host, an AI/LLM/VM cross platform workspace manager / IDE, shared multiplayer terminals in Minecraft and Godot
The goal is to make every recipe foolproof on the first try, similar to when you walk into a restaurant and just pick what you want to eat without thinking about the details. The goal is to have the same experience, just pick what you want to eat, with recipes that tells you exactly what to do with no magic involved.
Technically it is probably very different from other recipe apps. The database is a huge graph that captures the relations between ingredients and processes. Imagine 'raw potato'->'peeled potato'->'boiled potato'->'mashed potato'. It is all the same ingredients but different processing. The lines between the nodes define the process and the nodes are physical things. Recipes are defined as subsets of the graph. The graph can also wrap around into itself, which is apparently needed to properly define some European dishes in this system. The graph also has multiple layers to capture different relationships that are not process related.
Why was it designed it in this way? Because food/cooking is complex to define. This design is the only way I have found that can capture enough of these complex relationships that the computer can also 'understand' what is going on.
My favourite thing about this is that each recipe is strictly defined in the graph. If the recipe skips a step, or something is undefined, the computer knows that the recipe is incomplete. It won't ask you to do 10 things at the same time and then have something magically appear out of nowhere. It is like compile time checking but for recipes.
It also enables some other superpowers, for example: • Exclude meat part of the graph = vegetarian. Same thing works with allergies. • Include meat part of graph = only show me recipes that contain meat. • Recursive search: search for 'potato' and the computer will know that french fries are made from potato. It can therefore tell you that you could make the hamburger meal, but you will need to complete the french fries recipe first, which should take 60 minutes. • Adjustable recipe difficulty (experimental): It knows which steps can be done in parallell, and which can't based on how the nodes connect. A beginner can get a slower paced recipe with breathing room between steps, while someone more experienced can do a faster pace and do more things in parallell.
If I knew what it would take to build this, I would never have gotten started. I completely underestimated the complexity of the problem I was trying to solve. But here we are, and now it is basically done and working.
The website captures the key points from a non-technical point of view, and you can enter your email and get notified when it will launch in your country.
Anyway I'll look into it when in need to expand/replace my bosch system. Kudos to your team to make the work more reparable :-)
I think in this era of coding agents, more people feel empowered to build their own workflow automation. But for vast majority of non-technical folks, Claude Code or even Replit are not easy to use solutions. So I am taking inspiration from spreadsheets and using that as the primary UX to build a coding agent.
I started this after volunteering at my kid’s tournaments and seeing how fragmented things are: • registrations in Google Forms • payments via Venmo/Zelle • pairings in SwissSys/WinTD • communication across email and text
Chess67 aims to unify that: • coaches can sell lessons and manage scheduling and payments • clubs can run events and communicate with players • tournaments can handle registrations, with pairing and USCF submission in progress
Still early. The main challenge is not building features but matching existing workflows, especially Swiss pairings, which are more nuanced than they look.
Also, cleaning up a microscope 4-axis micro-positioning stage project control-loop.
Finding spare time to deal with a backlog of various other small projects. =3
2 products released (merge conflicts/codeowners) and now working on workflow automation. Basically trying to use Cloudflare Workers for a different paradigm of executing workflows instead of the traditional n8n VM.
It's called Inkfeed
https://www.geosystemsdev.com/products/hodlings/
In essence, it runs on your mobile device and stores all your data locally. It only connects to the freely available CoinGecko API (for latest prices) and GitHub (for reference and historical data). A background job updates GitHub ref data hourly. There's no login, no cloud, no ads, etc.
It's an LLM-webapp-builder, sure, but different from the rest! I have the LLM write python code when it needs to modify an HTML file for example (it'll use beautifulsoup; then I run the code: it parses the source into a data structure, modifies the data structure, and then outputs the resulting html).
It's also a marketplace where you can publish your llm-powered webapp, and earn $ on the token margins (I charge 2x token rates) when people use your site.
The idea: describe any problem in plain language (voice or text), and AI codifies it into a structured program with the right people, steps, timeline, and agents to get it done. It's a 5-step wizard: Define Problem → Codify Solution → Setup Program → Execute Program → Verify Outcome.
It runs across 50+ domains — codify.healthcare (EMR backend), codify.education (LMS backend), codify.finance, codify.careers (HRM backend), codify.law, plus 13 city domains (codify.nyc, codify.miami, codify.london, codify.tokyo, etc.). Each domain tailors the AI assessment and program output to that sector.
The platform is Project20x — think of it as the infrastructure layer. If Codify is the verb ("codify your healthcare problem into a care program"), Project20x is the operating system that runs it all: multi-tenant governance, AI agent orchestration, and domain-specific sys-cores for healthcare, education, city services, etc.
Every US federal agency and state-level department has a subdomain — ed.usa.project20x.com (Dept of Education), doj.usa.project20x.com, hhs.usa.project20x.com, etc. — with AI agents representing each agency's mandate. Same structure at the state level.
The political side: Project20x hosts policy management for both parties — dnc.project20x.com and rnc.project20x.com — where legislative intent gets codified into executable governance through a 10-step policy lifecycle. Right now I'm building out the multi-agent environment so agency agents can negotiate with each other, make deals, and send policy proposals up to the HITL (human-in-the-loop) politician for approval. Each elected official has a profile (e.g. https://project20x.com/u/donald-trump) where constituents can engage and where policy proposals land for review.
The name is a nod to structured policy frameworks, but the goal is nonpartisan infrastructure: democratically governed essential services delivered as AI-native social programs.
Stack: Nuxt 2/Vue 2 frontend, Laravel 10 API, Python/LangGraph agent orchestration, Flutter mobile app. Currently live across all domains.
https://project20x.com | https://codify.healthcare | https://codify.education | https://dnc.project20x.com | https://rnc.project20x.com etc...
the Indie Internet Index - https://iii.social
No file contents are accessed, only metadata, fully client-side API calls (browser to google API).
Direction - I’m trying to teach people how to do all the other stuff that you need to know, other than writing code, about delivering real products and not just a bunch of junk and slop that can’t be maintained
ShowHN: https://news.ycombinator.com/item?id=47721469
I’m also trying to make it really super simple so it’s week to week pricing, and have a discord community that grows out of it.
It’s literally just four two hour courses on Monday of each week and a demo day.
you walk through what you’re gonna do, how you’re gonna do it, how you’re gonna use your AI assistants to help you, where it can help you, and where it can’t help you, how to talk to it about teaching you instead of just doing it for you, and at the end of it you have something tangible to show for it.
There’s no subscription this is just straight up teaching product and project development that comes with a community and the community grows as much as it chooses to.
You can read the vision and roadmap on the site as well
Over the past few weeks, I have been building an AI coding tool in Go. The core loop is straightforward: accept a natural-language instruction, let the LLM interpret intent, then execute coding work through tools such as file read/write, code search, and terminal commands.
As of now, I haven't come across any agent coding tools written in Go, but I have always thought that Go is an excellent language and is very suitable for building any CLI tools.
Currently, I have added harness constraints to the agent by exposing hooks and implementing monitoring during the agent's working lifecycle. I think this will enable a clear division of responsibilities between the agent and the harness. The agent is the smallest execution core, while the harness acts as the execution agent for the agent and imposes constraints on its behavior.