It's an auction website for schools, charities etc without the exploitative transaction fees.
My wife and I are pretty heavily involved in our son's school PTA (parent teacher association) and have helped run school fundraising events for a few years, so we feel sort of like domain experts in this area :)
I started developing a city builder called Metropolis 1998 [1], but wanted to take the genre in new directions, building on top of what modern games have to offer:
- Watch what's happening inside buildings and design your own (optional)
- Change demand to a per-business level
- Bring the pixel art 3D render aesthetic back from the dead (e.g RollerCoaster Tycoon) [2]
I just updated my Steam page with some recent snapshots from my game. Im really happy with how the game is turning out!
[1] https://store.steampowered.com/app/2287430/Metropolis_1998/
[2] The art in my game is hand drawn though
> Both adults in a family will now own a car. This is required since there are not other transportation options, and sidewalks are optional.
Is this temporary or are you planning to release it like this? SimCity leaned into euclidean zoning (separate industrial/residential/commercial zones) and pocketable cars which needed no parking, and thus failed to properly showcase how ugly car-centric cities actually are. I’m sure they did it because it made for an easy gameplay loop/balancing but I’d hope we could come up with more realistic and interesting mechanics in 2026
Will you do a native Linux release, or has it been tested with Proton?
Also, just from watching the video and screenshots in the Steam page, it seems like a crazy amount of work. Are you doing everything by yourself?
Didn't realise you'd swapped to isometric, it's looking fabulous!
Did you roll your own engine, I know Godot has issues scaling past a certain number of simulations.
Pipeline so far has gone like this:
* Use the search engine's API to query a bunch of depravity
* Use qwen3.5 to label the search results and generate training data
* Try to use fasttext to create a fast model
* Get good results in theory but awful results in practice because it picks up weird features
* Yolo implement a small neural net using hand selected input features instead
* Train using fasttext training data
* Do a pretty good job
* for (;;) Apply the model to real a world link database and relabel positive findings with qwen to provide more training data
Currently this is where I'm at
Accuracy: 90.90%
True Positive: 1021
False Positive: 154
True Negative: 2816
False Negative: 230
Precision: 0.8689
Recall: 0.8161
F1: 0.8417
There's a lot of vague middle ground and many of the false positives are arguably just mislabeled.We’ve continued to get some paid customers and have exited beta last week, given everyone seemed to be quite satisfied and there hadn't been requests for changes, only some specific search providers.
Because of bots there isn’t a free trial easily available, but if you’re a human and you’d like to try it for a couple of days for free, reach out with your account number and we’ll set that up!
Thanks.
P.S.: Because people have asked before, our tech stack is intentionally very "boring" (as in, it generates and serves the HTML + bits of JS to enhance settings and such — search can be done without JS), using Deno in the backend (for easier TypeScript), PostgreSQL for the DB, and Docker for easier deploying.
One bit of feedback from me, take it or leave it, but the name doesn't feel appealing or memorable. What does it mean?
Also working on a language for embedded bare-metal devices with built-in cooperative multitasking.
A lot of embedded projects introduce an RTOS and then end up inheriting the complexity that comes with it. The idea here is to keep the mental model simple: every `[]` block runs independently and automatically yields after each logical line of code.
There is also an event/messaging system:
- Blocks can be triggered by events: `[>event params ...]`
- Blocks can wait for events internally
- Events can also be injected from interrupts
This makes it easy to model embedded systems as independent state machines while still monitoring device state.
Right now it’s mostly an interpreter written in Rust, but it can also emit C code. I’m still experimenting with syntax.
Example:
module WaterTank {
type Direction = UP|DOWN
let direction = UP
let current = 0
[>open_valve direction |> direction]
[>update level |> current]
[
for 0..30 |> iteration {
when direction {
UP -> !update level=current + 1 |> min(100)
DOWN -> !update level=current - 1 |> max(0)
} ~
%'{iteration} {current}'
}
]
[>update level |> when {
0..10 -> %'shallow'
11..15 -> %'good'
16.. -> %'too much!' then !open_valve direction=DOWN
}
]
}The content is hand picked from tiktok, Instagram, Facebook, Reddit and other AI generating platforms.
Honestly I don't know where I'm going with this, but I felt the urge to create it, so here it is.
I learned how to optimize serving assets on CloudFlare.
Feedback welcome.
EDIT: Hm, I switched tab, away to write this comment, now that I switched back, it showed me that I clicked correctly. So it seems, that sometimes it just has huge delay in accepting my choice?
Edit: I don't see slow traces in Sentry. No idea what caused this. Also, voting goes through redis and the dB load is low. Weird. I probably have to add gunicorn workers.
Edit2: Bumped gunicorn workers from 2 to 4. Should be fine now, under the current load. Again, thank you for reporting!
I dunno if/how this could be taught, but I feel like half the battle is critical thinking with an adversarial mindset towards media -- who would make this, why would they want to show me, do I see anything that makes this impossible, is it worth engaging with in the first place, can I fact check this.
I'm trying to gamify the training to make the experience more appealing.
I store a "proof URL" on the backend, but I don't know if it makes sense to serve it to the end user. Also, a Reddit discussion is not necessarily a proof one wants. A fingerprint would be better, but not all images are generated with Google. That's another problem to be solved.
It's SFW and localized to the most popular languages.
I have been working on it as side project for over two years and now, with funding from the EU for the next 2.5 years, I hope I can make of it a real product for everyone to use that can compete with the likes of Excel and Googl;e Sheets.
I can oly say, I am overly, off the Moon excited
edit: nm, rtfm, it was on the landing page: Horizon Europe programme
NLnet is just amazing and can keep you going if you are a student or have some extra sources of income
HORIZON is a huge grant but fairly hard to obtain. Generally related to reasearch grants in academia
As you see there is a huge component of sheer luck
[1]: https://nlnet.nl/project/IronCalc/ [2]: https://nextgraph.org/ [3]: https://elfaconsortium.eu/
Rendering is done by a go server. I wanted to learn go for quite a while and this is the perfect excuse.
const app = new App("com.apple.finder")
and then query for elements: const window = app.$({role: "window"})
const someButton = window.$(/* another query */)
and then do stuff with it: someButton.press()
and you can bind everything to very specific shortcuts like "press and hold cmd, then scroll mouse wheel up"Targeted towards music producers and AI (there's one collection of snippets that starts an MCP server and exposes some basic functionality) in the beginning.
Im also building https://www.keepfiled.com, a microsaas to save emails (or email attachments) to google drive
I almost forgot, I also built https://statphone.com - One emergency number that rings your whole family and breaks through DND.
I love building. I built all these for myself. unfortunately I suck at marketing so I barely have customers.
It's an addictive slot machine where I pull the lever and the dials spin as I hope for the sound of a jackpot. 999 out of 1000 winning models do so because of look-ahead bias, which makes them look great but are actually bad models. For example, one didn't convert the time zone from UTC to EST, so five hours of future knowledge got baked into the model. Another used `SELECT DISTINCT`, which chose a value at random during a 0–5 hour window — meaning 0–5 hours of future knowledge got baked in. That one was somehow related to Timescale hypertables.
Now I'm applying the VIX formula to TSLA options trades to see if I can take research papers about trading with VIX and apply them to TSLA.
Whatever the case, I've learned a lot about working with LLM agents and time-series data, and very little about actually trading equities and derivatives.
(I did 100% beat SPY with a train/out-of-sample test, though not by much. I'll likely share it here in a couple weeks. It automates trading on Robinhood, which is pretty cool.)
Best of luck. Super fun!
PS: Just a follow-up. There was a post here a few days ago about a research breakthrough where they literally just had the agent iterate on a single planning doc over and over. I think pushing chain of thought for SOTA foundational models is fertile ground. That may lead to an algorithmic breakthrough if you start with some solid academic research.
Ended up using ClickHouse - much smaller on disk, and much faster on all metrics.
GitHub: https://github.com/pranshuchittora/simvyn
Do give it a try, Thanks!
The problem was the ML dependencies. The backend uses BGE-small-en-v1.5 for embeddings and FAISS for vector search. Both are C++/Python. Using them from Go means CGO, which means a C toolchain in your build, platform-specific binaries, and the end of go get && go build.
So I wrote both from scratch in pure Go.
goformer (https://www.mikeayles.com/blog/goformer/) loads HuggingFace safetensors directly and runs BERT inference. No ONNX export step, no Python in the build pipeline. It produces embeddings that match the Python reference to cosine similarity > 0.9999. It's 10-50x slower than ONNX Runtime, but for my workload (embed one short query at search time, batch ingest at deploy time) 154ms per embedding is noise.
goformersearch (https://www.mikeayles.com/blog/goformersearch/) is the vector index. Brute-force and HNSW, same interface, swap with one line. I couldn't justify pulling in FAISS for the index sizes I'm dealing with (10k-50k vectors), and the pure Go HNSW searches in under 0.5ms at 50k vectors. Had to settle for HNSW over FAISS's IVF-PQ, but at this scale the recall tradeoff is fine.
The interesting bit was finding the crossover point where HNSW beats brute-force. At 384 dimensions it's around 2,400 vectors. Below that, just scan everything, the graph overhead isn't worth it. I wrote it up with benchmarks against FAISS for reference.
Together they're a zero-dependency semantic search stack. go get both libraries, download a model from HuggingFace, and you have embedding generation + vector search in a single static binary. No Python, no Docker, no CGO.
Is it better than ONNX/FAISS? Heck no. I just did it because I wanted to try out Go.
goformer: https://github.com/MichaelAyles/goformer
goformersearch: https://github.com/MichaelAyles/goformersearch
It started as a client problem, then something which I also experienced so decided to built it. It's just one small script and work seamlessly across platforms.
So FormBeep is designed to work with WhatsApp since most people prefer whatsapp notifications!
while word_count < x: write_next_chapter(outline, summary_so_far, previous_chapter_text)
It worked well enough that the novels were better than the median novel aimed at my son's age group, but I'm pretty sure we can do better.
There are web-based tools to help fiction authors to keep their stories straight: they use some data structures to store details about the world, the characters, the plot, the subplots etc., and how they change during each chapter.
I am trying to make an agent skill that has two parts:
- the SKILL.md that defines the goal (what criteria the novel must satisfy to be complete and good) and the general method
- some other md files that describe different roles (planner, author, editor, lore keeper, plot consistency checker etc.)
- a python file which the agent uses as the interface into the data structure (I want it to have a strong structure, and I don't like the idea of the agent just editing a bunch of json files directly)
For the first few iterations, I'm using cheap models (Gemini Flash ones) to generate the stories, and Opus 4.6 to provide feedback. Once I think the skill is described sufficiently well, I'll use a more powerful model for generation and read the resulting novel myself.
some other md files that describe different roles (planner, author, editor, lore keeper, plot consistency checker etc.)
- What are these meant to be exactly? are these sub agents in the workflow or am i completely misunderstanding?
One thing that I've been very happy with has been "org-people", now on MELPA, which allows contact-management within Emacs via org-mode blocks and properties. It works so well with the native facilities that it's a joy to work on.
I've been learning a lot of new things while I've been expanding it now it has a bigger audience (e.g. "cl-defstruct" was a pleasant surprise).
The front bump out leaks when we get driving rain. I installed some flashing but that wasn't enough, it's still leaking. So I'm working on that so I can close up the big hole in the ceiling some day.
The prior owners filled in the old coal chute with literal bags of cement sort of artistically placed in the hole in the brick foundation. So I'm trying to figure out what masonry tools and skills I'll need to close it up proper.
I'd like to build my kids a playhouse of some sort, sketching out some designs for that.
Very exciting on the playhouse. What kind of things will it have?
I'm expecting my first this year so have a ways to go before I get to work on that project
* adding local conversation memory for LLM
* improving Word spec compliance
* adding/extending table and image manipulation
* bug fixes!
Supports multiple-accounts (track as a family or even as an advisor), multi-currency, a custom sheet/calculator to operate on your accounts (calculate taxes etc) and much more. Most recently, we added support for benchmarking (create custom dashboards tracking nav and value chart of subsets of your portfolio) and US stocks, etfs etc.
We also write about like:
How fund performance explain part of returns, rest is explained by timing. And ways to tease those out: https://finbodhi.com/docs/blog/benchmark-scenarios
Or, understanding double entry account: https://finbodhi.com/docs/understanding-double-entry
I have a fairly novel approach to operating it, and in the case of one time theft prevention security through obscurity is actually a great approach. The assailant only has a short time to pull the car apart and solve the puzzle, couple that with genuine security techniques, a physical aspect, and it should be pretty foolproof.
It can still be towed away, etc, not much to be done there except brute force physical blocks. Most cars get stolen here to do crime in that night so it's not as common.
Free Math Sheets is a tool to generate math worksheet PDFs (and the answer keys if required). Currently it supports K-5 but I want to expand it to higher levels of math (Calculus, Physics, you name it!). You select a bunch of different options and then generate it. All in the front-end. No back-end or login in required. https://www.freemathsheets.com
If you are interested in helping out or forking it, here is the github repo github.com/sophikos/free-math-sheets
The paid project is Numerikos. I am going for something in between Khan Academy and Math Academy. I like the playfulness and answer input methods from Khan Academy (but it is linear, doesn't have a good way to go back and practice, etc.). I like Math Academy's algorithm (but it has multiple choice answers, yuck! and is easy to get stuck and doesn't have a good way to explore on your own). Currently Numerikos supports 4th and 5th grade math lessons and practice. The algorithm is based on mastery learning like Numerikos, but you can also see a list of all the skills and practice whatever you want. I am also working on a dashboard system where you can build your own daily/weekly practices for the skills you care about. Next up is 6th grade math and placement tests.
The idea is that _any_ user-facing tool, whether an app, worksheet generator or whatever, will need something like this for content, so I'm making this available for free and hoping for others to build on top of it.
I'm sticking to university-level stuff because I feel that school-level, especially math, is over-saturated already.
Technically, it is currently built as a React app, but that is mostly me sticking to tools that get out of my way. Generating PDFs or Anki files should be relatively straightforward.
Also been spending some time on my old side project https://infrabase.ai, an directory of AI infra related tools. Redesigned the landscape page (https://infrabase.ai/landscape), going through product submissions and content, optimizing a bit for seo/geo.
Environment variable checker - pretty niche.
What would make you use this? does this miss anything useful?
One of the issues I encountered initially was that the LLMs were repeating a small set of actions and never trying some of the more experimental actions. With a bit of prompt tweaking I was able to get them to branch out a bit, but it still feels like there's a lot of room for improvement on that front. I still haven't figured out how to instill a creative spark for exploration through my prompting skills.
It has been quite exciting to see how quickly a few simple rules can lead to emergent storytelling. One of the actions I added was the ability for the agents to pray to the creator of their world (i.e. me) along with the ability for me to respond in a separate cycle. The first prayer I received was from an agent that decided to wade into a river and kneel, just to offer a moment in stillness. Imagining it is still making me smile.
Unfortunately, I don't have access to enough compute to run a bigger experiment, but I think it would be really interesting to create lots of seed worlds / codebases which exist in a loop. With the twist being that after each cycle the agents can all suggest changes to their world. This would've previously been quite difficult, but I think it could be viable with current agentic programming capabilities. I wonder what a world with different LLM distributions would look like after a few iterations. What kind of worlds would Gemini, Claude, Grok, or ChatGPT create? And what if they're all put in the same world, which ones become the dominant force?
It just won an award! It was awarded Players' Choice out of 700 daily web games at the Playlin awards: https://playlin.io/news/announcing-the-2025-playlin-awards-w...
Right now around 3,500 people play every day which kind of blows my mind!
It's free, web-based, and responsive. It was inspired by board games and crosswords.
I've been troubleshooting some iOS performance issues, working on user accounts, and getting ready to launch player-submitted puzzles. It's slow going though because I have limited free time and making the puzzles is time consuming!
Here's an article with more info about the award: https://cogconnected.com/2026/03/tiled-words-crowned-the-pla...
Thank you so much for keeping it going!
- I applied to showcase the game at the Portland Retro Gaming Expo with the Portland Indie Game Squad. They accepted me so I was able to showcase it at the expo for a day. This got me some players right off the bat
- I shared it on HN, Reddit, Mastodon, etc.
- The website Thinky Games wrote an article about it
- The YouTube channel Cracking the Cryptic shared it which got a lot of new players. More recently a couple of other YouTubers (Timotab and Stro Solves) have been posting videos regularly
- I link to it from my blog, and this unrelated rant went semi-viral in web dev circles: https://paulmakeswebsites.com/writing/shadcn-radio-button/
- Winning the award gave me more visibility and players
I've also tried using things like Instagram and Discord but haven't had much luck there. I don't really get how those platforms work.
To be honest I'm not great at marketing. I've just been experimenting and seeing what works.
---
I would say the most important thing is the game itself:
- I've worked hard to gather feedback and incorporate it into the gameplay.
- I focus on keeping the puzzles fresh and striking the right difficulty level. (Challenging but something most people can do in 10 minutes.)
- I built a sharing feature that ~300 or so people use a day
I think all my marketing would have been useless if people didn't like the game and want to play again and share it with their friends.
AIOs are a black hole - we dont know when they appear and whats in it. so i creates a tool thats starts with GSC data and enriches it via AIO data
works good and the major finding by now
the best AIOs you can get are ..... none.
doesn't matter if you are in it or not - as soon as they show up the CTR to tour web-property goes down massively ~60% to 70%
the CTR on the AIOs are ~0%
I recently converted a bunch of stuff to be client side instead of server side (turns out running a real-time MMORPG server is expensive) so there's a new round of bugs I'm still resolving, but it's still fun to play:
Primarily to use in conjunction with OpenVPN. Like secretive or /usr/lib/ssh-keychain.dylib[2], but not just for SSH.
The idea is: you join a meeting, hit start on the app, minimize, and go do actual work (or go make a coffee). When someone says your name or any keyword(s) you set, you get a native macOS notification with enough context to jump back in without looking lost. It uses whisper and is 100% local and doesnt leave traces, also very OE friendly.
Would love to hear what you think, especially if you're drowning in meetings too.
It’s like netflix for language, where users can select/create their personal bilangual stories.
I had quite a lot of feedback from HN, friends, random people on the internet and trying to solve the common pain points and find my way around to make it geniunely useful.
- Most people said it’s hard to come up with a story, so I added url grounding. Also added buttons (including HN :)) so people can just click click and get their stories at their level with their interests.
- Made sure people can generate stories without ever signing up
- Each word is highlighted while being read, and the meanings can be checked with a tap. I also added an option for users to read the sentence for being checked how good their pronounciation is.
- Benchmarked 7 different models to get the fastest & highest quality story generation (it’s gemini now) and it’s insanely fast. I might share more about it on the webpage because I am an engineer and I enjoy this stuff lol.
- Added CSV import in Use my words so Anki users can just import their words to study.
- Also people can download their stories as pdf so they can send it to their kindles.
- I am working on a ChatGPT app, so people can just say “@DuoBook give me a Dutch/English story on latest Iranian events” within ChatGPT, but I am a bit afraid that it might be costly lol.
I started small as a toy project, but gradually implemented full support for proper block context, flexbox layout, CSS variables, tables, etc. to the point where I have almost full support of all major CSS features (even math functions like calc(), min(), max()).
I'm cleaning up the code right now and will upload it later today or maybe tomorrow here: https://github.com/PureGoPDF
Type "cheapest flight from London to Tokyo, flexible on dates in April" and it returns live results with real pricing. I compared a few against Google Flights and they matched. Not mocked data.
The part I found interesting: it runs on a dedicated VM so it keeps context across the conversation. If you say "actually make that business class" or "what about flying into Osaka instead" it knows what you were looking at. Most chat-based search tools lose that between messages.
I didn't build it from scratch — it's a pre-built app in the SuperNinja App Store that I deployed and have been extending. The deploy itself took about 60 seconds. The extending part is what I've been spending time on: describing changes in plain text and watching them go live without touching a repo.
Still figuring out what the right UX is for flexible-date search. Curious if anyone has opinions on that.
Interesting findings include Mistral doing better than Gemini 3 Pro in certain usescases, cross-LLM works better than one LLM to another, oh and - the cost all of of this. So, so expensive.
I also intend to dig into how to integrate Emacs with tools such as yt-dlp and patreon-dl to grab Latin-language audio content from the Internet, transcode the audio with ffmpeg, load it into the LLM's context window, and send it off for transcription. If the essay isn't already too long, I'll demonstrate how to gather forced-alignment data using local models such as wav2vec2-latin so I can play audio snippets of Latin texts directly from a transcription buffer in Emacs. Lastly, I want show how to leverage Gemini to automatically create multimedia flash cards in Org mode using the anki-editor Emacs minor mode for sentence mining.
Why? Many yarncrafters painstakingly build spreadsheets, or try to bend existing general purpose pixel editors to their will. It's time consuming & frustrating.
Along the way, I've solved a bunch of problems:
- Automatic decreases (shapes the hat) / overstitching markers (shows when multiple colors are used in the same row)
- Parameterized designs, like waves, trees, geometric shapes. No more manually moving an object by a couple of pixels, it's a simple click & drag.
- Color palette merging (can't delete a color if you already use it in a pattern!)
- Export to PDF (so you can print it or stick it on a tablet)
- Repeat previews (visualize the pattern as it repeats horizontally)
The core feature that makes this more useful than most general purpose editors is that the canvas is continuous.If you drag a shape near the right edge of the canvas, you'll see it "wrapping around" onto the right edge.
This reflects the 3D reality of a hat!
I got it all done in probably an hour or two. But done in 10-15 min blocks over many days.
Today working on adding chat history search (FTS5) and OpenRouter Nano Banana 2 support.
https://breaka.club/blog/why-were-building-clubs-for-kids
The recent Netflix Games edition of Overcooked with K-Pop Demon Hunters is cool, but not nearly as cool as kids coding and playing their way through Overcooked levels in our custom educational mod for Overcooked:
I'm also maintaining GodotJS, strongly typed TypeScript bindings for Godot, which is used to build the Breaka Club RPG (see first link):
https://github.com/godotjs/GodotJS
And last week I also put together the first release of MoonSharp in ~10 years; Lua runtime for Unity. That's not for Breaka Club though, I also consult for Berserk Games on Tabletop Simulator:
Basically OpenClaw but with investing dashboards for my portfolio, additional tools specifically for investing, and exploring an AI-Human collaboration on researching economics (check the 'community' tab).
The data models are all in markdown and Excel so that there's no lockin and you can manually edit positions, personalities, etc.
This comes from frustration around most investing tools basically scraping your personal data + forcing you to lock into subscriptions. I think it's now possible to just vibe code most of what one needs, aside form raw data subscriptions.
It's all open source, too: https://github.com/wgryc/athena-os
Here's a link to the jam if anyone else is interested, and I recommend joining the Discord server too because the organizers and participants are really great and fun to hang around! - https://itch.io/jam/flame-game-jam-2026
Each game adds more building blocks to the editor: multiplayer, event systems, NPC behaviors, pathfinding, etc. I build a system once, and then anyone using the editor can use it in a click.
Since my last month, I shipped the asset marketplace and the LLM builder. Artists can now upload tilesets and characters, and unlike itch.io, assets drop directly into the editor. You can preview how they'll actually look in-game before using them [1].
An other problem I kept running into: even with a no-code editor, users don't know where to start. So now I'm extending it with a coding agent. Describe the game you want, and it assembles it — pulling assets from the marketplace, wiring up the event system, and using all the building blocks I've spent the past year extracting. Multiplayer, mobile controls, pathfinding, NPC behaviors — the agent doesn't build any of it, just reaches for what's already there.
Once the LLM assembles it, users will have a game ready to work on, and will still be able jump into the editor and tweak everything [2]. Here's an example of what it can already make [3] (after a lot of prompting), and the goal is to reach games like this one I built with the manual editor[4].
Hoping to release the AI mode in a week or two. The manual editor is live at https://craftmygame.com in the meantime.
[1] https://craftmygame.com/asset/mossy-cavern-JdYWai1
[2] https://youtu.be/6I0-eTmoHwQ
I've wanted this for a long time, so I finally started building it. I've had a lot fun!
- Graph-based signal flow: Products become nodes, connections are edges inferred from port compatibility (digital, analog, phono, speaker-level domains)
- Port profile system: Standardized port definitions (direction, domain, connector, channel mode) enable automatic connection inference
- Rule engine: Pluggable rules check completeness, power matching, phono stage requirements, DAC needs, and more
https://housepricedashboard.co.uk - shows a visualisation of house prices in England and Wales since the 90s, with filters for house types, real vs nominal, and change views over time
https://councilatlas.co.uk - similar structure to the above, but focusing on local council datasets. The idea is to make it easier to compare your local council's performance against the rest of the country.
The problem: every agent (Cline, Aider, Codex, Claude Code) has unrestricted access to your filesystem, shell, and network. When they process untrusted content — a cloned repo, a dependency README — they’re prompt injection vectors with full machine access. No existing tool evaluates what the agent actually does at the syscall level.
grith wraps any CLI agent without modification. OS-level interception captures every file open, network call, and process spawn, then runs it through 17 independent security filters in parallel across three phases (~15ms total). Composite score routes each call: auto-allow, auto-deny, or queue for async review. Most will auto approve - which eliminates approval fatigue.
Also does per-session cost tracking and audit trails as a side effect of intercepting everything.
It pulls a list of birds reported on eBird in your county in the last 2 weeks and you ask preselected questions like the the color or size to whittle down the possibilities. I also made a matching game that uses the same list and you have to match the name to a picture of the bird. I set it up for California for now. I wanted to get more comfortable with SQL and APIs.
Feedback welcome.
It's a bad ripoff of the much, much more fun metazooa (https://metazooa.com/play/game). I kike it but it gets real annoying when your down to 1 of 10 bats or something. I've been using it to read and edit Wikipedia articles for undeveloped pages.
Over the last year I've been hacking on Table Slayer [0] a web tool for projecting DnD maps on purpose built TV-in-table setups. Right now I'm working on making hardware that supports large format touch displays.
Since I also play boardgames, this past month I threw together Counter Slayer [1], which helps you generate STLs for box game inserts.
Both projects are open source and available on GitHub. I've had fun building software for hobbies that are mostly tactile.
(If you're a local reading this and enjoy DnD w/ roleplay and acting, email's in my profile)
The stock firmware is horrible but the community has this firmware called CrossPoint. I wanted to be able to upload, manage files etc. from my iPhone on the go and also send over web articles. So I build this app CrossPoint Sync https://crosspointsync.com to do just that.
I've already published it on App Store and pending publishing on Android. The community is niche and has also been using the app, so its been fun building for my use and in turn also getting good feedback from community.
If you are using the Xteink and CrossPoint firmware, then give the app a try.
iOS App Store: https://apps.apple.com/app/crosspoint-sync/id6758985427
Android Beta: https://crosspointsync.com/android/join-beta
https://milliondollarchat.com a reimagining of the million dollar homepage for the AI age. Not useful, but fun. A free to use chatbot that anyone can influence by adding to the context. The chatbot's "thoughts" are streamed to all visitors.
KPT is a language app specifically targeted at explainable verb conjugation for highly inflected/agglutinative languages. Currently works for Finnish, Ukrainian, Welsh, Turkish and Tamil.
These are really hard languages to learn for most speakers of European languages, particularly English - we're not used to complex verb conjugations, they're hard to memorise and the rules often feel quite arbitrary. Every other conjugation practice app just tells you right/wrong with no explanation, which doesn't really help you learn when there are literally hundreds of rules to get right.
The interesting part was using an LLM to create a complete machine-executable set of conjugation rules, which are optimized for human explainability, and an engine to diagnose which rule is at fault when you get it wrong. There's several hundred rules needed for each language in order to cover all exceptions.
NB as a bonus it also works fully offline because my best practice hours are when I'm travelling and have poor connectivity.
Making improvements on this tetris meets block puzzle game
Provisional patents went in recently so don't mind broadcasting to a wider audience beyond my poor, unknowing, testers
You can see it working here: https://www.youtube.com/watch?v=G5Xup3kB1D0 and I literally put up a holding page for some media related surges (as it's all self hosted etc and I didn't want to mix my functional stuff with my spikey stuff) here ( name to be worked on, but "NUTS" is the current one) : https://buttonsqueeze.com
My favorite features are: - custom layout and drag and drop to change window - auto resume to last working session on app starting - notifications - copy and paste images directly to Claude Code/Codex/Gemini CLI - file tree with right click to insert file path to the session directly
OH and it works on both Windows and MacOS! Fully open source too!
I have worked with data for a while. I feel like our tools could be much better when it comes to "flow". I want an experience where you don't need to alt+tab to slack/images/another query. What if we put it all on a canvas? That's what Kavla is all about!
Since last month I've done a lot of improvements to the editor to make the "flow" better.
I've also read up on HMAC, Nonces and fun encryption stuff to create read only boards.
Here's one where I look at stack overflow survey for databases: https://app.kavla.dev/v/mqhg54o319doya4.67dbfee1ccd6caf638d3...
Snowflake users apparently make the most money!
PS - The results are entirely obvious.
- golang based architecture
- information is dynamically mapped into one central directed knowledge graph
- default multithreading
- utilizes existing tools (such as nmap/nuclei/katana/wfuzz/....) instead of reinventing the wheel
- architecture is (tldr) a self supervising logic in which every worker is also a scheduler that based on delta causality uses cartesian fanout and graph overlay mapping including local only witness nodes to dispatch new "jobs" without having a central scheduler or the necessity to scan a central total job queue to prevent duplicate executions.
In this architecture every "action" that can be executed defines an input structure necessary. If the previously mentioned mechanic identifies a possible job execution it will create a job input payload which will automatically be picked up by a worker an executed. Therefor every action is a self containing logic. This results in a organically growing knowledge graph without defining a full execution flow. It is very easy to extend.
I worked on this for the past ~10 years (private time). The sad truth tho is, while this project was initially planned to be open sourced - after i not to long ago for quite some bugs consulted a lawyer, i basically was presented with the fact that if i would publish it i could get sued due to germany's hacker and software reliability laws. So for now its only trapped on my disk and maybe will never see daylight.
Im right now working on a blog article (thats why i even mention it) about the whole thing with quite more detailed description and will also contain some example visual data. Maybe will post it on hackernews will see.
PS:The tool does not need llm/nn.
I posted about it recently on HN (https://news.ycombinator.com/item?id=47199062):
It is at a fairly early stage of development, so it's quite rough around the edges. It is developed and hosted in EU.
I have started developing it as a slim wrapper around Git to serve my own code, but it grew to such extent that I decided to give it a try and offer it as a service. It doesn't have much at the moment, but it already has basic pull requests. Accessibility is high priority.
It will be a paid service, (free for contributors) but since it's an early start, an "early adopter discount" is applied – 6 months for free. No card details required.
I would be happy if you give it a try and let me know what do you think, and perhaps share what you lack in existing solutions that you would like to see implemented here.
- https://github.com/rumca-js/Internet-Places-Database - database of domains and youtube channels
- https://github.com/rumca-js/crawler-buddy - web crawling / web scraping tool
- https://github.com/rumca-js/webtoolkit - web crawling toolki
- https://github.com/rumca-js/Internet-feeds - feeds databse
- https://github.com/rumca-js/Django-link-archive - RSS reader
Group chat photobooks. Automatic layouts, no editor/app, unlimited free previews. Build a hardcover (up to 1000 image) and ship it in minutes.
Wanted a physical souvenir for everyone in my long running signal chat but didn’t want to spend hours curating in editors.
The result is an experiment called fesh. It works strictly as a deterministic pre-processor pipeline wrapping LZMA (xz). The AI kept identifying "structural entropy boundaries" and instructed me to extract near-branches, normalize jump tables, rewrite .eh_frame DWARF pointers to absolute image bases, delta-encode ELF .rela structs with ZigZag mappings, and force column transpositions before compressing them in separated LZMA channels.
Surprisingly, it actually works. The CI strictly verifies that compression is perfectly reversible (bit-for-bit identity match) across 103 Alpine Linux x86_64 packages. According to the benchmarks, it consistently produces smaller payloads than xz -9e --x86 (XZ BCJ), ZSTD, and Brotli across the board—averaging around 6% smaller than maximum XZ BCJ limits.
I honestly have no idea how much of this is genuinely novel versus standard practices in extreme binary packing (like Crinkler/UPX).
Repo: https://github.com/mohsen1/fesh
For those who know this stuff:
Does this architecture have any actual merits for standard distribution formats, or is this just overfitting the LZMA dictionary to Alpine's compiler outputs? I'd love to hear from people who actually understand compression math.
I have no illusions that this is actually something in capable of building to an actual release-able state but it’s fun to tinker with.
I wrote a CLI utility last year to control my SoundBlasterx G6 DAC (can only control LED colour and EQ bands) without needing to use Creative's windows only program (I am mostly a Mac + occasional Linux) user.
Recently downloaded Qwen3-coder-next 80b model and been vibing with it to introduce Qt6 and write a dead simple (aka ugly) crossplatform GUI to it so that other people can use it on their Macs and Linux machines. Letting a LLM wreak havoc on your project feels bad, I constantly have to reign it in and rollback the repo once it starts looping due to writing something that doesn't compile, making it going back and forth between doing and undoing changes.
A lot of existing databases are storage first, with everything else built around them. I have been exploring what it looks like if the database is closer to the application runtime itself, where state is live, queryable, and easier to reason about directly.
One thing I am prototyping right now is database-native tests.
Basically: what if integration tests were a database primitive?
CREATE TEST test::insert { INSERT test::users [{ id: 99, name: "Ghost" }]; FROM test::users | FILTER id == 99 | ASSERT { name == "Ghost" }; };
So not a wrapper, not a framework, not an external test runner.
A real test object inside the database.
The idea is that you could run these before schema changes, and make stored procedures or other database logic much easier to test without leaving the database model.
Still early, but it feels like one of those things that should just exist, especially for databases built around live application state.
Give it a try!
[0]: https://aibenchy.com
Now I've written quite a few posts (and given talks), I thought of writing a book. Just wrote two chapters. The draft lives here: https://www.jjude.com/books/hs/
https://greenmtnboy.github.io/sf_tree_reporting/#/
For all the places it's bad at, AI has been fantastic for making targeted data experiences a lot more accessible to build (see MotherDuck and dives, etc), as long as you can keep the actual data access grounded. Years of tableau/looker have atrophied my creativity a bit, trying to get back to having more fun.
This is an app that’s been bouncing around in my head for over a decade but finally got it working well enough for my own purposes about a year and a half ago.
I was thinking of a google maps kind of "here you are, here's your walking path of interesting trees" potentially, or something else that could tie the overview to the street experience - on the backlog!
I’m a Senior Full Stack Engineer with over 8 years of experience building and scaling production systems using Node.js, TypeScript, React, and Python. I’ve worked in remote, product-focused environments where I’ve led architectural improvements, including migrating a monolithic system to microservices, reducing deployment time by around 50% and improving scalability and reliability.
I’m comfortable owning features end-to-end — from system design and API development to deployment, performance optimization, and production support. I’ve also implemented CI/CD pipelines, improved database performance (PostgreSQL), and contributed to cloud-native infrastructure on AWS using Docker and Kubernetes. In addition, I’ve worked on AI-driven workflows and LLM integrations for modern product capabilities.
I’m currently exploring new remote opportunities and would love to connect if you’re building or scaling a product where strong backend architecture, clean execution, and ownership matter.
If it makes sense, I’d be happy to schedule a short conversation. Thank you.
The problem: every agent framework bolts together a vector DB for recall, a KV store for state, maybe a graph DB for relationships, and then hopes the duct tape holds. You get one retrieval path (similarity search), no decay, no consolidation, and the agent forgets everything the moment context gets trimmed.
HEBBS replaces that stack with a single embedded binary (RocksDB underneath, ONNX for local embeddings). Nine operations in three groups: write (remember, revise, forget), read (recall, prime, subscribe), and consolidate (reflect, insights, policy). The interesting part is four recall strategies — similarity, temporal, causal, and analogical — instead of just "nearest vector."
Some technical decisions I'm happy with:
- No network calls on the hot path. Embeddings run locally via ONNX; LLM calls only happen in the background reflect pipeline.
- recall at 2ms p50 / 8ms p99 at 10M memories on a 2 vCPU instance.
- Append-only event model for memories — sync is conflict-free, and forget is itself a logged event (useful for GDPR).
- Lineage tracking: insights link back to source memories, revisions track predecessors.
SDKs for Python, TypeScript, and Rust. CLI with a REPL. gRPC + REST.
There's a reference demo — an AI sales agent that uses HEBBS for multi-session memory, objection handling recall, and background consolidation of conversation patterns.
Still early. The part I'm wrestling with now is tuning the reflect pipeline — figuring out when and how aggressively to consolidate episodic memories into semantic insights without losing useful detail. Curious if anyone working on agent memory has opinions on that tradeoff, or if you've found other approaches that work.
The system allows users to submit a JSON payload containing geocoordinates and mission requirements (e.g., capture_type: "4K_video" | "IR_photo"), the backend then handles the fleet logistics, selecting the optimal VTOL units from distributed sub-stations based on battery state-of-charge and proximity.
POST a list of URLs, get back a signed download link. Your server never touches the file data.
Built it because I kept writing the same server-side ZIP streaming boilerplate on every project — memory spikes, connection timeouts, ZIP64 edge cases. Now it's one API call.
Zero egress fees, ZIP64 support (4 GB+, up to 5,000 files/job), signed expiring URLs. Free tier included, no credit card needed.
It is a pretty fun project
Sounds like a fun project -- like a more interactive version of Football Manager.
I first used DynamoDB 8 years ago and have been designing single-table schemas heavily since. For me, the best way to create drafts was always pen and paper (and then excel/confluence tables), but in reality it's a process (based on The DynamoDB Book) that can be automated to an extent.
Decided to build an app while on paternity leave. You define entities and access patterns, create (or get suggested) key and GSI design, and generate code for access patterns (TypeScript and Python), infrastructure (CDK, CloudFormation, Terraform), and documentation you can share with stakeholders.
There's more I want to build beyond the MVP - things around understanding and validating designs that you can't get from a chatbot - but for now focusing on the core.
If anyone wants to try it out, sign up for the waitlist on the landing page. MVP should be ready in the next few weeks.
https://www.inclusivecolors.com/
The current web tool lets you export to CSS, Tailwind and Figma, and uses HSLuv for the color picker. HSL color pickers that most design tools like Figma use have the very counterintuitive property that the hue and saturation sliders will change the lightness of a color (which then impacts its WCAG contrast), which HSLuv fixes to make it much easier to find accessible color combinations.
I'm working on a Figma plugin version so you can preview colors directly on a Figma design as you make changes. It's tricky shrinking the UI to work inside a small plugin window!
As many here, I've found that a single text file is all that I really need, but found that it makes it difficult to keep track of a variety of things. I was also trying to use the file as a simple project tracker, adding some tags like [BUG-N], and updating them by hand. Eventually, it became difficult to track the progress of things, since I had to jump around the file to look for updates.. or use grep.
I condensed the idea to just that - a very simple tool which manages "trackers", and has a simple filtering built in to "trace" the updates. I've been using it, since I've added the BE, and dogfooding it a bunch. Would love for fellow note takers to take a look. It's not perfect, but I'm keeping it around for myself :)
It's like a carfax but for your home, although the intention is more to create an interesting historical narrative that inspires people to care about the history of their home rather than as a tool for inspecting home issues before buying.
My target customer is realtors who want to inspire buyers to take on historic homes that may need a lot of work. Also home owners themselves of course.
If this became the norm, somehow, it would be a really helpful tool for both buyers and sellers.
I have built npm for LLM models, which lets you install & run 10,000+ open sourced large language models within seconds. The idea is to make models installable like packages in your code:
llmpm install llama3
llmpm run llama3
You can also package large language models together with your code so projects can reproduce the same setup easily.
Follow the docs here: https://www.llmpm.co/docs
Pro tip for your use case: Checkout the `llmpm serve` section
So I built YouLingua (https://youlingua.world). Paste any YouTube video and get a word-by-word interactive transcript. Click a word to save it with the exact video moment — "muddy puddles" isn't a flashcard, it's Peppa jumping in one. Saved words then power mini-games: a space shooter, hex puzzles, TikTok-style review shorts...
Browser-based, no install. Login with a Web3 wallet — no grand reason yet, just something I'm interested in. Dream is to eventually make it fully decentralized so you truly own your learning data.
Still early, but my son now asks to "play the word game." That feels like a win.
Warning: Do not lick on the link.
1. Live Kaiwa — real-time Japanese conversation support
I live in a rural farming neighborhood in Japan. Day-to-day Japanese is fine for me, but neighborhood meetings were a completely different level. Fast speech, local dialect, references to people and events from decades ago. I'd leave feeling like I understood maybe 5% of what happened.
So I built a tool for myself to help follow those conversations.
Live Kaiwa transcribes Japanese speech in real time and gives English translations, summaries, and suggested responses while the conversation is happening.
Some technical details:
* Browser microphone streams audio via WebRTC to a server with Kotoba Whisper * Multi-pass transcription: quick first pass, then higher-accuracy re-transcription that replaces earlier text * Each batch of transcript is sent to an LLM that generates translations, summary bullets, and response suggestions * Everything is streamed back to the UI live * Session data stays entirely in the browser — nothing stored server-side
---
2. Cooperation Cube — a board game that rotates the playing field
Years ago I built a physical board game where players place sticks into a wooden cube to complete patterns on the faces.
The twist: the cube rotates 90° every round, so patterns you're building suddenly become part of someone else's board. It creates a mix of strategy, memory, and semi-cooperative play.
I recently built a digital version.
Game mechanics:
* 4 players drafting cards and placing colored sticks on cube faces * The cube rotates every 4 actions * Players must remember what exists on other faces * Cooperation cards allow two players to coordinate for shared bonuses * Game ends when someone runs out of short sticks
---
Both projects mostly started as things I wanted to exist for myself. Curious what people think.
Over the past weeks, we consistently get 5-6 submissions per week. The newsletter and number of visitors are growing.
I’ve come to treat this as a pet project but realized that for indie devs who get very little marketing attention, being featured in the newsletter, top of the daily list, etc. can be another burst of users.
Think OpenClaw, but durable, with long-term state, and enterprise-ready. We've been using it internally to build agents for a while now and have decided to open-source it.
It's a free USCIS form-filling web-app(no Adobe required). USCIS forms still use XFA PDFs, which don’t let you edit in most browsers. Even with Adobe, fields break, and getting the signature is hard.
So I converted the PDF form into modern, browser-friendly web forms - and kept every field 1:1 with the original. You fill the form, submit it, and get the official USCIS PDF filled.
I found out SimpleCitizen(YC S16) offers a DIY plan for $529 [2]
So, a free (and local-only) version might be a good alternative
[1] https://fillvisa.com/demo [2] https://www.simplecitizen.com/pricing/
I have also taken an interest in learning distributed paradigms like MPI and am using it on my own cluster of rPis
Open-source plugins for Ghidra, Binary Ninja, and IDA Pro that bring LLM reasoning, autonomous agents, and semantic knowledge graphs directly into your analysis workflow.
Coming soon: A supporting online service. The VirusTotal for reverse engineering. A cloud-native symbol store and knowledge graph service designed for the reverse engineering community.
- Submit files for automated reverse engineering and analysis
- Query shared symbols, types, and semantic knowledge
- Accelerate analysis with community-contributed intelligence
- Versioned, deduplicated symbols with multi-contributor collaboration
We're constantly pulling info from official sources, and using AI to group and summarize into stories, and continue to share reporting from trusted, vetted journalists.
The result is news with the speed and breadth of getting updates straight from the source, and the perspective and context that reporting provides.
Still ramping up, but I'd love to hear feedback:
- VR version of Surface Browser (3d internet browser): https://boxc.net/surfacebrowser.html
- Crowd Strike: faster self-driving: an exhibition where the visitors help autonomous drones target a different visitor each minute with lasers
and also Wingman: a dating app secretary (privacy focus, runs locally on your computer for any dating app that has a web site. It tells you if favourites have messaged you): https://boxc.net/wingman_app.png I'll open source this one if interest.
I don't think what I am doing is really original, but it's shaping nicely.
I am working on:
- feature folders (one folder per feature, with changelog, issues, summaries etc)
- coworkers (cli-agents, with session management)
- agents intra-response messaging
In general the goal is forcing Claude to behave, which is quite ambitious :).
- UI for sandbox-exec to protect filesystem - Network sandbox per domain - Secrets filter via gitleaks - Vertical tabs option
It's highly customizable. You generate native macOS app wrappers for each terminal app, each with its own rules and customizations.
I needed a way to use and push my own artifacts in Meson projects. WrapDB is fine for upstream deps, but I wanted to publish my packages and depend on them with proper versioning and a lockfile, without hand-editing wrap files.
Collider builds on Meson’s wrap system: you declare deps in collider.json, run collider lock for reproducible installs, and push your projects as wraps to a local or HTTP repo. It’s compatible with WrapDB, so existing workflows still work: you just get a clear way to use and push your own stuff. Apache-2.0.
To me good is - Pre-determined lists of words - Audio examples - Sentence examples - Native app with offline support
most importantly: - No business model that requires a subscription
I'm trying to see it more as writing a text-book, than starting a business
If you operate in the EU and want to avoid heavy fines, this is for you. Once integrated, it allows users to report legal content issues directly to you, which you can then manage via a dedicated dashboard following official EU procedures. Without such a system, users are much more likely to file complaints through official state or EU channels, which can trigger investigations.
Today, if you search for "what size should I get in Nike Air Max 90" you'll find size charts. We have it, and for 200+ brands across 70+ retailers. When users tell us which shoes they own and what size fits them we’re slowly building crowdsourced fit recommendations which are personal and more accurate compared to size charts.
We're two coders who've built an almost fully autonomous platform. AI agents build, debug and deploy crawlers on their own. We went from 4 crawlers to 280+ in about a month, and the whole thing runs on a home server. When new shoes are discovered, the platform publishes new pages with relevant info automatically. Agents get access to platform metrics and SEO data via custom MCPs to identify the right opportunities on their own. Currently at about 3000 MAU and about 100 size recommendations/day.
Example: https://www.getsize.shoes/en/shoes/nike-air-jordan-1-low-se-...
This is the kind of project that creates something from as little as possible, where the only things you need to get started are a very basic RISC-V assembler and a computer or emulator to run it on.
I don't have anything interesting to show yet because I just started yesterday, but one day I will show you.
It started because my wife watches Chinese dramas and new episodes never have subtitles for our language. Turns out thousands of people have the same problem — Arabic speakers watching anime, Russian speakers following Turkish series, Persian speakers catching up on K-dramas.
Supports 40+ languages, works with any video link or direct file upload. There's also a Mini App inside Telegram for a more visual experience.
If any of you have already figured out a tool/workflow for this, I'd love to learn from your experience.
I'm finding language auto-detection to be a bit wonky (for example, it repeatedly identified Ladykracher audio as English instead of German). I ended up having to force a language instead. The only show in my library where this approach doesn't work is Parlement[1], but I can live with that.
On the whole this is looking quite promising. Thanks for the idea.
We've found most early-stage startups ignore social media until after a launch. Things like “$0 spent on ads” sound cool, but they don’t help if no one knows your product exists.
I'm building Appents to provide a done-for-you social media solution for startups.
Would love feedback: https://appents.com/
https://feedbun.com - a browser extension that decodes food labels and recipes on any website for healthy eating, with science-backed research summaries and recommendations.
https://rizz.farm - a lead gen tool for Reddit that focuses on helping instead of selling, to build long-lasting organic traffic.
https://persumi.com - a blogging platform that turns articles into audio, and to showcase your different interests or "personas".
Put One In for Johnny Minn (https://store.steampowered.com/app/3802120/Put_One_In_for_Jo...) - A small soccer game all about scoring nice goals. While I don’t expect it to do well, I’m very happy with how it came out, and it’s the first game I’ve made that I’ll release on Steam! Comes out on Thursday (March 12th).
HeartRoutine (https://www.heartroutine.com/) - I built this a few months ago to help me stay on top of my heart health. I enter my numbers on the (offline) app, and then configure my goals (like “lower Apo B through diet and exercise”), and then the server emails me every morning asking me what I ate yesterday, how I exercised, etc. The goal is to stay on track, and to be able to bring a cardiologist a very detailed report.
Main gig: Trusted agents. We just shipped hardware based signing to web bot auth protocol.
It’s been available as a free tool for years, growing to over 45k active installs. I just rolled out the Pro extension to offer more advanced features, and the early traction has exceeded my expectations. If you're running e-commerce in Europe, this is a must-have for staying compliant with EU law.
I've given myself 6 months
It's a bit scary basically 180ing like this but I figure if I don't try it now I never will
I've already started prototyping various ideas, and to be honest just sitting down and spending time doing this has been really quite lovely
One thing I'm finding fun is slowly unearthing what I actually find interesting
I started with messing around in minecraft and tinkering with rimworld-like game ideas, but I'm slowly moving away from them as I've been tinkering more and more
Don't get me wrong, I do want to revisit them at some point in the future, but I do find myself circling more around narrative, simulations and zachlikes
It's a bit of an odd mix and in some ways they look like paradox style games, but I'm well aware that taking one of those behemoths on is going to be a bit silly, so I'm trying to slim down until I get to a kernel that I actually find enjoyable tinkering with
A toy if you will
Currently I'm trying to work out if there's anything interesting in custom unit design, basically unpicking how games like rollercoaster tycoon's coaster design maps to stats like excitement ratings and seeing how that might mix with old school point buy systems
It feels like it might be small enough to be a good toy and I'm having fun tinkering with it, but I have no idea whether other people will xD
It might honestly be too niche for anyone and I've successfully optimised for an audience of one :shrug:
I have some barebones content at https://struggle-meals.wonger.dev/ and will be working on the design over the next few weeks. Some decisions I'm thinking about:
- balancing between personal convenience and brevity vs being potentially useful for other people. E.g. should I tag everything that's vegan/vegetarian/GF/dairyfree/halal/etc? Should I take pictures of everything? (I'd rather not)
- how simple can I make a recipe without ruining it? E.g. can I omit every measurement? should I separate nice-to-have ingredients from critical ingredients? how do I make that look uncomplicated? (Sometimes the worst thing is having too many options)
- if/how to price things? Depends on region, season, discounts, etc
https://github.com/skorokithakis/stavrobot
It's like OpenClaw but actually secure, without access to secrets, with scoped plugin permissions, isolation, etc. I love it, it's been extremely helpful, and pairs really well with a little hardware voice note device I made:
- 2FA, PassKey, and password-based login for folks that hate magic links
- Moved my entire API from GraphQL to REST so I can fully dogfood the API I offer
- Added an audit log as standard on all plans
- Built a terraform provider (https://github.com/OnlineOrNot/terraform-provider-onlineorno...), and a way to download existing config into terraform files
- Started iterating on a CLI (https://github.com/OnlineOrNot/onlineornot)
Published a demo/experiment under MalleableTodo [1] - and so far seen some pretty strange use cases...
Essentially, just allows each user to use an LLM to rewrite their own UI to add features/customisation.
Also launching a supabase security scanner. If someone wants a free scan hit me up. Includes POCs and verification before and after remidiation. Goodbye false positives.
https://github.com/nickbarth/closedbots/ I was also trying to do a simplified openclaw type gui using codex. The idea being its just desktop automation, but running through codex by sending codex screenshots and asking it to complete the steps in your automation via clicks and keypresses via robotgo.
I was stuck on this conversation problem. First version had a dead-end search box: six starter prompts, one referencing a tool that didn't exist. No follow-ups. No guided flows. Users got an answer and had to invent the next question from scratch.
Now the assistant explores your library with you. Tag discovery, color browsing, weekly digests, smart collections that auto-curate as you save.
Semantic search runs hybrid, keyword matching plus pgvector cosine similarity on 768-dim embeddings. Streaming responses.
Almost there. https://bookmarker.cc/
The paper in question: https://arxiv.org/abs/2602.05274 (published in the Journal of Mathematical Biology)
The idea is pretty simple: SQLite is amazing, but once it’s running in production you basically have zero observability. If something weird happens (unexpected writes, schema changes, background jobs touching tables, etc.) you only find out after the fact. It tries to solve that without touching application code. It's a Rust agent that runs next to your sqlite file, and connects to the server where everything is logged in. My current challenge right now is encryption and trust, mostly.
Curious if others here are running SQLite in production and if you would be interested in something like this.
There is an API, and it’s a straightforward task, but one thing led to another and I’m also improving the app UI. The update will take some time but I hope it will only be better.
[1]: https://lab174.com/nonodle/
[2]: https://apps.apple.com/app/nonoverse-nonogram-puzzles/id6748...
Originally I made it for my grandpa, but I got a lot of interest so I made it into a full commercial product.
Just yesterday I published a set of 3 mini tutorials if you want to see how it works - https://youtube.com/playlist?list=PLKt1F5TvOjAHE07oBDlPXcrHc...
New features shipped last month:
- Adaptive practice: LLM generates and grades questions in real-time, then uses Item Response Theory (IRT) to estimate your ability and schedule the optimal next question. Replaces flashcards; especially for math and topics where each question needs to be fresh even when covering the same concept. - Interactive math graphs (JSXGraph) that are gradable - Single-image Docker deployment for easy self-hosting
Open source: https://github.com/SamDc73/Talimio
It also features a recipe manager with family/friends sync. This makes it possible to upload your grandmother’s cookbook and share them with your whole family.
I've been building a collaborative docs tool called Docules. The short version: it's a team documentation tool that doesn't have any embedded AI features. I use Claude Code daily, but putting LLMs into every workflow and charging for it is kinda insane. Every docs tool is adding AI auto-complete, AI summaries, "generate a page" buttons. Docules has an API and an MCP server instead, so you connect whatever AI tools you actually want to use. The core product focuses on being a fast, solid docs tool. Real-time collab, fast — no embedded databases or heavy view abstractions, hierarchical docs, drag-and-drop, semantic search, comments, version history, public sharing, SSO, RBAC, audit logs, webhooks, etc. The stack is React, Hono, PostgreSQL, WebSockets. The MCP server is a separate package that exposes search, document CRUD, and comments — so Claude/ChatGPT can work with your docs without us reimplementing a worse version of what they already do. Happy to talk architecture or the MCP integration.
We use landing.ai to parse the PDF, as well as useworkflow.dev to durably perform other work such as rendering PDF pages for citations, and coordinating a few lightweight agents and deterministic checks that flag for inconsistencies, rule violations, bias, verify appraiser credentials, etc. etc. Everything is grounded in the input document so it makes it pretty fast and easy. We’re going to market soon and have an approval sign up gate currently. Plenty of new features and more rigorous checks planned to bring us to and exceed parity with competition and human reviewers.
There’s plenty of margin for cost and latency versus manual human review, which takes an hour or more and costs $100 or more.
So I build these two app to create items and spell cards and print them out.
At this moment I’m working on improving the logic that decides when/how much to throttle the network.
But anyway, I've started to learn Go. By doing a vertical scrolling shooter with embiten. Kinda like fitting a square peg into a round hole. No, it's not public and will probably never be.
Studying how do do a memory pool for actors, since it doesn't look like garbage collection and hundreds of short lived bullet objects will mix well.
Stack is 15+ Go microservices on k3s. Cross-lingual semantic search is fun. Spanish query returns English calls with no translation code.
- Tablex (https://www.tablex.pro) - seat arrangement app for weddings, seminars, conferences.
- Kardy (https://www.kardy.app) - group card app I've always wanted to build.
- Jello (https://www.jello.app) - Create games with your own photos and sound effects!
Been working on this for about 4 years. It has some cool features, like letting you create your own PDF templates with HTML/CSS. Most users love that it's so simple and just a one-time purchase.
Currently thinking about how to implement an Obsidian-style cloud sync feature since that gets requested a lot.
[1]: https://github.com/loopmaster-xyz/loopmaster
[2]: https://loopmaster.xyz/tutorials
An LLM observability SDK that let's you store pre and post request metadata with every call in as lightweight an SDK as possible.
Stores to S3 in batched JSON files, so can easily plug into existing tooling like DuckDB for analysis.
It's designed to answer questions like; "how do different user tiers of my services rate this two different models and three different systems prompts?". You can capture all the information required to answer this in the SDK and do some queries over the data to get the answers.
I do wonder if the problem is not so much having a place to find LAN events but actually just having enough people put on LAN events in the first place. It feels like a thing of the past with how much less people interact in person these days. It's a shame because LANs are awesome!
Have you thought about ways to make it easier for people to host LAN events? Or does this solve that as well? I guess a solution would require matching random people together. Happy to discuss more - nick at onthe.town
While the main app is closed sourced, the rails engine that handles all the rss feeds is open sourced here: https://github.com/dchuk/source_monitor
I have another version of source monitor getting by published soon with some nice enhancements
It's mainly for censorship evasion (should be much harder to block than the regular centralized VPNs), but also for expats to access geo-blocked domestic services.
It's at the MVP stage and honestly it evoked much less interest in people than I hoped it would, but I'm still going on despite my better judgement.
Also used the new Navigation API (and some Shadow DOM) to build a cheap, custom client-side rendering (sort of) into my site (https://taro.codes), and some other minor refactors and cleanup (finally migrated away from Sass to just native CSS, improved encapsulation of some things with Shadow roots, etc).
I've been wanting to write a simple AI agent with JS and Ollama just for fun and learning, but haven't started, yet...
This is a "full rewrite," because I need to migrate away from my previous server, which was developed as a high-security, general-purpose application server, and is way overkill for this app.
Migration is likely to take a couple more years, but this is a big first step.
I've rewritten the server, to present a much smaller API. Unfortunately, I'm not yet ready to change the server SQL schema yet, so "behind the curtain" is still pretty hairy. Once the new API and client app are stable, I'll look at the SQL schema. The whole deal is to not interfere with the many users of the app.
I should note that I never would have tried this, without the help of an LLM. It has been invaluable. The development speed is pretty crazy.
Still a lot of work ahead, but the server is done, and I'm a good part of the way through the client communication SDK.
First release was in December for 1D cuts. Last month I released sheet cutting for 2D cut calculation. It's been working well for my own projects and it started getting consistent daily users since my last update in February. You can save projects now on the site for you to come back to later.
Any feedback is welcome. I'm always looking for what features to add next.
Currently adding support for exposing Postgres schemas for each app to use. The goal is that with a shared Postgres instance, each app should be able to either get a dedicated schema or get limited/full access to another app's schema, with row level security rules being supported.
Built and adding few add on features on the way: copy card numbers and view notes.
With Rust, bwc-cli - it decrypts vault into zeroize and provides near instant search with hotkey.
All data lives in your browser (IndexDB) - https://buyitlater.vercel.app
A hobby project I started putting together late last year; a little spot on the internet for prayer and reflection. I've just shipped a small feature where you get a Bible reading (KJ only for now) in response to a prayer.
https://dugnad.stavanger-digital.no/
A pro bono tech consultancy for local (Stavanger, Norway) non profits. The idea is to help them use tech to better deliver on their mission. Last week I built a little bookmarklet for a non-profit to surface some of their data buried in a SaaS tool ... which will make their apple pressing operation easier.
Not templates with names swapped in. Every story and illustration is made from scratch. You can go from "dinosaurs soccer" or write out a whole storyline. Pick an art style, optionally upload reference photos of your kid, and it builds a 28 page book in a few minutes.
Bilingual in 38 languages. We handle RTL (Arabic, Hebrew), CJK, and less common languages like Estonian, Maltese, Irish where there's not much available for kids.
Tech side for the curious: LangGraph orchestrates the pipeline, Celery workers do image generation and text rendering in parallel, and LLMs critique the illustrations for consistency mistakes and can trigger regenerations automatically.
Printed in Germany, booklet around 20 EUR, hardcover around 40 EUR.
It's a reference to https://xkcd.com/2883/, which I've always liked and was suprised there was no tool to check when you last had astronauts over for dinner.
Looking up the location of the ISS at a specific time is easy. Looking up the closest passes of the ISS to a specific location for the last 30 years on-demand is more complicated.
Since last time, added a "landing-page" kind of website [0], added annotations with BGP events, support for IPv6, and finishing TLS for every communication between probes and central servers.
About to open for beta testers, and still very much interested in comments esp. regarding the UI.
It was inspired by tamagotchis of yesteryear (and my two cats). It uses a small common monochrome SSD1306 display with 128x64 pixels of resolution.
All of the pixel art is my own. And the cat features a bunch of different animated poses and behaviors, as well as different environments. And there are minigames (a chrome dino clone - but with a cat!, a breakout clone, a random maze generator, a tic-tac-toe game, and I plan to add more.)
I'm currently working on tweaking the stats so that they go up and down over time in a realistic way and encourage the player to feed and interact with the pet to keep stats from going too low. Then I plan on adding some wireless features, like having the pet scan WiFi names to determine if its home or traveling, or using ESP-NOW to let pets communicate with each other when they're nearby.
I made a reddit post with a video of it a few weeks ago [1] and have various prototypes of artwork for these little screens on my blog [2].
[1] https://www.reddit.com/r/arduino/comments/1r8i1vx/progress_o...
Email address in profile.
[1] https://www.crowdsupply.com/scope-creep-labs/hoopi-pedal [2] https://scopecreeplabs.com/blog/?tag=hoopi
Some of these are present here: https://github.com/vamsipavanmahesh/claude-skills/
Planning to package this as a workshop, so companies could be benefit from AI Native SDLC.
Put together the site yesterday https://getainative.com
Couple of the people I have worked with in the past agreed to meet me for a coffee, will pitch this. Fingers crossed.
Also moving to Sveltia as my CMS (Astro markdown blog), after exploring multiple other options. Changed the structure of my Obsidian vault, will write about that also.
I’m also still working on a few projects:
- https://game.tolearnkorean.com/
A few years ago it started as colorguesser.com - which is not much maintained, but since there are many new users enjoying this small game, I decide to invest more time and add more feature.
Currently optimized for restaurant menus but maybe we should expand to other type of long-form materials like brochures and infographics.
* Telephone handset for my mobile phone with side talk.
* First draft of a book / workbook on Work Flow. Outcrop of the work flow consulting I do, stuff I've learned, and so on.
* Short film script - trying to convince a local actor to play the lead before we lose the rainy season here - otherwise we'll need special effects or just wait until the fall.
* Polishing firmware, OSX, and iOS suite for a wearable neuromodulator unit. Deadline in a week!
* Nmemonic community and app - been poking at this for years and finally had a breakthrough on the UI. My first app to release in the wild, so pretty exciting.
Supports only YouTube as the data source, and Gemini 3.1 Flash Lite for processing, but it can easily be tweaked. Runs locally with Docker compose.
Also working on https://www.kinoko.sh/. An agentic engineering platform built from the ground up for agents. Custom language and architecture and a layer of formal verification on top. Also working on a custom inference engine that produces well typed programs
Coffee Roaster Aggregation ETL using fastapi, nextjs, bs4 etc etc. It's been fun, just finished up the oauth for discord that pairs nicely with the info required to make Discord dm notifications function. attempting to charge 6$ for the instant notifications, but doubt many people will be interested. up to 75 roasters and all of them are checked every 10 mins for new products.
Considering reusing the repo as a framework for other industries if this project ever gains any traction. Also was considering adding a goofy rag discord bot to the server just because i love tossing in a rag layer everywhere lately, and feel like i fall a bit short on my filters for stuff like origin/flavor notes and all that junk. Semantic search with solid chunk strategies might create a better solution than if i did get all the filters working as well as possible.
I've been building a collaborative docs tool called Docules. The short version: it's a team documentation tool that doesn't have any embedded AI features. I use Claude code daily, but putting LLM’s into every workflow and charging for it is kinda insane. Every docs tool is adding AI auto-complete, AI summaries, "generate a page" buttons. Docules has an open API and ships an MCP server, so it connects to whatever you want to use LLM-wise. They can read, search, create, and edit documents through the API. The core product is just a docs tool that tries to be good at being a docs tool:
- Real-time collab with live cursors
- Fast — no embedded databases or heavy view abstractions slowing things down
- Hierarchical docs, drag-and-drop, semantic search
- Comments, version history, public sharing
- SSO, RBAC, audit logs, webhooks
Stack is React, Hono, PostgreSQL, WebSockets. The MCP server is a separate package so it's not coupled to the main app. I keep seeing docs tools bolt on half-baked AI features and call it innovation. I'd rather build a solid foundation and let you plug in whatever AI workflow actually makes sense for your team. Happy to answer questions about the architecture or the MCP integration.I've also used tweakcc to make this work in Calude Code and plan to also do one for open source coding agents - codex, pi, Gemini, etc. And I'm also doing Livestreams of the development process.
* https://sprout.vision/ - AI generated Go-To-Market Strategy for launching your next venture. I have a Tech background with limited GTM experience, so I experimented with AI to learn about different strategies and decided to turn it into a simple product that will generate a comprehensive plan (500+ pages) to help you launch your next venture. Try it out, would love to hear your feedback, use the HN50 promo code for 50% off your order.
* https://pubdb.com/ - Reviving a 10 year old project, it’s meant to make research publications more accessible to mere mortals with the help of AI. I have lots of ideas I want to try out here but haven’t gotten around to it yet. Currently focused on nailing down the basics with an OCR indexing pipeline and generating AI summaries.
This was an excuse to ship a mobile app for the first time and get familiar with supabase.
After these last few bugs are fixed, its ready for a semi-public TestFlight with our friends who have kids.
Yes, I got addicted to playing Smashkarts (over 2 hours/day). Now it is capped at 30 minutes.
inspired by the karpathy/twitter posts on running (semi) autonomous research loops, I build https://github.com/tnguyen21/labrat to be able to try and replicate some paper results over night. still early stages but I'm getting some use out of it already.
also spending a lot of time thinking about how you "close the loop" on software projects. right now figuring out how you can combine static analysis + review heuristics to let LLMs course correct the codebase when they over-engineer or produced unwieldy abstractions.
Do you plan on submitting it to the store?
It's gone a long way to solve the "review" bottleneck people have been experiencing (though admittedly it doesn't fix all of it), and I'm in the process of adding support for Mac and Windows (WSL for now, native some other time).
Some of the features I've had for a while, like multi-project agent worktrees, have been added as a part of the Codex App, so it's good to see that this practice is proliferating because it makes it so much easier to manage the clusterf** that is managing 20+ agents at once without it.
I'm feeling the itch to have this working on mobile as well so I might prioritize that, and I'm planning to have a meta-agent that can talk to Tenex over some kind of API via tool calls so you can say things like "In project 2, spawn 5 agents, 2 codex, 2 claude, 1 kimi, use 5.2 and 5.4 for codex, use Opus for the claudes, and once kimi is finished launch 10 review agents on its code".
Already launched biz-in-a-box.org and a life-in-a-box.org spinoff as frameworks to replace every entity's QuickBooks. I'm using them myself for every project my agents are spinning up.
Stealth project is related to classpass but for another category of need that won't go away even in the age of AI that really is only possible with critical mass of supply to meet existing demand. Super excited cus there's no better time to build with unlimited agents that scale without people problems.
Lastly, can't wait to run local LLMs so no longer limited by tokens/money.
I'm enjoying building solitaire and puzzle games.
My goal is to make this project the largest online collection of free modern solitaire games available for all kinds of devices.
But once all the low level operations are done, my plan is to implement an A2A Agent as the sole Agent listed in the AgentCard at $SERVER_ROOT/.well-known/agent-card.json, which is itself an "AgentListerAgent". So you can send messages to that Agent to receive details about all the registered Agents. Keeps everything pure A2A and works around the point that (at least in the current version) A2A doesn't have any direct support for the notion of putting multiple Agents on the same server (without using different ports). There are proposals out there to modify the spec to support that kind of scenario directly, but for my money, just having an AgentListerAgent as the "root" Agent should work fine.
Next steps will include automatically defining routes in a proxy server (APISIX?) to route traffic to the Agent container. And I think I'll probably add support for Agents beyond just A2A based Agents.
And of course the basic idea could be extended to all sorts of scenarios. Also, right now this is all based on Docker, using the Docker system events mechanism, but I think I'll want to support Kubernetes as well. So plenty of work to do...
Multitrack field recorder with automatic cloud sync for iPhone. I use it for hi-fi recording of band practice and sharing demos with bandmates/collaborators. Great way to send stems too as it runs on the Mac as well and has a built in mixer. There's a social graph so you can send someone a session by typing in their handle and granting access.
A site for anti patterns in online discourse.
Example: https://odap.knrdd.com/patterns/strawman-disclaimer
Need to gather more patterns then create tooling around making it easier to use.
The goal is to raise the quality of comments/posts in forums where the intent is productive discussion or persuasion.
My favorite output so far is that I asked it what life was and in a random stroke of genius, it answered plainly: "It is.".
It's able to answer simple questions where the answer is in the question with up to 75% accuracy. Example success: 'The car was red. Q: What was red? ' |> 'the car' - Example failure: 'The stars twinkled at night. Q: What twinkled at night? ' |> 'the night'.
So nothing crazy, but I'm learning and having fun. My current corpus is ~17mb of stories, generated encyclopedia content, json examples, etc. JSON content is new from this weekend and the model is pretty bad at it so far, but I'm curious to see if I can get it somewhere interesting in the next few weeks.
* https://theblue.social — TheBlue.social, provides Bluesky native tools
* https://stacknaut.com — Stacknaut, SaaS starter kit to build on a solid foundation with AI, includes provisioning on Hetzner, deployment with Kamal 2 and dev with coding agents
* https://codevetta.com — Codevetta, Architecture and code reviews service
* https://myog.social — MyOG.social, OG Image Generation Service
I've been planning a new idea with that and possibly future ideas based on the future (and near future) where there are more and more "agent" users.
It's a collection of 40 (and growing) tools for text processing, data cleaning, conversions, dev utils etc. Everything runs in the browser and is completely free.
Started this partly to learn SEO from scratch on a fresh domain, partly because i am lazy with regards to doing basic data cleaning using pandas and i found myself repeatedly using similar online tools that are completely riddled with ads.
I built this using Flask + Vanilla JS. I don't think there was any need to overcomplicate it. And for fun, i vibe coded a windows 95 desktop mode where all the tools open as draggable windows. https://textkit.dev/desktop
Its been pretty fun cosplaying as an network engineer, and now I'm building out an Anycast network for a few ideas that I'm working on.
Its nothing too revolutionary or new, but I'm proud that I've built them from ground up and all running on my own infrastructure.
- DNS Authoritative Hosting - https://thelittlehost.com/dns/ - Quietnet - A family-focused internet filter - https://quietnet.app
I'm also getting ready to launch https://relaye.io, which was my personal tool I built to support my devops consultancy.
The original Python icloudpd is looking for a new maintainer. I’ve been building a ground-up Rust replacement with parallel downloads, SQLite state tracking, and resumable transfers. 5x faster downloads in benchmarks, single binary, Docker and Homebrew ready.
Downloaded and parsed a bunch of the pgsql-hackers mailing list. Right now it’s just a pretty basic alternative display, but I have some ideas I want to explore around hybrid search and a few other things. The official site for the mailing list has a pretty clean thread display but the search features are basic so I’m trying to see how I can improve on that.
The repo is public too: https://github.com/jbonatakis/pginbox
I’ve mostly built it using blackbird [1] which I also built. It’s pretty neat having a tool you built build you something else.
It has a few core libraries built in rust with a web app and a terminal UI. Android app is in the works. The persistence layer is intended to be offline first using a CRDT with an optional sync server. I'm also trying to integrate "bring your own AI" assistants to help tweak recipes or make suggestions.
It's been a fun way to sharpen my claude skills but also to see how feasible it is to maintain multiple frontend applications with a large amount of shared code. Still a lot to do, particularly the core calculations are not yet on par with existing offerings.
^^ project with my daughter
Parallel agents debate your ideas/work: https://murderboards.ai
^^ solo project / code review agents for non-coders / inspired by Compound Engineering plugin’s code review flow
Chronomaly is a flexible and extensible Python library for time series forecasting and anomaly detection using Google TimesFM.
i have also started experimenting with qwen3.5 0.8B model, my goal is to create agents with small models that are as robust as their commercial counterparts for specialized tasks. currently trying it for file editing.
To date it's handled more than 70k orders, ingested nearly 10m telemetry records, has been extremely reliable, is almost entirely self-contained (including the routing stack so no expensive mapping dependencies) and is very efficient on system resources.
It handles everything from real-time driver tracking, public order tracking links, finding suitable drivers for orders, batch push notifications for automatic order assignment, etc.
I originally made it a couple of years ago as a small proof of concept. A couple of weeks ago I started it over and have been using it as a project to work with Claude and learn approaches to coding with AI.
It's been a lot of fun.
The main thing I'm currently working on is a platform for organizing and discovering in-person events. Still not certain about the boundaries for "Phase 1", but I have a bunch of ideas in that space that I've been incubating for a while. One subset of features will be roughly similar to that app you've probably heard of that starts with 'M' and ends with 'p', but hopefully an improvement, at least for the right audience. But wait, there's more. :)
Currently building it; it's not public yet, so no link. Next month.
Thinking about how to grow the userbase is intimidating, but I think it might end up being fun.
I only got to the point of having code and data as \verbatim in \LaTeX. Next step is CWEB.
Here is an example (with C and Rust code in \verbatim)
https://ontouchstart.github.io/rabbit-holes/llm_rabbit_hole_...
The ultimate goal is machine and human readable proofs on algorithms.
The server is a rust binary so you can toss it on any container/computer and connect to it in the app.
My philosophy isn't to replace my other tools I love like emacs, ghostty, etc. But I am taking a stab at "real time code review" and have some crummy magit-like code review built in that I need to revisit.
The basic idea is that when one failure fans out across 20 services, you often end up with 20 alerts and 20 separate investigations, even though there is really just one root cause. I’m using distributed tracing to build a live model of how errors propagate through the system, and then exposing that context directly at each affected service.
Longer term, I want this to become a very high-precision RCA engine. Right now I’m looking to try it with a few early design partners that already have a lot of tracing data, especially OpenTelemetry or Datadog APM users. I'll love to chat with some folks who would be willing to try it out!
- Urbanism Now - I run https://urbanismnow.com, a weekly newsletter highlighting positive urbanism stories from around the world. It’s been exciting to see it grow and build an audience. I'm thinking of adding a jobs board soon that'll be built in astro.
- Open Library - I’ve been helping the Internet Archive migrate Open Library from web.py to FastAPI, improving performance and making the codebase easier for new contributors to work with.
- Publishing project - I’m also working on a book with Lab of Thought as the publisher, which has been a great opportunity to spend more time working with Typst.
These projects sit at the intersection of technology, cities, and knowledge sharing, exactly where I’m hoping to focus more of my time going forward.
working on an AI-native Kubernetes sidekick that watches your pods, reads the logs, and turns failures into clear fixes before they become outages
Demo fase, showing the branded version to potential clients. We iterate on it with their feedback.
So I built a test for the same and its called Orlog. You can check it out here: https://orlog-test.netlify.app/
Looking for your feedback.
(Sign up and I’ll send out beta codes tomorrow!)
I’ve had a few friends call it “fun”, and one said it’s “Scrabble that doesn’t drag”.
Working on some social media shareable replays you can post after matches tonight, thanks to Remotion:
Aside from that, I need to document and properly release one of the pieces that PAPER is relying on (some generic tree-processing code that makes operations on directory trees a lot nicer than with the standard library "walk"s), and work on others (in particular, a "bytecode archive" format for Python that speeds up imports for some projects, mainly by avoiding filesystem work at import time — I want to offer it as an install-time option in PAPER, and later have `bbbb` make wheels with the bytecode precompiled that way).
Phase 1: Download the student's code from their submitted github repo URL and run a series of extractions defined as skills. Did they include a README.md? What few-shot examples they provided in their prompt? Save all of it to a JSON blob.
Phase 2: Generate a series of probe queries for their agent based on it's system prompt and run the agent locally testing it with the probes. Save the queries and results to the JSON blob.
Phase 3: For anything subjective, surface the extraction/results to the grader (TA), ask them to grade them 1-5.
The final rubric is 50% objective and 50% subjective but it's all driven by the agent.
- http://sharpee.net : Text Adventure authoring platform in Typescript
- https://github.com/ChicagoDave/tsf : A multi-target npm build tool
- https://devarch.ai : Claude Code guardrail workflow including hooks, agents, and skills
In progress:
- unnamed project to disrupt commercial site hosting including a new marketplace
Full encryption for notes (uses local encryption before you even sent the note to the server).
I wanted a mixture of Github Gists (sans Git) and 1Password shares so I've been using it eitj great success at my current company to share snippets and private stuff.
Might open source in the future, just need to gauge interest.
I just released support for dashboards. I've kept a devlog for the past 6 months.
And the biggest update is coming soon, DB Pro Cloud, which will let you connect to and manage any database through your browser as well as collaborate with your team.
Imagine Postman but for databases.
Info (not recent) available here: https://awz.us/docs
It's written in Golang and acts as a simple desktop app that creates a web server and then opens the site in your default browser. This way it's easily multi-platform and can also be hosted as a SaaS for larger production houses.
- No sign-up, works entirely in-browser
- Live PDF preview + instant PDF download
- Flexible Tax Support: VAT, GST, Sales Tax, and custom tax formats with automatic calculations
- Shareable invoice links
- Multi-language (10+) and 120+ currencies
- Multiple templates (incl. Stripe-style)
- Mobile-friendly
- QR Code Support: Add payment QR codes with any invoice-related information (payment links, UPI, contact details, custom data)
- Multi-Page PDFs: Seamless multi-page support with automatic pagination and page breaks
GitHub: https://github.com/VladSez/easy-invoice-pdf
Would love feedback, contributions, or ideas for other templates/features.
PS: e-invoice support is wip
Under the hood it uses a cool legal reasoning agent primarily designed for understanding litigation claims and objectives.
Sometimes I do wish I had a slack channel of like 30 attorneys so I can ask them questions and get feedback.
Physical engineering is a huge welcome transition for me from what coding has become in the last couple years.
There’s something nice about the realities of creating a model, then printing it, then seeing that exact is too exact, then reprinting, then eight more times, and then that feeling when it all comes together properly.
A few weeks ago I was working on an adapter for an airbrush to use on a standard pancake air compressor. Learning to create threads in blender was really neat! I learned a lot about the physical construction of threads, something I have never put much thought into before.
I had been doing lots of time-based work for a blog post and ended up annoyed that so many clocks around me were visually out of sync. Especially my microwave and oven clocks. Using the tool I got them synced up beyond what I could perceive.
Now shifting to established SaaS companies adding AI assistants to their existing products. Some of them literally have people reading chats full time, so they actually value the experience.
Building https://lenzy.ai - 2 paid customers, 2 pilots, looking for more and figuring out positioning.
Run your agents contained (container or VM, Linux or Mac), with all restrictions removed.
Workflow:
┌─ YOLO shell ──────────────────────┬─ Outer shell ─────────────────────┐
│ │ │
│ yoloai new myproject . -a │ │
│ │ │
│ # Tell the agent what to do, │ │
│ # have it commit when done. │ │
│ │ yoloai diff myproject │
│ │ yoloai apply myproject │
│ │ # Review and accept the commits. │
│ │ │
│ # ... next task, next commit ... │ │
│ │ yoloai apply myproject │
│ │ │
│ │ # When you have a good set of │
│ │ # commits, push: │
│ │ git push │
│ │ │
│ │ # Done? Tear it down: │
│ │ yoloai destroy myproject │
└───────────────────────────────────┴───────────────────────────────────┘
Sandboxes: Docker (Linux, Mac), Seatbelt (Mac), Tart (Mac)Everything's contained in a single go binary. Just build and run.
I went through the Software Architecture Patterns for Serverless Systems book, which I think it is fantastic. I learned a lot but I still had a lot of doubts to actually use the ideas in real life. So I started dissecting the companion framework, which is in written in Typescript. I have been going piece by piece and converting to Kotlin which I think it is more expressive (and fun) and it is allowing me to understand how everything fits together.
Typescript framework: https://github.com/jgilbert01/aws-lambda-stream
A problem that we had at my last startup was that we got stuck between not wanting to spend too much time on devops, and getting price gouged by Heroku.
We were too big for the deploy to a VPS type options like coolify, but too small to justify hiring a full time Devops.
Eventually a few of us had to just suck it up and learn Kubernetes properly. Was pleasantly surprised how elegant it all was.
I was surprised there wasn’t something that “just worked” and plugged into our Kubernetes cluster, made it user friendly, teams, roles, etc.
Blog, news, chat, video, mail, web. Basically all the daily habits as little micro apps in one thing. I find it quite useful. Not sure anyone else does yet though.
Also separately worked on Reminder.dev which is a Quran app and API that bakes in LLM based search and motivational reminders.
AM3 - (Allied MasterComputer or Artificial Mind, version 3) - An attempt to make a symbolic AI that approaches the capacities of a LLM. An LLM makes variations on the same code and schedules those variations to play in "games". The results allow the LLM to make further changes.
Agents can search for design inspiration from production websites using semantic search. Since this inspiration comes from live websites, their design tokens; colors, typography usage, layout data are also available.
It's designed to integrate with Maven projects, to bring in the benefits of tools like Gradle and Bazel, where local and remote builds and tests share the same cache, and builds and tests are distributed over many machines. Cache hits greatly speed up large project builds, while also making it more reliable, since you're not potentially getting flaky test failures in your otherwise identical builds.
The problem I kept seeing: freelancers have happy clients but almost no testimonials on their site. Asking is awkward, clients say "sure!" and then never write anything.
SocialProof gives you one shareable link. Client clicks it, fills a short form (name, text, optional photo), you approve it, it embeds anywhere. No login required for the client.
The interesting technical bit: it's entirely on Cloudflare Workers + D1 + Pages. The collection form and embed widget are edge-served globally with no origin server. Been curious whether anyone else is building purely on Cloudflare's stack and what they've run into.
Still pre-revenue (just launched today). If you're a freelancer or run a small agency and have thoughts on how you currently handle testimonials, I'd genuinely love to hear it.
Version 3 will add more feed types: Podcasts, Mastodon, Twitter/X, Calendar, Reminders, Weather, Finance tickers and more.
It will have a new UI, new features like notifications and local transcripts and summaries and many quality of life improvements.
It’s a drop-in replacement for Redis written in Rust. Most if not all of your client code should work without issues. Outperforms in many areas and has more out of the box features like proto storage, raft/swim, and encryption at rest.
I’m pretty proud of it, and I hope you’ll give it a shot and open bug reports. :)
The workflow: Upload doc → LLM extracts structured data → Generate new doc from template.
It’s API-first, includes webhooks, and is built to be self-hosted/self-provisioned for privacy. Still very much a WIP, but looking for feedback on the feature set and the extraction accuracy.
URL: https://fetchtext.io
https://videohubapp.com/ & https://github.com/whyboris/Video-Hub-App
It is based on 20+ years of experience maintaining a similar system in Perl.
It's on Hex.pm already, looking for people to test and comment!
As Codex would say:
Selecto is an open-source SQL query builder for Elixir that helps you generate complex queries from clean, domain-based configs. It supports advanced joins, CTEs, subqueries, and analytics-friendly patterns, with companion packages for LiveView interfaces (selecto_components) and code generation (selecto_mix). If your app is data-heavy, Selecto gives you SQL-level power without brittle hand-written query strings.
Jive Data: https://jivedata.com
Financial and Investing data
Random Data Monster: https://randomdata.monster
Random Data (also available as a Google Sheets Add-on)
WhatIsMyIPAddress.Monster: https://whatismyipaddress.monster
A clean website to get your IP Address. Also available as an API.
Phone Monster: https://phone.monster
Caller ID, but on steroids
It basically pulls OpenTelemetry data from your infra and renders chernoff faces, so you can spot anomalies at a glance.
It's heavily supported by Claude Code, but much fun.
Actually not built on this yet I think, but I could switch over, haven't made anything more of it since it's still a bit rough around the edges, and I keep finding various issues during actual usage: https://binschema.net/
Several readers have asked for an easy way to get recommendations without working through long-form review articles.
Here's the first iteration of a simple recommender: https://bcmullins.github.io/reading/
Very early days but will keep updating them & adding more.
Check out this twisty vase demo: https://nodillo3d.com/s/VmP0nJdKRcPazQ1g
You can also share you files and create sharable configurations as well. Here is the same vase as a configurator: https://nodillo3d.com/v/a9REIEZIDYGtzZRA
I would like to do a more detailed intro class to help people learn how to model with nodes.
Hope you enjoy it!
It has gained a little traction in Reddit and grateful for the several paying users currently giving me lots of feedback. One of the features is that you get to import your own font using any otf, ttf files. App is 100% native too written in SwiftUI, AppKit and UIKit.
I just wanted my own interpretation of an RSS Reader app, I have been a heavy user of both Reeder and NNW but the interface is just the same and I got bored a lot.
- Portable Secret (https://alcazarsec.github.io/portable-secret/) - self-contained HTML files that decrypt in the browser.
- Dead Man's Switch (https://alcazarsec.com/deadmanswitch) - sends messages when you stop checking in.
- Flare (https://alcazarsec.com/) - silent alert when your device is accessed without authorization.
I ask because I was recently thinking about how to preserve information for the future like this
This seems unlikely, however, since our infrastructure costs for the dead man's switch are covered by just a handful of subscriptions. Besides, we host it next to our other more profitable main product, so it gets free maintenance.
We are up for the challenge of making this last for many decades, though. It is a beautiful mission.
https://x.com/ZDi____/status/2013655958027669958
Right now, I only have single speaker checkpoints (as per the old video). That will change soon.
It seems like you have been working on this application for sometime, i will go through your code , but could you provide some context about upgradations/changes you have made, or some post describing your efforts.
Cool nonetheless!
A project that I launched on HN that became a business. Simplescraper rode the no-code wave of a few years back ('instant structured data without parsing html').
Now working on increasing the surface area for AI agents: MCP support, screenshots API, and (experimentally) x402[1]
It suggests to me that the underlying architecture probably isn't too complicated, so I'd wish for an open-source solution
The insight: the friction in getting testimonials isn't that clients don't want to help – it's that a blank "leave a review" box produces mediocre one-liners. SocialProof guides them through structured questions ("what was your situation before?" / "what changed?") so you get a compelling before/after narrative automatically.
Free tier: unlimited testimonials. Just launched and looking for feedback from anyone who deals with client testimonials.
It makes connecting user domains to your app easy and reliable at any scale. Each Approximated user gets the own globally distributed, managed cluster of servers with its own dedicated IPv4 address. Includes (unlimited) edge rule features, DDoS protection, webhooks, and more. Make a simple API call, tell the user to point an A record at the IP, and it’s connected to your app with its own SSL certificates.
Built/building with elixir and phoenix, which has been fantastic.
I put together a pretty basic portal clone. I think its pretty cool to see it come together, animations, level creation, portal jumps.
The basic hardware on the ds makes 3d pretty approachable. Ive found opengl overwhelming in the past. It seems like a fun platform to make games on, but idk if there is any active ds homebrew communities. Anyway sharing because i thought it was cool, hard to find anyone that seems to be to interested. I thought about getting a 3ds but they are surprisingly expensive now
I've been working on a solution to automate solar+battery use to arbitrage the market. I'm on a real-time utility plan but even if you're on TOU it can save you $1+ per day by strategically planning when to use the battery and when to conserve or charge the battery. So far it's limited to a few providers and only FranklinWH batteries but I'm eagerly looking for someone to help me get Powerwall support working and other ESS. It's open-source on GitHub as well.
When training I like to have every day mapped out with how many miles to run, at what pace, etc as an event in my calendar. My actual workout gets uploaded into Garmin and Strava, but I always wanted it back in the calendar so I could see at a glance the consistency over time. It's been really fun to see other people use and get value out of something I built for myself.
I went full TDD with the app so it was easy enough to build the logical parts of it. The UI is fairly simple, but whenever I found that Claude did not understand exactly what I wanted, I gave it a screenshot/image of a design and it did things pretty well.
Now at 350k lines. Native and wasm binaries (you can try the limited wasm version online). Currently adding a full CPython test suite benchmark.
Just for fun, not trying to replace CPython here. Mainly to test the limits of current coding agents.
There is a surprising amount of edge cases that can cause ChatGPT or others to misunderstand your pages. Some models can handle div based tables, some want alt tags but cannot understand title tags, etc.
I built the tool to check your site as close as possible to what a human would see and then compare it with LLM's.
It was a weird journey trying to tease this info out of the models, they will happily lie, skip checking sites or just make things up.
I wrote about it here: https://jdsemrau.substack.com/p/pair-programming-superbill-w...
The goal is to get consistent synthesis to 450MHz such that I can use a narrower 256-bit instead of a 512-bit interface, while maintaining full bandwidth. I've got it working at an FMax ranging 440-490MHz, though there's still some edge cases I need to hammer out.
More context https://lepisma.xyz/2026/02/24/harp/
https://dnsisbeautiful.com - clean, ad free dns lookup tool.
https://evvl.ai - combination of Github Gists and AI output comparisons (evals)
https://finalfinalreallyfinaluntitleddocumentv3.com/ - free mac app to intelligently rename any kind of file (photos, videos, audio, text) based upon their contents.
Zero ops S3 based log search: https://github.com/amr8t/blobsearch
Build enterprise grade applications (in Elixir) with AI the right way.
Secure. Scalable. Reliable.
Built based on a senior engineer's experience. Uses 10 years of battle-tested patterns, not just LLMs:
1. Uses algorithms over AI whenever possible.
2. No external library dependencies whenever possible.
3. Old school over shiny new toys. Use the right solution for the problem (Eg. SQL vs NoSQL).
For now it uses UX patterns to make it easy to remove uninteresting articles and keeps a record of your read and saved articles. All locally of course.
I’d like to make it into something we can share quality content with one another eventually. For now I’m focusing on making it good enough my entourage will want to use it
There is a Vulkan based browser which you can use to connect to the only public site so far, a playable breakout clone.
I want to treat my Downloads folder (or some other one) like an "Inbox" where I can just dump everything, and then the program knows where exactly in my (Johnny Decimal) file system the file should land.
I do no tracking, no analytics, just help you cross the airgap between web and mobile app so you can send users to the right place (and track them however you deem necessary)
If we can nail this one, then an entire Oxford-grade education in the same style.
- look for feedback on the Freelance Rates calculator https://heygopher.ai/tools/freelance-rate-calculator
One thing I've noticed building in this space: freelancers are remarkably bad at collecting testimonials from clients (who usually love them!). The workflow ends after the invoice is paid and nobody ever goes back to ask for a written review. Worth thinking about whether that's a hook you could add — "invoice sent, client paid → automated ask for a testimonial."
I'm building something adjacent to that problem: socialproof.dev. Would be curious what your users say when you ask how they handle testimonials.
More movement than CS, less than quake
Focused on infiltration mode - one team stealing a briefcase back to base with the other on the defense
Export your Apple Health data directly to Markdown files in your iOS file system.
Open-sourced it at https://github.com/CodyBontecou/health-md.
Fun little vibe-coded app that has made a lot of users happy.
Platform for running web apps.
Single static binary and SQLite
lua for now (WASM future)
DEMO:
Lots of this is going to involve getting people more up to speed on CS, can't wait.
I'm working on socialproof.dev which automates that step — shareable link, structured form, one-click approve and embed. Wondering if that kind of tool would fit into what a growth agency delivers to clients, or if it's something you'd rather solve with AI prompts and an email sequence.
My kid played it, and didn't stop for 45 minutes so I think that's a win :-)
Modifications to my Land Cruiser j90
- LED daytime running lights / off-road LED light bar
- Winch
- Front left – tie rod end (both)
- Rear axle – pinion bearing (loud while driving)
- Right rear brake caliper – brake fluid leaking from the piston
- Boost chip (chip + turbo tee), Kill switch
No accounts required, all data is yours and lives on your computer.
Check it out: https://greentea.app
Yes, you can use your own API key as well.
I have found out that it is very efficient to use phaser.js/three.js for fast, vibe coding, because it handles everything without having to setup a unity scene manually or unreal blueprint. I really recommend to make web apps instead for vibe coding. I love how easy it goes.
Wondering if there are other similar tools out there which people love, and why ChatGPT/Gemini/Claude won't let you do the same in their native apps.
https://www.ai-proof-careers.com/
Super annoyed by the "AI will take your jobs" hysteria, so I pulled BLS data and analyzed talks by AI researchers and a few industry folks, and ranked 900+ BLS jobs by AI resilience.
Example : https://shorturl.at/We3dH
Check it out here : https://github.com/VoidenHQ/voiden
I’ve already completed the research, business model, competitive analysis, feature set, branding, and the full UI (40+ screens).
The MVP/V1 is currently in development. When the V1 is ready I’m planning to do a Show HN with this account.
It's my first product. Any feedback or questions are very welcome, even if it's just based on the idea and the screenshots on the site, since the product isn’t available to try yet.
Shamelessly trying to attract new monthly sponsors and people willing to buy me the occasional pizza with my crap HTML skills.
A theme hospital meet silicon valley game :D https://burn-rate.pages.dev/
Anyone can learn to type fast - I think it just takes the right tools to make it interesting enough for the users to use daily
Using a webcam, monitor finger movements and find mistakes (using some sort of AI video analysis) to help user figure out how to improve. It's a hard thing to build but if you build it there is going to be paying customers. You can even sell hardware and subscriptions with it. Lots of schools want this!
- Plask ( https://plask.dev ) — Google Analytics (GA4) connected analytics dashboard for people who ship multiple products. I got tired of manually checking separate GA4 properties for all my apps and SaaS projects, and setting up individual MCP integrations for each felt like overkill when I just wanted a quick overview. So I built a single dashboard that connects all your GA4 properties, runs statistical anomaly detection, sends alerts when something breaks, and generates AI weekly digests. Free tier for 2 properties, Pro at $9/mo.
- Kvile ( https://kvile.app ) — A lightweight desktop HTTP client built with Rust + Tauri. Native .http file support (JetBrains/VS Code/Kulala compatible), Monaco editor, JS pre/post scripts, SQLite-backed history. Sub-second startup. MIT licensed, no cloud, your requests stay on your machine. Think Postman without the bloat and login walls.
- APIDrift ( https://apidrift.dev ) — Monitors changelogs for APIs, SDKs, and libraries you depend on so you don't get blindsided by upstream breaking changes. Scrapes docs, diffs changes, classifies severity with AI, and sends digest emails. Track your dependencies, get alerted when something breaks. Free tier covers 3 sources with weekly digests. Built with Next.js, Supabase, and Gemini Flash.
- Mockingjay ( https://apps.apple.com/app/id6758616261 ) — iOS app that records video and streams AES-256-GCM encrypted chunks to your Google Drive in real-time. By the time someone takes your phone, the footage is already safe in the cloud. Built for journalists, activists, and anyone who needs tamper-proof evidence. Features a duress PIN that wipes local keys while preserving cloud backups, and a fake sleep mode that makes the phone look powered off during recording.
- Stao ( https://stao.app ) — A simple sit/stand reminder for standing desk users. Runs in the system tray, tracks your streaks, zero setup. Available on macOS, Windows, Linux, iOS, and Android.
- MyVisualRoutine ( https://myvisualroutine.com ) — This one is personal. I have three kids, two with severe disabilities. Visual schedules (laminated cards, velcro boards) are a lifeline for non-verbal children, but they're a nightmare to manage and they don't leave the house. So I built an app that lets you create a full visual routine in about 20 seconds and take it anywhere. Choice boards, First/Then boards, day plans, 50+ preloaded activities, works fully offline. Free tier is genuinely usable. Available on iOS and Android.
- Linetris ( https://apps.apple.com/app/id6759858457 ), a daily puzzle game where you fill an 8x8 grid with Tetris-like pieces to clear lines. Think Wordle meets Tetris. Daily challenges, leaderboards, and competititve play against friends.
Not similar to linetris but its tetris meets a block puzzle
Arch Asxent https://github.com/mikko-ahonen/arch-ascent - tool for analyzing large microservice networks with hundres of microservices and creating architectural vision for them, and steps to reach the vision
Product link: https://ziva.sh/
Golang inference engine from scratch that can run a bunch of models with vulkan acceleration.
Automated personal outreach app for job seekers, integrated with Gmail.
Art search for magic cards
Built a Cythonized Icecast2 implementation I've wanted for years: https://github.com/lukeb42/cycast
Built a p2p Kanban board that fits in a single .html file and uses only the Python stdlib for LAN discovery https://github.com/lukeb42/kanban_p2p
Developed a p2p legislature that scales from a small team of 3 users to countries of tens of millions of people: https://gist.githubusercontent.com/LukeB42/deb887691f13dee9c...
Developed a small SPA framework inspired by React, Ractive-Load and hn.js: https://lukeb42.github.io/vertex-manual.html
Updated a news archival service for Python 3.x: https://github.com/lukeb42/harvest
Made a scriptable IRC client inspired by irssi and mIRC: https://github.com/lukeb42/scroll
and worked on a couple of my company's products.
I've been on/off working on a Forth compiler for the NES. It will be open source soon enough but I'm not happy with the code right now as it's extremely messy, repetitive, and buggy, but I think it's turning out ok. I am resisting the urge to use Claude to do all the work for me, since that's depressing.
I've also been working on a clone of the old podcasting website TalkShoe. It's nothing too complicated. It's mostly an excuse to learn a bit more about Asterisk and telephony stuff. I'm hoping to have something fully usable in about a month or two.
I forked the main MiSTer binary due to some disagreements I had with Sorg in how he's running things [1]. My fork was largely done by Codex and Claude, but the tl;dr of it is that it has automatic backup of your saves, tagging and versioning of your saves, and it abuses the hell out of SQLite to give better guarantees of write safety than the vanilla MiSTer binary gives you. I've been using it for a few weeks now and it seems to work fine, and it's neat to be able to tag and version saves.
I think that's mostly it. I'm always hacking on something so there might be a straggler there.
[1] https://github.com/Tombert/Main_MiSSus/blob/master/README.md
We're also trying to use AI more thoughtfully than just bolting on a chatbot. We're planning to consider each workflow our customers need and how AI might help speed them up - even letting them build custom AI workflows. I think most businesses (especially smaller businesses) don't want to work at the level of Claude Code, Codex, etc. They want to work on higher level problems - build this dashboard, connect these data sources, invoice this customer, etc.
Aside from that, we've noticed that the basics really matter, so we're trying to nail that first.
We're definitely a bit delusional, we're just 3 people, we're doing it without funding and the competition is stiff, but we really believe in the product. Additionally, I think a lot of CRMs go south by taking on too much VC that naturally pushes them to prioritize ROI instead of continually improving the product.
It incorporates also complaints from a static analyzer for Python and Javascript that detects 90+ vibe slop anti-patterns using mostly ASTs, and in some cases AST + small language models. The complaints give the local class and methods a sense of how much pain they are in, so I give the code a sense of its own emotional state.
I also build data flow schematics of the entire system so I can visualize the project as a wire diagram, which is very helpful to quickly see what is going on.
- Named `gg` for grep-ibility and ease of typing.
- However Claude has been inserting most calls for me (and can now read back the client-side results without any dev interaction!)
- Here is how Claude used gg to fix a layout bug in itself (gg ships with an optional dev console): https://github.com/Leftium/gg/blob/main/references/gg-consol...
---
# I've been prototyping realtime streaming transcription UX: https://rift-transcription.vercel.app
- Really want to use dictation app in addition to typing on a daily basis, but the current UX of all apps I've tried are insufficient.
---
# https://veneer.leftium.com is a thin layer over Google forms + sheets
- If you can use Google forms, you can publish a nice-looking web site with an optional form
- Example: https://www.vivimil.com
- Example: https://veneer.leftium.com/s.1RoVLit_cAJPZBeFYzSwHc7vADV_fYL...
- DEMO (feel free to try the sign up feature): https://veneer.leftium.com/g.chwbD7sLmAoLe65Z8
https://craft-burgers.openship.org/
Github:
https://github.com/openshiporg/openfront-restaurant
We're actually building an opensource SaaS for every vertical. We shipped our Shopify alternative end of last year and after restaurant, we have hotels, grocery, and gyms next.
It also comes with nice features and benchmarking abilities. For running evals, it has a companion called Calibra https://calibra.swival.dev
It is a forum application where each community is invite only. Think a cross between reddit/discord. The invite only architecture reduces trolls, spam, AI slop and promotes more substantive discussions.
Right now invitations are limited to 1 per day for each user in a community. You don't need an invite to join at the global level - but to join any community you must have received an invitation link. Still a major work in progress, right now working on expanding the flexibility of community creation and invitation logic. (allowing bulk invites, adding flexible invitation cool downs, etc).
Also I gave that blog post to Claude Code and asked to implement the API and it made terrible terrible mistakes. Just saying.
About an hour ago I was dismissed as AI slop on the r/rust Reddit. Whatever.
This tool is my line of defense in case `trunk` goes dead, which it seems to be increasingly likely. It helps me build fullstack sites using Actix Web and Yew.
Using it now to see if I can re-invent my blog site for the umpteenth time. :)
Also a bunch of other smaller projects and ideas.
so far, ive spent a lot of manual time labeling and matching RGB and LWIR images, and trying to figure out first ways to get better pose matches when the flights arent the same.
that, and many different attempts at getting torch to work using my laptop's GPU and NPU. i think im close, without having to build torch from source woo.
Ive been having an eye towards getting better llm generation quality for python too, but havent put a focus on it yet. im fed up with it making one off script after one off script and instead of just making a react app, making some raw html and making a new html file with the new and old bugs every time i want to do something interactive. its maddening.
my last month of gettin claude code ro play pokemon webt well and ive about learned skills pretty well now, but it keeps wanting to do like a challenge run of sticking with a single pokemon.
button1 = create_button('Hello world!')
button1.on_clicked {
the_hello_world_button_was_clicked
}
# this is the verbose variant in a pseudo-DSL,
# I like things being explicit. In most code
# I may omit some parts e. g.
_ = button('Hello world!') { :the_hello_world_button_was_clicked }
It defaults to ruby and what ruby supports (including
jruby-swing) but two additional languages to use are
python and java. Anyway.I recently added the possibility to describe what kind of widgets are to be used via a yaml file, as an option. This may not sound like a huge win, but so far what I like here is that it becomes easier to modify individual widgets without having to sift through code; and it works for more programming languages too. Any customization for the widget, including method-invocations if necessary, can be done via a yaml file now. There is of course a trade off in that the yaml file can become a bit complex (if the GUI uses many widgets), so for the most part I use this for smaller widgets/components that do one specific functionality (or, few specific functionalities). For instance, a GUI over wget. Then if other larger programs need that, I make this small widget more useful and flexible.
The distant goal is to actually use a simple DSL that also would allow average Joe to customize everything in a very easy manner; and to have a widget set that can be used for as many different parts possible including wonky ideas such as having a whole operating system as a GUI available one day (a bit like webmin, but not limited to what webmin does; for instance, I'd also have games such as solitaire, reversi and so forth). I'd like to see how far that idea can go, but it is just a hobby so I can only invest little time into it.
Tech details: I found that used, small form-factor Dell Optiplexes are great for product protoytyping. I'm in Medellin Colombia, and found that you can buy these for about $200 USD - they are often former Point of Sale (POS) or office computers, from about 10 years ago. They have SSDs, run quiet, and are very reliable.
For project Affirmator, I installed Linux Mint Debian Edition (LMDE). Using Cron and Mpv to shuffle-play activity-specific folders of MP3s at the same time each day. For example, for the chill jazz music - I've got a folder of 40+ song MP3s. Cron plays those at 06:30. So it's like a calm, upbeat alarm clock. I'm not a morning person, so this is a "friendly" way for me to wake myself up!
For the vocal affirmation part - I built a Python tool that reads 200+ text affirmations from a markdown/text file. It then uses AWS Polly text-to-speech API to vocalize the affirmations into MP3s. Next, I use `ffmpeg` to add a variable silent spacer gap to the ends of all the MP3s. This allows your to hear a voice affirmation ("I am fit, athletic, and strong!", "I am a confident piano player."), and then there is silent space for you to say it out loud, or repeat in your head.
This project incorporates ideas & routines from: The Strangest Secret by Earl Nightingale, Tony Robbins Personal Power II, Think and Grow Rich by Napoleon Hill, and Atomic Habits by James Clear
https://codeberg.org/jro/Affirmator-app
(2) PROJECT "LINGOFREQ" - Language learning tool. Uses language-specific high-frequency word lists. Generates example sentences according to a theme/topic. Translates the word & example phrases to English / Spanish / Chinese. Uses Text-to-speech to vocalize the phrases into each language. These phrases are ordered by frequency. When you want to improve your language skills, you set a "window" range of frequency you want to practice, and Lingofreq will play audio files in this range. You can learn Chinese & Spanish while doing the dishes, at the gym, or before going to bed!
Code: https://codeberg.org/jro/LingoFreq-app/src/branch/main/apps
(3) Medellin COMMUNITY MAKER-SPACE / CREATIVE ENTREPRENEUR LAB I'm at Medellin Colombia - my mission is to create the best maker-space. I was a member of ASMBLY Maker-space in Austin Texas (great space!) and worked at Pivotal Labs (agile product prototyping / software lab) - so I'm aiming to combine the best ideas from those.
BACK-BURNER projects:
Documenting my Knowledge as "Public Knowledge Base" - https://codeberg.org/jro/Knowledge - Here are my notes on Python, Git - I'm bouncing between Obsidian Sync / Publish / Markdown (currently easiest way), and some sort of open-source knowledge base website (VSCodium + Markdown + FOAM + MkDocs + RClone). I haven't found a solution I'm happy with yet...
Open-source CNC router tech stack: - I have a CNC router (robotic drill which can carve 3D shapes into wood). Last year I challenged myself to operate it completely through an open-source tech stack. This took me on a journey of learning Inkscape (2d vector design tool, SVG), FreeCAD (3D product design / CAD / CAM tool), G-code (format of text instructions which tell CNC tools where to move and what to do), Universal G-Code Sender (a tool which imports CAM - computer aided manufacturing - designs, connects to the CNC router tool, and actually operates machine. It's quite exciting to play with! Used Kiri-moto (web-based CAD / CAM tool) to convert 2D/SVG designs into 3D shapes). Used OBS (screen recording/streaming tool) and a bunch of web-cams to live-stream tool usage to PeerTube Live (similar to YouTube).
Being "principled" about using open-source tools can be so challenging, but its quite rewarding on the long run.
LEARNING SPANISH - What's working for me... trying to read spanish books before bed. Handwriting a few paragraphs from a book into journal. Highlighting words I don't know. Looking them up later. Reading a book while listening to its audio book at the same time.
If anyone's interesting in contributing to these projects, I would warmly welcome that. Design, product, sales, project management, engineering/coding, marketing - need tons of help in all these areas.
Gracias! // JRO
Oracle's plugin allows you access Fusion REST Endpoints (your business data) from within an Excel workbook but it only works on Windows machines and has some other limitations.
Also added a plugin for inspecting punchout payloads for RSSP [2]
RSS and podcast Google Readerish clone mostly for myself
It has one SQLite database per queue.
In golang.
Why? Because Rabbit is slow and resource hungry and needs configuration.
Problems I'm having: - Getting enriched vectors because the definitions to some of the words are absolute garbage - Finding a good open source embedding model, currently using nomic-embed-text
Goal: Find me words originating from X city and it not giving me results that match X
I’m a Senior Full Stack Engineer with over 8 years of experience building and scaling production systems using Node.js, TypeScript, React, and Python. I’ve worked in remote, product-focused environments where I’ve led architectural improvements, including migrating a monolithic system to microservices, reducing deployment time by around 50% and improving scalability and reliability.
I’m comfortable owning features end-to-end — from system design and API development to deployment, performance optimization, and production support. I’ve also implemented CI/CD pipelines, improved database performance (PostgreSQL), and contributed to cloud-native infrastructure on AWS using Docker and Kubernetes. In addition, I’ve worked on AI-driven workflows and LLM integrations for modern product capabilities.
I’m currently exploring new remote opportunities and would love to connect if you’re building or scaling a product where strong backend architecture, clean execution, and ownership matter.
If it makes sense, I’d be happy to schedule a short conversation. Thank you.