An LLM observability SDK that let's you store pre and post request metadata with every call in as lightweight an SDK as possible.
Stores to S3 in batched JSON files, so can easily plug into existing tooling like DuckDB for analysis.
It's designed to answer questions like; "how do different user tiers of my services rate this two different models and three different systems prompts?". You can capture all the information required to answer this in the SDK and do some queries over the data to get the answers.
A lot of existing databases are storage first, with everything else built around them. I have been exploring what it looks like if the database is closer to the application runtime itself, where state is live, queryable, and easier to reason about directly.
One thing I am prototyping right now is database-native tests.
Basically: what if integration tests were a database primitive?
CREATE TEST test::insert { INSERT test::users [{ id: 99, name: "Ghost" }]; FROM test::users | FILTER id == 99 | ASSERT { name == "Ghost" }; };
So not a wrapper, not a framework, not an external test runner.
A real test object inside the database.
The idea is that you could run these before schema changes, and make stored procedures or other database logic much easier to test without leaving the database model.
Still early, but it feels like one of those things that should just exist, especially for databases built around live application state.
It pulls a list of birds reported on eBird in your county in the last 2 weeks and you ask preselected questions like the the color or size to whittle down the possibilities. I also made a matching game that uses the same list and you have to match the name to a picture of the bird. I set it up for California for now. I wanted to get more comfortable with SQL and APIs.
Feedback welcome.
Today, if you search for "what size should I get in Nike Air Max 90" you'll find size charts. We have it, and for 200+ brands across 70+ retailers. When users tell us which shoes they own and what size fits them we’re slowly building crowdsourced fit recommendations which are personal and more accurate compared to size charts.
We're two coders who've built an almost fully autonomous platform. AI agents build, debug and deploy crawlers on their own. We went from 4 crawlers to 280+ in about a month, and the whole thing runs on a home server. When new shoes are discovered, the platform publishes new pages with relevant info automatically. Agents get access to platform metrics and SEO data via custom MCPs to identify the right opportunities on their own. Currently at about 3000 MAU and about 100 size recommendations/day.
Example: https://www.getsize.shoes/en/shoes/nike-air-jordan-1-low-se-...
I started developing a city builder called Metropolis 1998 [1], but wanted to take the genre in new directions, building on top of what modern games have to offer:
- Watch what's happening inside buildings and design your own (optional)
- Change demand to a per-business level
- Bring the pixel art 3D render aesthetic back from the dead (e.g RollerCoaster Tycoon) [2]
I just updated my Steam page with some recent snapshots from my game. Im really happy with how the game is turning out!
[1] https://store.steampowered.com/app/2287430/Metropolis_1998/
[2] The art in my game is hand drawn though
Supports multiple-accounts (track as a family or even as an advisor), multi-currency, a custom sheet/calculator to operate on your accounts (calculate taxes etc) and much more. Most recently, we added support for benchmarking (create custom dashboards tracking nav and value chart of subsets of your portfolio) and US stocks, etfs etc.
We also write about like:
How fund performance explain part of returns, rest is explained by timing. And ways to tease those out: https://finbodhi.com/docs/blog/benchmark-scenarios
Or, understanding double entry account: https://finbodhi.com/docs/understanding-double-entry
Also working on a language for embedded bare-metal devices with built-in cooperative multitasking.
A lot of embedded projects introduce an RTOS and then end up inheriting the complexity that comes with it. The idea here is to keep the mental model simple: every `[]` block runs independently and automatically yields after each logical line of code.
There is also an event/messaging system:
- Blocks can be triggered by events: `[>event params ...]`
- Blocks can wait for events internally
- Events can also be injected from interrupts
This makes it easy to model embedded systems as independent state machines while still monitoring device state.
Right now it’s mostly an interpreter written in Rust, but it can also emit C code. I’m still experimenting with syntax.
Example:
module WaterTank {
type Direction = UP|DOWN
let direction = UP
let current = 0
[>open_valve direction |> direction]
[>update level |> current]
[
for 0..30 |> iteration {
when direction {
UP -> !update level=current + 1 |> min(100)
DOWN -> !update level=current - 1 |> max(0)
} ~
%'{iteration} {current}'
}
]
[>update level |> when {
0..10 -> %'shallow'
11..15 -> %'good'
16.. -> %'too much!' then !open_valve direction=DOWN
}
]
}I posted about it recently on HN (https://news.ycombinator.com/item?id=47199062):
It is at a fairly early stage of development, so it's quite rough around the edges. It is developed and hosted in EU.
I have started developing it as a slim wrapper around Git to serve my own code, but it grew to such extent that I decided to give it a try and offer it as a service. It doesn't have much at the moment, but it already has basic pull requests. Accessibility is high priority.
It will be a paid service, (free for contributors) but since it's an early start, an "early adopter discount" is applied – 6 months for free. No card details required.
I would be happy if you give it a try and let me know what do you think, and perhaps share what you lack in existing solutions that you would like to see implemented here.
while word_count < x: write_next_chapter(outline, summary_so_far, previous_chapter_text)
It worked well enough that the novels were better than the median novel aimed at my son's age group, but I'm pretty sure we can do better.
There are web-based tools to help fiction authors to keep their stories straight: they use some data structures to store details about the world, the characters, the plot, the subplots etc., and how they change during each chapter.
I am trying to make an agent skill that has two parts:
- the SKILL.md that defines the goal (what criteria the novel must satisfy to be complete and good) and the general method
- some other md files that describe different roles (planner, author, editor, lore keeper, plot consistency checker etc.)
- a python file which the agent uses as the interface into the data structure (I want it to have a strong structure, and I don't like the idea of the agent just editing a bunch of json files directly)
For the first few iterations, I'm using cheap models (Gemini Flash ones) to generate the stories, and Opus 4.6 to provide feedback. Once I think the skill is described sufficiently well, I'll use a more powerful model for generation and read the resulting novel myself.
Im also building https://www.keepfiled.com, a microsaas to save emails (or email attachments) to google drive
I almost forgot, I also built https://statphone.com - One emergency number that rings your whole family and breaks through DND.
I love building. I built all these for myself. unfortunately I suck at marketing so I barely have customers.
The content is hand picked from tiktok, Instagram, Facebook, Reddit and other AI generating platforms.
Honestly I don't know where I'm going with this, but I felt the urge to create it, so here it is.
I learned how to optimize serving assets on CloudFlare.
Feedback welcome.
Each game adds more building blocks to the editor: multiplayer, event systems, NPC behaviors, pathfinding, etc. I build a system once, and then anyone using the editor can use it in a click.
Since my last month, I shipped the asset marketplace and the LLM builder. Artists can now upload tilesets and characters, and unlike itch.io, assets drop directly into the editor. You can preview how they'll actually look in-game before using them [1].
An other problem I kept running into: even with a no-code editor, users don't know where to start. So now I'm extending it with a coding agent. Describe the game you want, and it assembles it — pulling assets from the marketplace, wiring up the event system, and using all the building blocks I've spent the past year extracting. Multiplayer, mobile controls, pathfinding, NPC behaviors — the agent doesn't build any of it, just reaches for what's already there.
Once the LLM assembles it, users will have a game ready to work on, and will still be able jump into the editor and tweak everything [2]. Here's an example of what it can already make [3] (after a lot of prompting), and the goal is to reach games like this one I built with the manual editor[4].
Hoping to release the AI mode in a week or two. The manual editor is live at https://craftmygame.com in the meantime.
[1] https://craftmygame.com/asset/mossy-cavern-JdYWai1
[2] https://youtu.be/6I0-eTmoHwQ
We use landing.ai to parse the PDF, as well as useworkflow.dev to durably perform other work such as rendering PDF pages for citations, and coordinating a few lightweight agents and deterministic checks that flag for inconsistencies, rule violations, bias, verify appraiser credentials, etc. etc. Everything is grounded in the input document so it makes it pretty fast and easy. We’re going to market soon and have an approval sign up gate currently. Plenty of new features and more rigorous checks planned to bring us to and exceed parity with competition and human reviewers.
There’s plenty of margin for cost and latency versus manual human review, which takes an hour or more and costs $100 or more.
It's an addictive slot machine where I pull the lever and the dials spin as I hope for the sound of a jackpot. 999 out of 1000 winning models do so because of look-ahead bias, which makes them look great but are actually bad models. For example, one didn't convert the time zone from UTC to EST, so five hours of future knowledge got baked into the model. Another used `SELECT DISTINCT`, which chose a value at random during a 0–5 hour window — meaning 0–5 hours of future knowledge got baked in. That one was somehow related to Timescale hypertables.
Now I'm applying the VIX formula to TSLA options trades to see if I can take research papers about trading with VIX and apply them to TSLA.
Whatever the case, I've learned a lot about working with LLM agents and time-series data, and very little about actually trading equities and derivatives.
(I did 100% beat SPY with a train/out-of-sample test, though not by much. I'll likely share it here in a couple weeks. It automates trading on Robinhood, which is pretty cool.)
Why? Many yarncrafters painstakingly build spreadsheets, or try to bend existing general purpose pixel editors to their will. It's time consuming & frustrating.
Along the way, I've solved a bunch of problems:
- Automatic decreases (shapes the hat) / overstitching markers (shows when multiple colors are used in the same row)
- Parameterized designs, like waves, trees, geometric shapes. No more manually moving an object by a couple of pixels, it's a simple click & drag.
- Color palette merging (can't delete a color if you already use it in a pattern!)
- Export to PDF (so you can print it or stick it on a tablet)
- Repeat previews (visualize the pattern as it repeats horizontally)
The core feature that makes this more useful than most general purpose editors is that the canvas is continuous.If you drag a shape near the right edge of the canvas, you'll see it "wrapping around" onto the right edge.
This reflects the 3D reality of a hat!
It just won an award! It was awarded Players' Choice out of 700 daily web games at the Playlin awards: https://playlin.io/news/announcing-the-2025-playlin-awards-w...
Right now around 3,500 people play every day which kind of blows my mind!
It's free, web-based, and responsive. It was inspired by board games and crosswords.
I've been troubleshooting some iOS performance issues, working on user accounts, and getting ready to launch player-submitted puzzles. It's slow going though because I have limited free time and making the puzzles is time consuming!
Here's an article with more info about the award: https://cogconnected.com/2026/03/tiled-words-crowned-the-pla...
Open-source plugins for Ghidra, Binary Ninja, and IDA Pro that bring LLM reasoning, autonomous agents, and semantic knowledge graphs directly into your analysis workflow.
Coming soon: A supporting online service. The VirusTotal for reverse engineering. A cloud-native symbol store and knowledge graph service designed for the reverse engineering community.
- Submit files for automated reverse engineering and analysis
- Query shared symbols, types, and semantic knowledge
- Accelerate analysis with community-contributed intelligence
- Versioned, deduplicated symbols with multi-contributor collaboration
The front bump out leaks when we get driving rain. I installed some flashing but that wasn't enough, it's still leaking. So I'm working on that so I can close up the big hole in the ceiling some day.
The prior owners filled in the old coal chute with literal bags of cement sort of artistically placed in the hole in the brick foundation. So I'm trying to figure out what masonry tools and skills I'll need to close it up proper.
I'd like to build my kids a playhouse of some sort, sketching out some designs for that.
https://breaka.club/blog/why-were-building-clubs-for-kids
The recent Netflix Games edition of Overcooked with K-Pop Demon Hunters is cool, but not nearly as cool as kids coding and playing their way through Overcooked levels in our custom educational mod for Overcooked:
I'm also maintaining GodotJS, strongly typed TypeScript bindings for Godot, which is used to build the Breaka Club RPG (see first link):
https://github.com/godotjs/GodotJS
And last week I also put together the first release of MoonSharp in ~10 years; Lua runtime for Unity. That's not for Breaka Club though, I also consult for Berserk Games on Tabletop Simulator:
I've wanted this for a long time, so I finally started building it. I've had a lot fun!
- Graph-based signal flow: Products become nodes, connections are edges inferred from port compatibility (digital, analog, phono, speaker-level domains)
- Port profile system: Standardized port definitions (direction, domain, connector, channel mode) enable automatic connection inference
- Rule engine: Pluggable rules check completeness, power matching, phono stage requirements, DAC needs, and more
working on an AI-native Kubernetes sidekick that watches your pods, reads the logs, and turns failures into clear fixes before they become outages
As many here, I've found that a single text file is all that I really need, but found that it makes it difficult to keep track of a variety of things. I was also trying to use the file as a simple project tracker, adding some tags like [BUG-N], and updating them by hand. Eventually, it became difficult to track the progress of things, since I had to jump around the file to look for updates.. or use grep.
I condensed the idea to just that - a very simple tool which manages "trackers", and has a simple filtering built in to "trace" the updates. I've been using it, since I've added the BE, and dogfooding it a bunch. Would love for fellow note takers to take a look. It's not perfect, but I'm keeping it around for myself :)
I have no illusions that this is actually something in capable of building to an actual release-able state but it’s fun to tinker with.
I recently converted a bunch of stuff to be client side instead of server side (turns out running a real-time MMORPG server is expensive) so there's a new round of bugs I'm still resolving, but it's still fun to play:
The system allows users to submit a JSON payload containing geocoordinates and mission requirements (e.g., capture_type: "4K_video" | "IR_photo"), the backend then handles the fleet logistics, selecting the optimal VTOL units from distributed sub-stations based on battery state-of-charge and proximity.
I've been building a collaborative docs tool called Docules. The short version: it's a team documentation tool that doesn't have any embedded AI features. I use Claude Code daily, but putting LLMs into every workflow and charging for it is kinda insane. Every docs tool is adding AI auto-complete, AI summaries, "generate a page" buttons. Docules has an API and an MCP server instead, so you connect whatever AI tools you actually want to use. The core product focuses on being a fast, solid docs tool. Real-time collab, fast — no embedded databases or heavy view abstractions, hierarchical docs, drag-and-drop, semantic search, comments, version history, public sharing, SSO, RBAC, audit logs, webhooks, etc. The stack is React, Hono, PostgreSQL, WebSockets. The MCP server is a separate package that exposes search, document CRUD, and comments — so Claude/ChatGPT can work with your docs without us reimplementing a worse version of what they already do. Happy to talk architecture or the MCP integration.
It’s like netflix for language, where users can select/create their personal bilangual stories.
I had quite a lot of feedback from HN, friends, random people on the internet and trying to solve the common pain points and find my way around to make it geniunely useful.
- Most people said it’s hard to come up with a story, so I added url grounding. Also added buttons (including HN :)) so people can just click click and get their stories at their level with their interests.
- Made sure people can generate stories without ever signing up
- Each word is highlighted while being read, and the meanings can be checked with a tap. I also added an option for users to read the sentence for being checked how good their pronounciation is.
- Benchmarked 7 different models to get the fastest & highest quality story generation (it’s gemini now) and it’s insanely fast. I might share more about it on the webpage because I am an engineer and I enjoy this stuff lol.
- Added CSV import in Use my words so Anki users can just import their words to study.
- Also people can download their stories as pdf so they can send it to their kindles.
- I am working on a ChatGPT app, so people can just say “@DuoBook give me a Dutch/English story on latest Iranian events” within ChatGPT, but I am a bit afraid that it might be costly lol.
One of the issues I encountered initially was that the LLMs were repeating a small set of actions and never trying some of the more experimental actions. With a bit of prompt tweaking I was able to get them to branch out a bit, but it still feels like there's a lot of room for improvement on that front. I still haven't figured out how to instill a creative spark for exploration through my prompting skills.
It has been quite exciting to see how quickly a few simple rules can lead to emergent storytelling. One of the actions I added was the ability for the agents to pray to the creator of their world (i.e. me) along with the ability for me to respond in a separate cycle. The first prayer I received was from an agent that decided to wade into a river and kneel, just to offer a moment in stillness. Imagining it is still making me smile.
Unfortunately, I don't have access to enough compute to run a bigger experiment, but I think it would be really interesting to create lots of seed worlds / codebases which exist in a loop. With the twist being that after each cycle the agents can all suggest changes to their world. This would've previously been quite difficult, but I think it could be viable with current agentic programming capabilities. I wonder what a world with different LLM distributions would look like after a few iterations. What kind of worlds would Gemini, Claude, Grok, or ChatGPT create? And what if they're all put in the same world, which ones become the dominant force?
Also working on https://www.kinoko.sh/. An agentic engineering platform built from the ground up for agents. Custom language and architecture and a later of formal verification on top. Also working on a custom inference engine that produces well typed programs always
A hobby project I started putting together late last year; a little spot on the internet for prayer and reflection. I've just shipped a small feature where you get a Bible reading (KJ only for now) in response to a prayer.
https://dugnad.stavanger-digital.no/
A pro bono tech consultancy for local (Stavanger, Norway) non profits. The idea is to help them use tech to better deliver on their mission. Last week I built a little bookmarklet for a non-profit to surface some of their data buried in a SaaS tool ... which will make their apple pressing operation easier.
Give it a try!
I have built npm for LLM models, which lets you install & run 10,000+ open sourced large language models within seconds. The idea is to make models installable like packages in your code:
llmpm install llama3
llmpm run llama3
You can also package large language models together with your code so projects can reproduce the same setup easily.
I also intend to dig into how to integrate Emacs with tools such as yt-dlp and patreon-dl to grab Latin-language audio content from the Internet, transcode the audio with ffmpeg, load it into the LLM's context window, and send it off for transcription. If the essay isn't already too long, I'll demonstrate how to gather forced-alignment data using local models such as wav2vec2-latin so I can play audio snippets of Latin texts directly from a transcription buffer in Emacs. Lastly, I want show how to leverage Gemini to automatically create multimedia flash cards in Org mode using the anki-editor Emacs minor mode for sentence mining.
It started because my wife watches Chinese dramas and new episodes never have subtitles for our language. Turns out thousands of people have the same problem — Arabic speakers watching anime, Russian speakers following Turkish series, Persian speakers catching up on K-dramas.
Supports 40+ languages, works with any video link or direct file upload. There's also a Mini App inside Telegram for a more visual experience.
The result is an experiment called fesh. It works strictly as a deterministic pre-processor pipeline wrapping LZMA (xz). The AI kept identifying "structural entropy boundaries" and instructed me to extract near-branches, normalize jump tables, rewrite .eh_frame DWARF pointers to absolute image bases, delta-encode ELF .rela structs with ZigZag mappings, and force column transpositions before compressing them in separated LZMA channels.
Surprisingly, it actually works. The CI strictly verifies that compression is perfectly reversible (bit-for-bit identity match) across 103 Alpine Linux x86_64 packages. According to the benchmarks, it consistently produces smaller payloads than xz -9e --x86 (XZ BCJ), ZSTD, and Brotli across the board—averaging around 6% smaller than maximum XZ BCJ limits.
I honestly have no idea how much of this is genuinely novel versus standard practices in extreme binary packing (like Crinkler/UPX).
Repo: https://github.com/mohsen1/fesh
For those who know this stuff:
Does this architecture have any actual merits for standard distribution formats, or is this just overfitting the LZMA dictionary to Alpine's compiler outputs? I'd love to hear from people who actually understand compression math.
Over the last year I've been hacking on Table Slayer [0] a web tool for projecting DnD maps on purpose built TV-in-table setups. Right now I'm working on making hardware that supports large format touch displays.
Since I also play boardgames, this past month I threw together Counter Slayer [1], which helps you generate STLs for box game inserts.
Both projects are open source and available on GitHub. I've had fun building software for hobbies that are mostly tactile.
[1] https://www.crowdsupply.com/scope-creep-labs/hoopi-pedal [2] https://scopecreeplabs.com/blog/?tag=hoopi
It's like a carfax but for your home, although the intention is more to create an interesting historical narrative that inspires people to care about the history of their home rather than as a tool for inspecting home issues before buying.
My target customer is realtors who want to inspire buyers to take on historic homes that may need a lot of work. Also home owners themselves of course.
I was stuck on this conversation problem. First version had a dead-end search box: six starter prompts, one referencing a tool that didn't exist. No follow-ups. No guided flows. Users got an answer and had to invent the next question from scratch.
Now the assistant explores your library with you. Tag discovery, color browsing, weekly digests, smart collections that auto-curate as you save.
Semantic search runs hybrid, keyword matching plus pgvector cosine similarity on 768-dim embeddings. Streaming responses.
Almost there. https://bookmarker.cc/
https://github.com/nickbarth/closedbots/ I was also trying to do a simplified openclaw type gui using codex. The idea being its just desktop automation, but running through codex by sending codex screenshots and asking it to complete the steps in your automation via clicks and keypresses via robotgo.
https://milliondollarchat.com a reimagining of the million dollar homepage for the AI age. Not useful, but fun. A free to use chatbot that anyone can influence by adding to the context. The chatbot's "thoughts" are streamed to all visitors.
My favorite features are: - custom layout and drag and drop to change window - auto resume to last working session on app starting - notifications - copy and paste images directly to Claude Code/Codex/Gemini CLI - file tree with right click to insert file path to the session directly
OH and it works on both Windows and MacOS! Fully open source too!
Main gig: Trusted agents. We just shipped hardware based signing to web bot auth protocol.
We're constantly pulling info from official sources, and using AI to group and summarize into stories, and continue to share reporting from trusted, vetted journalists.
The result is news with the speed and breadth of getting updates straight from the source, and the perspective and context that reporting provides.
Still ramping up, but I'd love to hear feedback:
I've been building a collaborative docs tool called Docules. The short version: it's a team documentation tool that doesn't have any embedded AI features. I use Claude code daily, but putting LLM’s into every workflow and charging for it is kinda insane. Every docs tool is adding AI auto-complete, AI summaries, "generate a page" buttons. Docules has an open API and ships an MCP server, so it connects to whatever you want to use LLM-wise. They can read, search, create, and edit documents through the API. The core product is just a docs tool that tries to be good at being a docs tool:
- Real-time collab with live cursors
- Fast — no embedded databases or heavy view abstractions slowing things down
- Hierarchical docs, drag-and-drop, semantic search
- Comments, version history, public sharing
- SSO, RBAC, audit logs, webhooks
Stack is React, Hono, PostgreSQL, WebSockets. The MCP server is a separate package so it's not coupled to the main app. I keep seeing docs tools bolt on half-baked AI features and call it innovation. I'd rather build a solid foundation and let you plug in whatever AI workflow actually makes sense for your team. Happy to answer questions about the architecture or the MCP integration.[1]: https://github.com/loopmaster-xyz/loopmaster
[2]: https://loopmaster.xyz/tutorials
Multitrack field recorder with automatic cloud sync for iPhone. I use it for hi-fi recording of band practice and sharing demos with bandmates/collaborators. Great way to send stems too as it runs on the Mac as well and has a built in mixer. There's a social graph so you can send someone a session by typing in their handle and granting access.
The stock firmware is horrible but the community has this firmware called CrossPoint. I wanted to be able to upload, manage files etc. from my iPhone on the go and also send over web articles. So I build this app CrossPoint Sync https://crosspointsync.com to do just that.
I've already published it on App Store and pending publishing on Android. The community is niche and has also been using the app, so its been fun building for my use and in turn also getting good feedback from community.
If you are using the Xteink and CrossPoint firmware, then give the app a try.
iOS App Store: https://apps.apple.com/app/crosspoint-sync/id6758985427
Android Beta: https://crosspointsync.com/android/join-beta
The idea is pretty simple: SQLite is amazing, but once it’s running in production you basically have zero observability. If something weird happens (unexpected writes, schema changes, background jobs touching tables, etc.) you only find out after the fact. It tries to solve that without touching application code. It's a Rust agent that runs next to your sqlite file, and connects to the server where everything is logged in. My current challenge right now is encryption and trust, mostly.
Curious if others here are running SQLite in production and if you would be interested in something like this.
I originally made it a couple of years ago as a small proof of concept. A couple of weeks ago I started it over and have been using it as a project to work with Claude and learn approaches to coding with AI.
It's been a lot of fun.
Info (not recent) available here: https://awz.us/docs
https://greenmtnboy.github.io/sf_tree_reporting/#/
For all the places it's bad at, AI has been fantastic for making targeted data experiences a lot more accessible to build (see MotherDuck and dives, etc), as long as you can keep the actual data access grounded. Years of tableau/looker have atrophied my creativity a bit, trying to get back to having more fun.
It is a pretty fun project
https://feedbun.com - a browser extension that decodes food labels and recipes on any website for healthy eating, with science-backed research summaries and recommendations.
https://rizz.farm - a lead gen tool for Reddit that focuses on helping instead of selling, to build long-lasting organic traffic.
https://persumi.com - a blogging platform that turns articles into audio, and to showcase your different interests or "personas".
All data lives in your browser (IndexDB) - https://buyitlater.vercel.app
I have some barebones content at https://struggle-meals.wonger.dev/ and will be working on the design over the next few weeks. Some decisions I'm thinking about:
- balancing between personal convenience and brevity vs being potentially useful for other people. E.g. should I tag everything that's vegan/vegetarian/GF/dairyfree/halal/etc? Should I take pictures of everything? (I'd rather not)
- how simple can I make a recipe without ruining it? E.g. can I omit every measurement? should I separate nice-to-have ingredients from critical ingredients? how do I make that look uncomplicated? (Sometimes the worst thing is having too many options)
- if/how to price things? Depends on region, season, discounts, etc
^^ project with my daughter
Parallel agents debate your ideas/work: https://murderboards.ai
^^ solo project / code review agents for non-coders / inspired by Compound Engineering plugin’s code review flow
https://github.com/skorokithakis/stavrobot
It's like OpenClaw but actually secure, without access to secrets, with scoped plugin permissions, isolation, etc. I love it, it's been extremely helpful, and pairs really well with a little hardware voice note device I made:
Also moving to Sveltia as my CMS (Astro markdown blog), after exploring multiple other options. Changed the structure of my Obsidian vault, will write about that also.
I’m also still working on a few projects:
- https://game.tolearnkorean.com/
PS - The results are entirely obvious.
At this moment I’m working on improving the logic that decides when/how much to throttle the network.
* Telephone handset for my mobile phone with side talk.
* First draft of a book / workbook on Work Flow. Outcrop of the work flow consulting I do, stuff I've learned, and so on.
* Short film script - trying to convince a local actor to play the lead before we lose the rainy season here - otherwise we'll need special effects or just wait until the fall.
* Polishing firmware, OSX, and iOS suite for a wearable neuromodulator unit. Deadline in a week!
* Nmemonic community and app - been poking at this for years and finally had a breakthrough on the UI. My first app to release in the wild, so pretty exciting.
Check out this twisty vase demo: https://nodillo3d.com/s/VmP0nJdKRcPazQ1g
You can also share you files and create sharable configurations as well. Here is the same vase as a configurator: https://nodillo3d.com/v/a9REIEZIDYGtzZRA
I would like to do a more detailed intro class to help people learn how to model with nodes.
Hope you enjoy it!
While the main app is closed sourced, the rails engine that handles all the rss feeds is open sourced here: https://github.com/dchuk/source_monitor
I have another version of source monitor getting by published soon with some nice enhancements
Also used the new Navigation API (and some Shadow DOM) to build a cheap, custom client-side rendering (sort of) into my site (https://taro.codes), and some other minor refactors and cleanup (finally migrated away from Sass to just native CSS, improved encapsulation of some things with Shadow roots, etc).
I've been wanting to write a simple AI agent with JS and Ollama just for fun and learning, but haven't started, yet...
- Tablex (https://www.tablex.pro) - seat arrangement app for weddings, seminars, conferences.
- Kardy (https://www.kardy.app) - group card app I've always wanted to build.
- Jello (https://www.jello.app) - Create games with your own photos and sound effects!
It was inspired by tamagotchis of yesteryear (and my two cats). It uses a small common monochrome SSD1306 display with 128x64 pixels of resolution.
All of the pixel art is my own. And the cat features a bunch of different animated poses and behaviors, as well as different environments. And there are minigames (a chrome dino clone - but with a cat!, a breakout clone, a random maze generator, a tic-tac-toe game, and I plan to add more.)
I'm currently working on tweaking the stats so that they go up and down over time in a realistic way and encourage the player to feed and interact with the pet to keep stats from going too low. Then I plan on adding some wireless features, like having the pet scan WiFi names to determine if its home or traveling, or using ESP-NOW to let pets communicate with each other when they're nearby.
I made a reddit post with a video of it a few weeks ago [1] and have various prototypes of artwork for these little screens on my blog [2].
[1] https://www.reddit.com/r/arduino/comments/1r8i1vx/progress_o...
https://www.inclusivecolors.com/
The current web tool lets you export to CSS, Tailwind and Figma, and uses HSLuv for the color picker. HSL color pickers that most design tools like Figma use have the very counterintuitive property that the hue and saturation sliders will change the lightness of a color (which then impacts its WCAG contrast), which HSLuv fixes to make it much easier to find accessible color combinations.
I'm working on a Figma plugin version so you can preview colors directly on a Figma design as you make changes. It's tricky shrinking the UI to work inside a small plugin window!
First release was in December for 1D cuts. Last month I released sheet cutting for 2D cut calculation. It's been working well for my own projects and it started getting consistent daily users since my last update in February. You can save projects now on the site for you to come back to later.
Any feedback is welcome. I'm always looking for what features to add next.
i have also started experimenting with qwen3.5 0.8B model, my goal is to create agents with small models that are as robust as their commercial counterparts for specialized tasks. currently trying it for file editing.
Coffee Roaster Aggregation ETL using fastapi, nextjs, bs4 etc etc. It's been fun, just finished up the oauth for discord that pairs nicely with the info required to make Discord dm notifications function. attempting to charge 6$ for the instant notifications, but doubt many people will be interested. up to 75 roasters and all of them are checked every 10 mins for new products.
Considering reusing the repo as a framework for other industries if this project ever gains any traction. Also was considering adding a goofy rag discord bot to the server just because i love tossing in a rag layer everywhere lately, and feel like i fall a bit short on my filters for stuff like origin/flavor notes and all that junk. Semantic search with solid chunk strategies might create a better solution than if i did get all the filters working as well as possible.
It basically pulls OpenTelemetry data from your infra and renders chernoff faces, so you can spot anomalies at a glance.
Currently adding support for exposing Postgres schemas for each app to use. The goal is that with a shared Postgres instance, each app should be able to either get a dedicated schema or get limited/full access to another app's schema, with row level security rules being supported.
Run your agents contained (container or VM, Linux or Mac), with all restrictions removed.
Workflow:
┌─ YOLO shell ──────────────────────┬─ Outer shell ─────────────────────┐
│ │ │
│ yoloai new myproject . -a │ │
│ │ │
│ # Tell the agent what to do, │ │
│ # have it commit when done. │ │
│ │ yoloai diff myproject │
│ │ yoloai apply myproject │
│ │ # Review and accept the commits. │
│ │ │
│ # ... next task, next commit ... │ │
│ │ yoloai apply myproject │
│ │ │
│ │ # When you have a good set of │
│ │ # commits, push: │
│ │ git push │
│ │ │
│ │ # Done? Tear it down: │
│ │ yoloai destroy myproject │
└───────────────────────────────────┴───────────────────────────────────┘
Sandboxes: Docker (Linux, Mac), Seatbelt (Mac), Tart (Mac)Everything's contained in a single go binary. Just build and run.
Full encryption for notes (uses local encryption before you even sent the note to the server).
I wanted a mixture of Github Gists (sans Git) and 1Password shares so I've been using it eitj great success at my current company to share snippets and private stuff.
Might open source in the future, just need to gauge interest.
Already launched biz-in-a-box.org and a life-in-a-box.org spinoff as frameworks to replace every entity's QuickBooks. I'm using them myself for every project my agents are spinning up.
Stealth project is related to classpass but for another category of need that won't go away even in the age of AI that really is only possible with critical mass of supply to meet existing demand. Super excited cus there's no better time to build with unlimited agents that scale without people problems.
Lastly, can't wait to run local LLMs so no longer limited by tokens/money.
This is a "full rewrite," because I need to migrate away from my previous server, which was developed as a high-security, general-purpose application server, and is way overkill for this app.
Migration is likely to take a couple more years, but this is a big first step.
I've rewritten the server, to present a much smaller API. Unfortunately, I'm not yet ready to change the server SQL schema yet, so "behind the curtain" is still pretty hairy. Once the new API and client app are stable, I'll look at the SQL schema. The whole deal is to not interfere with the many users of the app.
I should note that I never would have tried this, without the help of an LLM. It has been invaluable. The development speed is pretty crazy.
Still a lot of work ahead, but the server is done, and I'm a good part of the way through the client communication SDK.
I've given myself 6 months
It's a bit scary basically 180ing like this but I figure if I don't try it now I never will
I've already started prototyping various ideas, and to be honest just sitting down and spending time doing this has been really quite lovely
One thing I'm finding fun is slowly unearthing what I actually find interesting
I started with messing around in minecraft and tinkering with rimworld-like game ideas, but I'm slowly moving away from them as I've been tinkering more and more
Don't get me wrong, I do want to revisit them at some point in the future, but I do find myself circling more around narrative, simulations and zachlikes
It's a bit of an odd mix and in some ways they look like paradox style games, but I'm well aware that taking one of those behemoths on is going to be a bit silly, so I'm trying to slim down until I get to a kernel that I actually find enjoyable tinkering with
A toy if you will
Currently I'm trying to work out if there's anything interesting in custom unit design, basically unpicking how games like rollercoaster tycoon's coaster design maps to stats like excitement ratings and seeing how that might mix with old school point buy systems
It feels like it might be small enough to be a good toy and I'm having fun tinkering with it, but I have no idea whether other people will xD
It might honestly be too niche for anyone and I've successfully optimised for an audience of one :shrug:
I've also used tweakcc to make this work in Calude Code and plan to also do one for open source coding agents - codex, pi, Gemini, etc. And I'm also doing Livestreams of the development process.
So I built a test for the same and its called Orlog. You can check it out here: https://orlog-test.netlify.app/
Looking for your feedback.
* https://sprout.vision/ - AI generated Go-To-Market Strategy for launching your next venture. I have a Tech background with limited GTM experience, so I experimented with AI to learn about different strategies and decided to turn it into a simple product that will generate a comprehensive plan (500+ pages) to help you launch your next venture. Try it out, would love to hear your feedback, use the HN50 promo code for 50% off your order.
* https://pubdb.com/ - Reviving a 10 year old project, it’s meant to make research publications more accessible to mere mortals with the help of AI. I have lots of ideas I want to try out here but haven’t gotten around to it yet. Currently focused on nailing down the basics with an OCR indexing pipeline and generating AI summaries.
Agents can search for design inspiration from production websites using semantic search. Since this inspiration comes from live websites, their design tokens; colors, typography usage, layout data are also available.
The problem I kept seeing: freelancers have happy clients but almost no testimonials on their site. Asking is awkward, clients say "sure!" and then never write anything.
SocialProof gives you one shareable link. Client clicks it, fills a short form (name, text, optional photo), you approve it, it embeds anywhere. No login required for the client.
The interesting technical bit: it's entirely on Cloudflare Workers + D1 + Pages. The collection form and embed widget are edge-served globally with no origin server. Been curious whether anyone else is building purely on Cloudflare's stack and what they've run into.
Still pre-revenue (just launched today). If you're a freelancer or run a small agency and have thoughts on how you currently handle testimonials, I'd genuinely love to hear it.
Some of these are present here: https://github.com/vamsipavanmahesh/claude-skills/
Planning to package this as a workshop, so companies could be benefit from AI Native SDLC.
Put together the site yesterday https://getainative.com
Couple of the people I have worked with in the past agreed to meet me for a coffee, will pitch this. Fingers crossed.
Under the hood it uses a cool legal reasoning agent primarily designed for understanding litigation claims and objectives.
Sometimes I do wish I had a slack channel of like 30 attorneys so I can ask them questions and get feedback.
(Sign up and I’ll send out beta codes tomorrow!)
I’ve had a few friends call it “fun”, and one said it’s “Scrabble that doesn’t drag”.
Working on some social media shareable replays you can post after matches tonight, thanks to Remotion:
* https://theblue.social — TheBlue.social, provides Bluesky native tools
* https://stacknaut.com — Stacknaut, SaaS starter kit to build on a solid foundation with AI, includes provisioning on Hetzner, deployment with Kamal 2 and dev with coding agents
* https://codevetta.com — Codevetta, Architecture and code reviews service
* https://myog.social — MyOG.social, OG Image Generation Service
I've been planning a new idea with that and possibly future ideas based on the future (and near future) where there are more and more "agent" users.
I had been doing lots of time-based work for a blog post and ended up annoyed that so many clocks around me were visually out of sync. Especially my microwave and oven clocks. Using the tool I got them synced up beyond what I could perceive.
It's a collection of 40 (and growing) tools for text processing, data cleaning, conversions, dev utils etc. Everything runs in the browser and is completely free.
Started this partly to learn SEO from scratch on a fresh domain, partly because i am lazy with regards to doing basic data cleaning using pandas and i found myself repeatedly using similar online tools that are completely riddled with ads.
I built this using Flask + Vanilla JS. I don't think there was any need to overcomplicate it. And for fun, i vibe coded a windows 95 desktop mode where all the tools open as draggable windows. https://textkit.dev/desktop
I just released support for dashboards. I've kept a devlog for the past 6 months.
And the biggest update is coming soon, DB Pro Cloud, which will let you connect to and manage any database through your browser as well as collaborate with your team.
Imagine Postman but for databases.
It's gone a long way to solve the "review" bottleneck people have been experiencing (though admittedly it doesn't fix all of it), and I'm in the process of adding support for Mac and Windows (WSL for now, native some other time).
Some of the features I've had for a while, like multi-project agent worktrees, have been added as a part of the Codex App, so it's good to see that this practice is proliferating because it makes it so much easier to manage the clusterf** that is managing 20+ agents at once without it.
I'm feeling the itch to have this working on mobile as well so I might prioritize that, and I'm planning to have a meta-agent that can talk to Tenex over some kind of API via tool calls so you can say things like "In project 2, spawn 5 agents, 2 codex, 2 claude, 1 kimi, use 5.2 and 5.4 for codex, use Opus for the claudes, and once kimi is finished launch 10 review agents on its code".
- http://sharpee.net : Text Adventure authoring platform in Typescript
- https://github.com/ChicagoDave/tsf : A multi-target npm build tool
- https://devarch.ai : Claude Code guardrail workflow including hooks, agents, and skills
In progress:
- unnamed project to disrupt commercial site hosting including a new marketplace
Its been pretty fun cosplaying as an network engineer, and now I'm building out an Anycast network for a few ideas that I'm working on.
Its nothing too revolutionary or new, but I'm proud that I've built them from ground up and all running on my own infrastructure.
- DNS Authoritative Hosting - https://thelittlehost.com/dns/ - Quietnet - A family-focused internet filter - https://quietnet.app
I'm also getting ready to launch https://relaye.io, which was my personal tool I built to support my devops consultancy.
Version 3 will add more feed types: Podcasts, Mastodon, Twitter/X, Calendar, Reminders, Weather, Finance tickers and more.
It will have a new UI, new features like notifications and local transcripts and summaries and many quality of life improvements.
This was an excuse to ship a mobile app for the first time and get familiar with supabase.
After these last few bugs are fixed, its ready for a semi-public TestFlight with our friends who have kids.
https://x.com/ZDi____/status/2013655958027669958
Right now, I only have single speaker checkpoints (as per the old video). That will change soon.
Downloaded and parsed a bunch of the pgsql-hackers mailing list. Right now it’s just a pretty basic alternative display, but I have some ideas I want to explore around hybrid search and a few other things. The official site for the mailing list has a pretty clean thread display but the search features are basic so I’m trying to see how I can improve on that.
The repo is public too: https://github.com/jbonatakis/pginbox
I’ve mostly built it using blackbird [1] which I also built. It’s pretty neat having a tool you built build you something else.
It has a few core libraries built in rust with a web app and a terminal UI. Android app is in the works. The persistence layer is intended to be offline first using a CRDT with an optional sync server. I'm also trying to integrate "bring your own AI" assistants to help tweak recipes or make suggestions.
It's been a fun way to sharpen my claude skills but also to see how feasible it is to maintain multiple frontend applications with a large amount of shared code. Still a lot to do, particularly the core calculations are not yet on par with existing offerings.
To date it's handled more than 70k orders, ingested nearly 10m telemetry records, has been extremely reliable, is almost entirely self-contained (including the routing stack so no expensive mapping dependencies) and is very efficient on system resources.
It handles everything from real-time driver tracking, public order tracking links, finding suitable drivers for orders, batch push notifications for automatic order assignment, etc.
Blog, news, chat, video, mail, web. Basically all the daily habits as little micro apps in one thing. I find it quite useful. Not sure anyone else does yet though.
Also separately worked on Reminder.dev which is a Quran app and API that bakes in LLM based search and motivational reminders.
I only got to the point of having code and data as \verbatim in \LaTeX. Next step is CWEB.
Here is an example (with C and Rust code in \verbatim)
https://ontouchstart.github.io/rabbit-holes/llm_rabbit_hole_...
The ultimate goal is machine and human readable proofs on algorithms.
A site for anti patterns in online discourse.
Example: https://odap.knrdd.com/patterns/strawman-disclaimer
Need to gather more patterns then create tooling around making it easier to use.
The goal is to raise the quality of comments/posts in forums where the intent is productive discussion or persuasion.
The server is a rust binary so you can toss it on any container/computer and connect to it in the app.
My philosophy isn't to replace my other tools I love like emacs, ghostty, etc. But I am taking a stab at "real time code review" and have some crummy magit-like code review built in that I need to revisit.
But once all the low level operations are done, my plan is to implement an A2A Agent as the sole Agent listed in the AgentCard at $SERVER_ROOT/.well-known/agent-card.json, which is itself an "AgentListerAgent". So you can send messages to that Agent to receive details about all the registered Agents. Keeps everything pure A2A and works around the point that (at least in the current version) A2A doesn't have any direct support for the notion of putting multiple Agents on the same server (without using different ports). There are proposals out there to modify the spec to support that kind of scenario directly, but for my money, just having an AgentListerAgent as the "root" Agent should work fine.
Next steps will include automatically defining routes in a proxy server (APISIX?) to route traffic to the Agent container. And I think I'll probably add support for Agents beyond just A2A based Agents.
And of course the basic idea could be extended to all sorts of scenarios. Also, right now this is all based on Docker, using the Docker system events mechanism, but I think I'll want to support Kubernetes as well. So plenty of work to do...
My favorite output so far is that I asked it what life was and in a random stroke of genius, it answered plainly: "It is.".
It's able to answer simple questions where the answer is in the question with up to 75% accuracy. Example success: 'The car was red. Q: What was red? ' |> 'the car' - Example failure: 'The stars twinkled at night. Q: What twinkled at night? ' |> 'the night'.
So nothing crazy, but I'm learning and having fun. My current corpus is ~17mb of stories, generated encyclopedia content, json examples, etc. JSON content is new from this weekend and the model is pretty bad at it so far, but I'm curious to see if I can get it somewhere interesting in the next few weeks.
More context https://lepisma.xyz/2026/02/24/harp/
Zero ops S3 based log search: https://github.com/amr8t/blobsearch
1. Live Kaiwa — real-time Japanese conversation support
I live in a rural farming neighborhood in Japan. Day-to-day Japanese is fine for me, but neighborhood meetings were a completely different level. Fast speech, local dialect, references to people and events from decades ago. I'd leave feeling like I understood maybe 5% of what happened.
So I built a tool for myself to help follow those conversations.
Live Kaiwa transcribes Japanese speech in real time and gives English translations, summaries, and suggested responses while the conversation is happening.
Some technical details:
* Browser microphone streams audio via WebRTC to a server with Kotoba Whisper * Multi-pass transcription: quick first pass, then higher-accuracy re-transcription that replaces earlier text * Each batch of transcript is sent to an LLM that generates translations, summary bullets, and response suggestions * Everything is streamed back to the UI live * Session data stays entirely in the browser — nothing stored server-side
---
2. Cooperation Cube — a board game that rotates the playing field
Years ago I built a physical board game where players place sticks into a wooden cube to complete patterns on the faces.
The twist: the cube rotates 90° every round, so patterns you're building suddenly become part of someone else's board. It creates a mix of strategy, memory, and semi-cooperative play.
I recently built a digital version.
Game mechanics:
* 4 players drafting cards and placing colored sticks on cube faces * The cube rotates every 4 actions * Players must remember what exists on other faces * Cooperation cards allow two players to coordinate for shared bonuses * Game ends when someone runs out of short sticks
---
Both projects mostly started as things I wanted to exist for myself. Curious what people think.
inspired by the karpathy/twitter posts on running (semi) autonomous research loops, I build https://github.com/tnguyen21/labrat to be able to try and replicate some paper results over night. still early stages but I'm getting some use out of it already.
also spending a lot of time thinking about how you "close the loop" on software projects. right now figuring out how you can combine static analysis + review heuristics to let LLMs course correct the codebase when they over-engineer or produced unwieldy abstractions.
The original Python icloudpd is looking for a new maintainer. I’ve been building a ground-up Rust replacement with parallel downloads, SQLite state tracking, and resumable transfers. 5x faster downloads in benchmarks, single binary, Docker and Homebrew ready.
The basic idea is that when one failure fans out across 20 services, you often end up with 20 alerts and 20 separate investigations, even though there is really just one root cause. I’m using distributed tracing to build a live model of how errors propagate through the system, and then exposing that context directly at each affected service.
Longer term, I want this to become a very high-precision RCA engine. Right now I’m looking to try it with a few early design partners that already have a lot of tracing data, especially OpenTelemetry or Datadog APM users. I'll love to chat with some folks who would be willing to try it out!
The main thing I'm currently working on is a platform for organizing and discovering in-person events. Still not certain about the boundaries for "Phase 1", but I have a bunch of ideas in that space that I've been incubating for a while. One subset of features will be roughly similar to that app you've probably heard of that starts with 'M' and ends with 'p', but hopefully an improvement, at least for the right audience. But wait, there's more. :)
Currently building it; it's not public yet, so no link. Next month.
Thinking about how to grow the userbase is intimidating, but I think it might end up being fun.
It's designed to integrate with Maven projects, to bring in the benefits of tools like Gradle and Bazel, where local and remote builds and tests share the same cache, and builds and tests are distributed over many machines. Cache hits greatly speed up large project builds, while also making it more reliable, since you're not potentially getting flaky test failures in your otherwise identical builds.
https://videohubapp.com/ & https://github.com/whyboris/Video-Hub-App
The workflow: Upload doc → LLM extracts structured data → Generate new doc from template.
It’s API-first, includes webhooks, and is built to be self-hosted/self-provisioned for privacy. Still very much a WIP, but looking for feedback on the feature set and the extraction accuracy.
URL: https://fetchtext.io
Now shifting to established SaaS companies adding AI assistants to their existing products. Some of them literally have people reading chats full time, so they actually value the experience.
Building https://lenzy.ai - 2 paid customers, 2 pilots, looking for more and figuring out positioning.
Aside from that, I need to document and properly release one of the pieces that PAPER is relying on (some generic tree-processing code that makes operations on directory trees a lot nicer than with the standard library "walk"s), and work on others (in particular, a "bytecode archive" format for Python that speeds up imports for some projects, mainly by avoiding filesystem work at import time — I want to offer it as an install-time option in PAPER, and later have `bbbb` make wheels with the bytecode precompiled that way).
Phase 1: Download the student's code from their submitted github repo URL and run a series of extractions defined as skills. Did they include a README.md? What few-shot examples they provided in their prompt? Save all of it to a JSON blob.
Phase 2: Generate a series of probe queries for their agent based on it's system prompt and run the agent locally testing it with the probes. Save the queries and results to the JSON blob.
Phase 3: For anything subjective, surface the extraction/results to the grader (TA), ask them to grade them 1-5.
The final rubric is 50% objective and 50% subjective but it's all driven by the agent.
Several readers have asked for an easy way to get recommendations without working through long-form review articles.
Here's the first iteration of a simple recommender: https://bcmullins.github.io/reading/
It's written in Golang and acts as a simple desktop app that creates a web server and then opens the site in your default browser. This way it's easily multi-platform and can also be hosted as a SaaS for larger production houses.
If we can nail this one, then an entire Oxford-grade education in the same style.
- Urbanism Now - I run https://urbanismnow.com, a weekly newsletter highlighting positive urbanism stories from around the world. It’s been exciting to see it grow and build an audience. I'm thinking of adding a jobs board soon that'll be built in astro.
- Open Library - I’ve been helping the Internet Archive migrate Open Library from web.py to FastAPI, improving performance and making the codebase easier for new contributors to work with.
- Publishing project - I’m also working on a book with Lab of Thought as the publisher, which has been a great opportunity to spend more time working with Typst.
These projects sit at the intersection of technology, cities, and knowledge sharing, exactly where I’m hoping to focus more of my time going forward.
It’s a drop-in replacement for Redis written in Rust. Most if not all of your client code should work without issues. Outperforms in many areas and has more out of the box features like proto storage, raft/swim, and encryption at rest.
I’m pretty proud of it, and I hope you’ll give it a shot and open bug reports. :)
- No sign-up, works entirely in-browser
- Live PDF preview + instant PDF download
- Flexible Tax Support: VAT, GST, Sales Tax, and custom tax formats with automatic calculations
- Shareable invoice links
- Multi-language (10+) and 120+ currencies
- Multiple templates (incl. Stripe-style)
- Mobile-friendly
- QR Code Support: Add payment QR codes with any invoice-related information (payment links, UPI, contact details, custom data)
- Multi-Page PDFs: Seamless multi-page support with automatic pagination and page breaks
GitHub: https://github.com/VladSez/easy-invoice-pdf
Would love feedback, contributions, or ideas for other templates/features.
PS: e-invoice support is wip
Jive Data: https://jivedata.com
Financial and Investing data
Random Data Monster: https://randomdata.monster
Random Data (also available as a Google Sheets Add-on)
WhatIsMyIPAddress.Monster: https://whatismyipaddress.monster
A clean website to get your IP Address. Also available as an API.
Phone Monster: https://phone.monster
Caller ID, but on steroids
A problem that we had at my last startup was that we got stuck between not wanting to spend too much time on devops, and getting price gouged by Heroku.
We were too big for the deploy to a VPS type options like coolify, but too small to justify hiring a full time Devops.
Eventually a few of us had to just suck it up and learn Kubernetes properly. Was pleasantly surprised how elegant it all was.
I was surprised there wasn’t something that “just worked” and plugged into our Kubernetes cluster, made it user friendly, teams, roles, etc.
I went through the Software Architecture Patterns for Serverless Systems book, which I think it is fantastic. I learned a lot but I still had a lot of doubts to actually use the ideas in real life. So I started dissecting the companion framework, which is in written in Typescript. I have been going piece by piece and converting to Kotlin which I think it is more expressive (and fun) and it is allowing me to understand how everything fits together.
Typescript framework: https://github.com/jgilbert01/aws-lambda-stream
There is a Vulkan based browser which you can use to connect to the only public site so far, a playable breakout clone.
Very early days but will keep updating them & adding more.
Physical engineering is a huge welcome transition for me from what coding has become in the last couple years.
There’s something nice about the realities of creating a model, then printing it, then seeing that exact is too exact, then reprinting, then eight more times, and then that feeling when it all comes together properly.
A few weeks ago I was working on an adapter for an airbrush to use on a standard pancake air compressor. Learning to create threads in blender was really neat! I learned a lot about the physical construction of threads, something I have never put much thought into before.
AM3 - (Allied MasterComputer or Artificial Mind, version 3) - An attempt to make a symbolic AI that approaches the capacities of a LLM. An LLM makes variations on the same code and schedules those variations to play in "games". The results allow the LLM to make further changes.
It's heavily supported by Claude Code, but much fun.
Actually not built on this yet I think, but I could switch over, haven't made anything more of it since it's still a bit rough around the edges, and I keep finding various issues during actual usage: https://binschema.net/
It is based on 20+ years of experience maintaining a similar system in Perl.
It's on Hex.pm already, looking for people to test and comment!
As Codex would say:
Selecto is an open-source SQL query builder for Elixir that helps you generate complex queries from clean, domain-based configs. It supports advanced joins, CTEs, subqueries, and analytics-friendly patterns, with companion packages for LiveView interfaces (selecto_components) and code generation (selecto_mix). If your app is data-heavy, Selecto gives you SQL-level power without brittle hand-written query strings.
Now at 350k lines. Native and wasm binaries (you can try the limited wasm version online). Currently adding a full CPython test suite benchmark.
Just for fun, not trying to replace CPython here. Mainly to test the limits of current coding agents.
A project that I launched on HN that became a business. Simplescraper rode the no-code wave of a few years back ('instant structured data without parsing html').
Now working on increasing the surface area for AI agents: MCP support, screenshots API, and (experimentally) x402[1]
It has gained a little traction in Reddit and grateful for the several paying users currently giving me lots of feedback. One of the features is that you get to import your own font using any otf, ttf files. App is 100% native too written in SwiftUI, AppKit and UIKit.
I just wanted my own interpretation of an RSS Reader app, I have been a heavy user of both Reeder and NNW but the interface is just the same and I got bored a lot.
- Portable Secret (https://alcazarsec.github.io/portable-secret/) - self-contained HTML files that decrypt in the browser.
- Dead Man's Switch (https://alcazarsec.com/deadmanswitch) - sends messages when you stop checking in.
- Flare (https://alcazarsec.com/) - silent alert when your device is accessed without authorization.
It makes connecting user domains to your app easy and reliable at any scale. Each Approximated user gets the own globally distributed, managed cluster of servers with its own dedicated IPv4 address. Includes (unlimited) edge rule features, DDoS protection, webhooks, and more. Make a simple API call, tell the user to point an A record at the IP, and it’s connected to your app with its own SSL certificates.
Built/building with elixir and phoenix, which has been fantastic.
I put together a pretty basic portal clone. I think its pretty cool to see it come together, animations, level creation, portal jumps.
The basic hardware on the ds makes 3d pretty approachable. Ive found opengl overwhelming in the past. It seems like a fun platform to make games on, but idk if there is any active ds homebrew communities. Anyway sharing because i thought it was cool, hard to find anyone that seems to be to interested. I thought about getting a 3ds but they are surprisingly expensive now
The insight: the friction in getting testimonials isn't that clients don't want to help – it's that a blank "leave a review" box produces mediocre one-liners. SocialProof guides them through structured questions ("what was your situation before?" / "what changed?") so you get a compelling before/after narrative automatically.
Free tier: unlimited testimonials. Just launched and looking for feedback from anyone who deals with client testimonials.
I've been working on a solution to automate solar+battery use to arbitrage the market. I'm on a real-time utility plan but even if you're on TOU it can save you $1+ per day by strategically planning when to use the battery and when to conserve or charge the battery. So far it's limited to a few providers and only FranklinWH batteries but I'm eagerly looking for someone to help me get Powerwall support working and other ESS. It's open-source on GitHub as well.
Build enterprise grade applications (in Elixir) with AI the right way.
Secure. Scalable. Reliable.
Built based on a senior engineer's experience. Uses 10 years of battle-tested patterns, not just LLMs:
1. Uses algorithms over AI whenever possible.
2. No external library dependencies whenever possible.
3. Old school over shiny new toys. Use the right solution for the problem (Eg. SQL vs NoSQL).
Example : https://shorturl.at/We3dH
When training I like to have every day mapped out with how many miles to run, at what pace, etc as an event in my calendar. My actual workout gets uploaded into Garmin and Strava, but I always wanted it back in the calendar so I could see at a glance the consistency over time. It's been really fun to see other people use and get value out of something I built for myself.
I wrote about it here: https://jdsemrau.substack.com/p/pair-programming-superbill-w...
The goal is to get consistent synthesis to 450MHz such that I can use a narrower 256-bit instead of a 512-bit interface, while maintaining full bandwidth. I've got it working at an FMax ranging 440-490MHz, though there's still some edge cases I need to hammer out.
I want to treat my Downloads folder (or some other one) like an "Inbox" where I can just dump everything, and then the program knows where exactly in my (Johnny Decimal) file system the file should land.
There is a surprising amount of edge cases that can cause ChatGPT or others to misunderstand your pages. Some models can handle div based tables, some want alt tags but cannot understand title tags, etc.
I built the tool to check your site as close as possible to what a human would see and then compare it with LLM's.
It was a weird journey trying to tease this info out of the models, they will happily lie, skip checking sites or just make things up.
https://dnsisbeautiful.com - clean, ad free dns lookup tool.
https://evvl.ai - combination of Github Gists and AI output comparisons (evals)
https://finalfinalreallyfinaluntitleddocumentv3.com/ - free mac app to intelligently rename any kind of file (photos, videos, audio, text) based upon their contents.
For now it uses UX patterns to make it easy to remove uninteresting articles and keeps a record of your read and saved articles. All locally of course.
I’d like to make it into something we can share quality content with one another eventually. For now I’m focusing on making it good enough my entourage will want to use it
- look for feedback on the Freelance Rates calculator https://heygopher.ai/tools/freelance-rate-calculator
My kid played it, and didn't stop for 45 minutes so I think that's a win :-)
More movement than CS, less than quake
Focused on infiltration mode - one team stealing a briefcase back to base with the other on the defense
I do no tracking, no analytics, just help you cross the airgap between web and mobile app so you can send users to the right place (and track them however you deem necessary)
Platform for running web apps.
Single static binary and SQLite
lua for now (WASM future)
DEMO:
Lots of this is going to involve getting people more up to speed on CS, can't wait.
Check it out here : https://github.com/VoidenHQ/voiden
Export your Apple Health data directly to Markdown files in your iOS file system.
Open-sourced it at https://github.com/CodyBontecou/health-md.
Fun little vibe-coded app that has made a lot of users happy.
Yes, you can use your own API key as well.
I’ve already completed the research, business model, competitive analysis, feature set, branding, and the full UI (40+ screens).
The MVP/V1 is currently in development. When the V1 is ready I’m planning to do a Show HN with this account.
It's my first product. Any feedback or questions are very welcome, even if it's just based on the idea and the screenshots on the site, since the product isn’t available to try yet.
Modifications to my Land Cruiser j90
- LED daytime running lights / off-road LED light bar
- Winch
- Front left – tie rod end (both)
- Rear axle – pinion bearing (loud while driving)
- Right rear brake caliper – brake fluid leaking from the piston
- Boost chip (chip + turbo tee), Kill switch
No accounts required, all data is yours and lives on your computer.
Check it out: https://greentea.app
Wondering if there are other similar tools out there which people love, and why ChatGPT/Gemini/Claude won't let you do the same in their native apps.
https://www.ai-proof-careers.com/
Super annoyed by the "AI will take your jobs" hysteria, so I pulled BLS data and analyzed talks by AI researchers and a few industry folks, and ranked 900+ BLS jobs by AI resilience.
Product link: https://ziva.sh/
Shamelessly trying to attract new monthly sponsors and people willing to buy me the occasional pizza with my crap HTML skills.
Arch Asxent https://github.com/mikko-ahonen/arch-ascent - tool for analyzing large microservice networks with hundres of microservices and creating architectural vision for them, and steps to reach the vision
Anyone can learn to type fast - I think it just takes the right tools to make it interesting enough for the users to use daily
Golang inference engine from scratch that can run a bunch of models with vulkan acceleration.
Art search for magic cards
Automated personal outreach app for job seekers, integrated with Gmail.
button1 = create_button('Hello world!')
button1.on_clicked {
the_hello_world_button_was_clicked
}
# this is the verbose variant in a pseudo-DSL,
# I like things being explicit. In most code
# I may omit some parts e. g.
_ = button('Hello world!') { :the_hello_world_button_was_clicked }
It defaults to ruby and what ruby supports (including
jruby-swing) but two additional languages to use are
python and java. Anyway.I recently added the possibility to describe what kind of widgets are to be used via a yaml file, as an option. This may not sound like a huge win, but so far what I like here is that it becomes easier to modify individual widgets without having to sift through code; and it works for more programming languages too. Any customization for the widget, including method-invocations if necessary, can be done via a yaml file now. There is of course a trade off in that the yaml file can become a bit complex (if the GUI uses many widgets), so for the most part I use this for smaller widgets/components that do one specific functionality (or, few specific functionalities). For instance, a GUI over wget. Then if other larger programs need that, I make this small widget more useful and flexible.
The distant goal is to actually use a simple DSL that also would allow average Joe to customize everything in a very easy manner; and to have a widget set that can be used for as many different parts possible including wonky ideas such as having a whole operating system as a GUI available one day (a bit like webmin, but not limited to what webmin does; for instance, I'd also have games such as solitaire, reversi and so forth). I'd like to see how far that idea can go, but it is just a hobby so I can only invest little time into it.
I've been on/off working on a Forth compiler for the NES. It will be open source soon enough but I'm not happy with the code right now as it's extremely messy, repetitive, and buggy, but I think it's turning out ok. I am resisting the urge to use Claude to do all the work for me, since that's depressing.
I've also been working on a clone of the old podcasting website TalkShoe. It's nothing too complicated. It's mostly an excuse to learn a bit more about Asterisk and telephony stuff. I'm hoping to have something fully usable in about a month or two.
I forked the main MiSTer binary due to some disagreements I had with Sorg in how he's running things [1]. My fork was largely done by Codex and Claude, but the tl;dr of it is that it has automatic backup of your saves, tagging and versioning of your saves, and it abuses the hell out of SQLite to give better guarantees of write safety than the vanilla MiSTer binary gives you. I've been using it for a few weeks now and it seems to work fine, and it's neat to be able to tag and version saves.
I think that's mostly it. I'm always hacking on something so there might be a straggler there.
[1] https://github.com/Tombert/Main_MiSSus/blob/master/README.md
It is a forum application where each community is invite only. Think a cross between reddit/discord. The invite only architecture reduces trolls, spam, AI slop and promotes more substantive discussions.
Right now invitations are limited to 1 per day for each user in a community. You don't need an invite to join at the global level - but to join any community you must have received an invitation link. Still a major work in progress, right now working on expanding the flexibility of community creation and invitation logic. (allowing bulk invites, adding flexible invitation cool downs, etc).
Tech details: I found that used, small form-factor Dell Optiplexes are great for product protoytyping. I'm in Medellin Colombia, and found that you can buy these for about $200 USD - they are often former Point of Sale (POS) or office computers, from about 10 years ago. They have SSDs, run quiet, and are very reliable.
For project Affirmator, I installed Linux Mint Debian Edition (LMDE). Using Cron and Mpv to shuffle-play activity-specific folders of MP3s at the same time each day. For example, for the chill jazz music - I've got a folder of 40+ song MP3s. Cron plays those at 06:30. So it's like a calm, upbeat alarm clock. I'm not a morning person, so this is a "friendly" way for me to wake myself up!
For the vocal affirmation part - I built a Python tool that reads 200+ text affirmations from a markdown/text file. It then uses AWS Polly text-to-speech API to vocalize the affirmations into MP3s. Next, I use `ffmpeg` to add a variable silent spacer gap to the ends of all the MP3s. This allows your to hear a voice affirmation ("I am fit, athletic, and strong!", "I am a confident piano player."), and then there is silent space for you to say it out loud, or repeat in your head.
This project incorporates ideas & routines from: The Strangest Secret by Earl Nightingale, Tony Robbins Personal Power II, Think and Grow Rich by Napoleon Hill, and Atomic Habits by James Clear
https://codeberg.org/jro/Affirmator-app
(2) PROJECT "LINGOFREQ" - Language learning tool. Uses language-specific high-frequency word lists. Generates example sentences according to a theme/topic. Translates the word & example phrases to English / Spanish / Chinese. Uses Text-to-speech to vocalize the phrases into each language. These phrases are ordered by frequency. When you want to improve your language skills, you set a "window" range of frequency you want to practice, and Lingofreq will play audio files in this range. You can learn Chinese & Spanish while doing the dishes, at the gym, or before going to bed!
Code: https://codeberg.org/jro/LingoFreq-app/src/branch/main/apps
(3) Medellin COMMUNITY MAKER-SPACE / CREATIVE ENTREPRENEUR LAB I'm at Medellin Colombia - my mission is to create the best maker-space. I was a member of ASMBLY Maker-space in Austin Texas (great space!) and worked at Pivotal Labs (agile product prototyping / software lab) - so I'm aiming to combine the best ideas from those.
BACK-BURNER projects:
Documenting my Knowledge as "Public Knowledge Base" - https://codeberg.org/jro/Knowledge - Here are my notes on Python, Git - I'm bouncing between Obsidian Sync / Publish / Markdown (currently easiest way), and some sort of open-source knowledge base website (VSCodium + Markdown + FOAM + MkDocs + RClone). I haven't found a solution I'm happy with yet...
Open-source CNC router tech stack: - I have a CNC router (robotic drill which can carve 3D shapes into wood). Last year I challenged myself to operate it completely through an open-source tech stack. This took me on a journey of learning Inkscape (2d vector design tool, SVG), FreeCAD (3D product design / CAD / CAM tool), G-code (format of text instructions which tell CNC tools where to move and what to do), Universal G-Code Sender (a tool which imports CAM - computer aided manufacturing - designs, connects to the CNC router tool, and actually operates machine. It's quite exciting to play with! Used Kiri-moto (web-based CAD / CAM tool) to convert 2D/SVG designs into 3D shapes). Used OBS (screen recording/streaming tool) and a bunch of web-cams to live-stream tool usage to PeerTube Live (similar to YouTube).
Being "principled" about using open-source tools can be so challenging, but its quite rewarding on the long run.
LEARNING SPANISH - What's working for me... trying to read spanish books before bed. Handwriting a few paragraphs from a book into journal. Highlighting words I don't know. Looking them up later. Reading a book while listening to its audio book at the same time.
If anyone's interesting in contributing to these projects, I would warmly welcome that. Design, product, sales, project management, engineering/coding, marketing - need tons of help in all these areas.
Gracias! // JRO
It incorporates also complaints from a static analyzer for Python and Javascript that detects 90+ vibe slop anti-patterns using mostly ASTs, and in some cases AST + small language models. The complaints give the local class and methods a sense of how much pain they are in, so I give the code a sense of its own emotional state.
I also build data flow schematics of the entire system so I can visualize the project as a wire diagram, which is very helpful to quickly see what is going on.
We're also trying to use AI more thoughtfully than just bolting on a chatbot. We're planning to consider each workflow our customers need and how AI might help speed them up - even letting them build custom AI workflows. I think most businesses (especially smaller businesses) don't want to work at the level of Claude Code, Codex, etc. They want to work on higher level problems - build this dashboard, connect these data sources, invoice this customer, etc.
Aside from that, we've noticed that the basics really matter, so we're trying to nail that first.
We're definitely a bit delusional, we're just 3 people, we're doing it without funding and the competition is stiff, but we really believe in the product. Additionally, I think a lot of CRMs go south by taking on too much VC that naturally pushes them to prioritize ROI instead of continually improving the product.
Oracle's plugin allows you access Fusion REST Endpoints (your business data) from within an Excel workbook but it only works on Windows machines and has some other limitations.
Also added a plugin for inspecting punchout payloads for RSSP [2]
- Named `gg` for grep-ibility and ease of typing.
- However Claude has been inserting most calls for me (and can now read back the client-side results without any dev interaction!)
- Here is how Claude used gg to fix a layout bug in itself (gg ships with an optional dev console): https://github.com/Leftium/gg/blob/main/references/gg-consol...
---
# I've been prototyping realtime streaming transcription UX: https://rift-transcription.vercel.app
- Really want to use dictation app in addition to typing on a daily basis, but the current UX of all apps I've tried are insufficient.
---
# https://veneer.leftium.com is a thin layer over Google forms + sheets
- If you can use Google forms, you can publish a nice-looking web site with an optional form
- Example: https://www.vivimil.com
- Example: https://veneer.leftium.com/s.1RoVLit_cAJPZBeFYzSwHc7vADV_fYL...
- DEMO (feel free to try the sign up feature): https://veneer.leftium.com/g.chwbD7sLmAoLe65Z8
RSS and podcast Google Readerish clone mostly for myself
Also I gave that blog post to Claude Code and asked to implement the API and it made terrible terrible mistakes. Just saying.
About an hour ago I was dismissed as AI slop on the r/rust Reddit. Whatever.
This tool is my line of defense in case `trunk` goes dead, which it seems to be increasingly likely. It helps me build fullstack sites using Actix Web and Yew.
Using it now to see if I can re-invent my blog site for the umpteenth time. :)
so far, ive spent a lot of manual time labeling and matching RGB and LWIR images, and trying to figure out first ways to get better pose matches when the flights arent the same.
that, and many different attempts at getting torch to work using my laptop's GPU and NPU. i think im close, without having to build torch from source woo.
Ive been having an eye towards getting better llm generation quality for python too, but havent put a focus on it yet. im fed up with it making one off script after one off script and instead of just making a react app, making some raw html and making a new html file with the new and old bugs every time i want to do something interactive. its maddening.
my last month of gettin claude code ro play pokemon webt well and ive about learned skills pretty well now, but it keeps wanting to do like a challenge run of sticking with a single pokemon.
Also a bunch of other smaller projects and ideas.
It has one SQLite database per queue.
In golang.
Why? Because Rabbit is slow and resource hungry and needs configuration.
Problems I'm having: - Getting enriched vectors because the definitions to some of the words are absolute garbage - Finding a good open source embedding model, currently using nomic-embed-text
Goal: Find me words originating from X city and it not giving me results that match X