Monday, October 27, 2025
Show HN: Easily visualize torch, Jax, tf, NumPy, etc. tensors https://ift.tt/anqfwJx
Show HN: Easily visualize torch, Jax, tf, NumPy, etc. tensors hey hn, wrote a python library for myself to visualize tensors. Makes learning and debugging deep learning code so much easier. Works seamlessly with colab/jupyter notebooks, and other python contexts. It's built on top of the graphics backend, chalk ( https://chalk-diagrams.github.io/ ). why? Understanding deep learning code is hard—especially when it's foreign, because it's hard to imagine tensor manipulations, e.g. `F.conv2d(x.unsqueeze(1), w.transpose(-1, -2)).squeeze().view(B, L, -1)` in your head. Printing shapes and tensor values only get me so far. tensordiagram lets me quickly diagram tensors. Other python libraries for creating tensor diagrams are either too physics and math focused, not notebook-friendly, limited to visualizing single tensors, and/or serve a wider purpose (so have a steep learning curve). https://ift.tt/kISOJeM October 28, 2025 at 03:00AM
Show HN: JSON Query https://ift.tt/KHCNSpc
Show HN: JSON Query I'm working on a tool that will probably involve querying JSON documents and I'm asking myself how to expose that functionality to my users. I like the power of `jq` and the fact that LLMs are proficient at it, but I find it right out impossible to come up with the right `jq` incantations myself. Has anyone here been in a similar situation? Which tool / language did you end up exposing to your users? https://ift.tt/mrc8wBK October 27, 2025 at 09:52PM
Show HN: MNML – Android Launcher (Open Testing Release) https://ift.tt/bEkwevz
Show HN: MNML – Android Launcher (Open Testing Release) Just got the first open testing release approved on google play store for a minimal android launcher i'm developing. I recently switched to a phone with an e-ink display and was missing features in current minimal launchers, so i decided to make my own. Would love to hear your opinion, improvements that could be made or features to change/add. Thanks! === Links === - https://ift.tt/g2RN5Tx - https://www.youtube.com/watch?v=6gcF9lSDBPY (video demo) === Current Features === - Multiple rows of apps - Material Font icons can be set as app names - When reducing or increasing apps you can re-order and select which apps to keep (this is missing from a lot of minimal launchers imo) - App shortcuts (long press on an app to show that app's own shortcut menu) - Backup your launcher settings with Export/Import - App drawer list (swipe up) - You can hide apps from the app list - You can uninstall apps from the app list - App rows can have their size individually set. === Planned features === - Background images or static colors - Show weather information - Screen time usage information - Possibly some limited widgets - google search bar on the bottom? - Notes on the main screen - App folders (so you can assign multiple apps to one slot) - App slot assignment other than apps - URL's - Files - Intents - Contacts - etc. - More gestures that you can assign actions to - 3 or 4 finger swipe - triple tap - pinch zoom - etc. - Find contacts and files in the app list - Localization October 28, 2025 at 12:50AM
Sunday, October 26, 2025
Show HN: Helium Browser for Android with extensions support, based on Vanadium https://ift.tt/x5yfbVG
Show HN: Helium Browser for Android with extensions support, based on Vanadium Been working on an experimental Chromium-based browser that brings 2 major features to your phone/tablet: 1. desktop-style extensions: natively install any extensions (like uBO) from the chrome web store, just toggle "desktop site" in the menu first. 2. privacy/security hardening: applies the full patch sets from Vanadium (with Helium's currently wip). Means you get both browsers' excellent privacy features, like Vanadium's webrtc IP policy option that protects your real IP by default, and security improvements such as JIT being disabled by default, all while being a reasonably efficient FOSS app that can be installed on any (modern) android. It's still in beta, and as I note in the README, it's not a replacement for the full OS-level security model you'd get from running the GrapheneOS Vanadium combo. However, goal was to combine privacy of Vanadium with the power of desktop extensions and Helium features, and make it accessible to a wider audience. (Passkeys from Bitwarden Mobile should also work straight away once merged in the list of FIDO2 privileged browsers) Build scripts are in the repo if you want to compile it yourself. You can find pre-built releases there too. Would love any feedback/support! https://ift.tt/FPVlNUa October 27, 2025 at 04:11AM
Show HN: The Legal Embedding Benchmark (MLEB) https://ift.tt/zi9KrXV
Show HN: The Legal Embedding Benchmark (MLEB) Hey HN, I'm excited to share the Massive Legal Embedding Benchmark (MLEB) — the first comprehensive benchmark for legal embedding models. Unlike previous legal retrieval datasets, MLEB was created by someone with actual domain expertise (I have a law degree and previously led the AI team at the Attorney-General's Department of Australia). I came up with MLEB while trying to train my own state-of-the-art legal embedding model. I found that there were no good benchmarks for legal information retrieval to evaluate my model on. That led me down a months-long process working alongside my brother to identify or, in many cases, build our own high-quality legal evaluation sets. The final product was 10 datasets spanning multiple jurisdictions (the US, UK, Australia, Singapore, and Ireland), document types (cases, laws, regulations, contracts, and textbooks), and problem types (retrieval, zero-shot classification, and QA), all of which have been vetted for quality, diversity, and utility. For a model to do well at MLEB, it needs to have both extensive legal domain knowledge and strong legal reasoning skills. That is deliberate — given just how important high-quality embeddings are to legal RAG (particularly for reducing hallucinations), we wanted our benchmark to correlate as strongly as possible with real-world usefulness. The dataset we are most proud of is called Australian Tax Guidance Retrieval. It pairs real-life tax questions posed by Australian taxpayers with relevant Australian Government guidance and policy documents. We constructed the dataset by sourcing questions from the Australian Taxation Office's community forum, where Australian taxpayers ask accountants and ATO officials their tax questions. We found that, in most cases, such questions can be answered by reference to government web pages that, for whatever reason, users were unable to find themselves. Accordingly, we manually went through a stratified sample of 112 challenging forum questions and extracted relevant portions of government guidance materials linked to by tax experts that we verified to be correct. What makes the dataset so valuable is that, unlike the vast majority of legal information retrieval evaluation sets currently available, it consists of genuinely challenging real-world user-created questions, rather than artificially constructed queries that, at times, diverge considerably from the types of tasks embedding models are actually used for. Australian Tax Guidance Retrieval is just one of several other evaluation sets that we painstakingly constructed ourselves simply because there weren't any other options. We've contributed everything, including the code used to evaluate models on MLEB, back to the open-source community. Our hope is that MLEB and the datasets within it will hold value long into the future so that others training legal information retrieval models won't have to detour into building their own "MTEB for law". If you'd like to head straight to the leaderboard instead of reading our full announcement, you can find it here: https://ift.tt/duy6TAP If you're interested in playing around with our model, which happens to be ranked first on MLEB as of 16 October 2025 at least, check out our docs: https://ift.tt/IWE1fZ2 https://ift.tt/hOWXJuz October 27, 2025 at 03:46AM
Show HN: MyraOS – My 32-bit operating system in C and ASM (Hack Club project) https://ift.tt/hoWER6B
Show HN: MyraOS – My 32-bit operating system in C and ASM (Hack Club project) Hi HN, I’m Dvir, a young developer. Last year, I got rejected after a job interview because I lacked some CPU knowledge. After that, I decided to deepen my understanding in the low level world and learn how things work under the hood. I decided to try and create an OS in C and ASM as a way to broaden my knowledge in this area. This took me on the most interesting ride, where I’ve learned about OS theory and low level programming on a whole new level. I’ve spent hours upon hours, blood and tears, reading different OS theory blogs, learning low level concepts, debugging, testing and working on this project. I started by reading University books and online blogs, while also watching videos. Some sources that helped me out were OSDev Wiki ( https://ift.tt/lZcGgPW ), OSTEP ( https://pages.cs.wisc.edu/~remzi/OSTEP ), open-source repositories like MellOS and LemonOS (more advanced), DoomGeneric, and some friends that have built an OS before. This part was the longest, but also the easiest. I felt like I understood the theory, but still could not connect it into actual code. Sitting down and starting to code was difficult, but I knew that was the next step I needed to take! I began by working on the bootloader, which is optional since you can use a pre-made one (I switched to GRUB later), but implementing it was mainly for learning purposes and to warm up on ASM. These were my steps after that: 1) I started implementing the VGA driver, which gave me the ability to display text. 2) Interrupts - IDT, ISR, IRQ, which signal to the CPU that a certain event occurred and needs handling (such as faults, hardware connected device actions, etc). 3) Keyboard driver, which enables me to display the same text I type on my keyboard. 4) PMM (Physical memory management) 5) Paging and virtual memory management 6) RTC driver - clock addition (which was, in my opinion, optional) 7) PIT driver - Ticks every certain amount of time, and also 8) FS (File System) and physical HDD drivers - for the HDD I chose PATA (HDD communication protocol) for simplicity (SATA is a newer but harder option as well). For the FS I chose EXT2 (The Second Extended FileSystem), which is a foundational linux FS structure introduced in 1993. This FS structure is not the simplest, but is very popular in hobby-OS, it is very supported, easy to set up and upgrade to newer EXT versions, it has a lot of materials online, compared to other options. This was probably the longest and largest feature I had worked on. 9) Syscall support. 10) Libc implementation. 11) Processing and scheduling for multiprocessing. 12) Here I also made a shell to test it all. At this point, I had a working shell, but later decided to go further and add a GUI! I was working on the FS (stage 8), when I heard about Hack Club’s Summer of Making (SoM). This was my first time practicing in HackClub, and I want to express my gratitude and share my enjoyment of participating in it. At first I just wanted to declare the OS as finished after completing the FS, and a bit of other drivers, but because of SoM my perspective was changed completely. Because of the competition, I started to think that I needed to ship a complete OS, with processing, GUI and the bare minimum ability to run Doom. I wanted to show the community in SoM how everything works. Then I worked on it for another 2 months, after finishing the shell, just because of SoM!, totalling my project to almost 7 months of work. At this time I added full GUI support, with dirty rectangles and double buffering, I made a GUI mouse driver, and even made a full Doom port! things I would've never even thought about without participating in SoM. This is my SoM project: https://ift.tt/JRslr58 . Every project has challenges, especially in such a low level project. I had to do a lot of debugging while working on this, and it is no easy task. I highly recommend using GDB which helped me debug so many of my problems, especially memory ones. The first major challenge I encountered was during the coding of processes - I realized that a lot of my paging code was completely wrong, poorly tested, and had to be reworked. During this time I was already in the competition and it was difficult keeping up with devlogs and new features while fixing old problems in a code I wrote a few months ago. Some more major problems occurred when trying to run Doom, and unlike the last problem, this was a disaster. I had random PFs and memory problems, one run could work while the next one wouldn’t, and the worst part is that it was only on the Doom, and not on processes I created myself. These issues took a lot of time to figure out. I began to question the Doom code itself, and even thought about giving up on the whole project. After a lot of time spent debugging, I fixed the issues. It was a combination of scheduling issues, Libc issues and the Qemu not having enough (wrongfully assuming 128MB for the whole OS was enough). Finally, I worked throughout all the difficulties, and shipped the project! In the end, the experience working on this project was amazing. I learned a lot, grew and improved as a developer, and I thank SoM for helping to increase my motivation and make the project memorable and unique like I never imagined it would be. The repo is at https://ift.tt/UXZramD . I’d love to discuss any aspect of this with you all in the comments! https://ift.tt/UXZramD October 27, 2025 at 02:13AM
Show HN: I Built DevTools for Blazor (Like React DevTools but for .NET) https://ift.tt/MyNbQFP
Show HN: I Built DevTools for Blazor (Like React DevTools but for .NET) Hi HN! I've been working on developer tools for Blazor that let you inspect Razor components in the browser, similar to React DevTools or Vue DevTools. The problem: Blazor is Microsoft's frontend framework that lets you write web UIs in C#. It's growing fast but lacks the debugging tools other frameworks have. When your component tree gets complex, you're stuck with Console.WriteLine debugging. What I built: A browser extension + NuGet package that: Shows the Razor component tree in your browser Maps DOM elements back to their source components Highlights components on hover Works with both Blazor Server and WASM How it works: The NuGet package creates shadow copies of your .razor files and injects invisible markers during compilation. These markers survive the Razor→HTML pipeline. The browser extension reads these markers to reconstruct the component tree. Current status: Beta - it works but has rough edges. Found some bugs when testing on larger production apps that I'm working through. All documented on GitHub. Technical challenges solved: Getting markers through the Razor compiler without breaking anything Working around CSS isolation that strips unknown attributes Making it work with both hosting models It's completely open source: https://ift.tt/A9TQNiZ Demo site where you can try it: https://ift.tt/trX0hvz Would love feedback, especially from anyone building production Blazor apps. What debugging pain points do you have that developer tools could solve? https://ift.tt/w3B6EHn October 26, 2025 at 10:04PM
Saturday, October 25, 2025
Show HN: Gisia – A Lightweight Self-Hosted DevOps Platform https://ift.tt/bexP8lA
Show HN: Gisia – A Lightweight Self-Hosted DevOps Platform Hi HN, I've been building Gisia – a lightweight, self-hosted DevOps platform for individuals and small teams. It's simpler, more lightweight and you can self-host. Key Features: - Git repository hosting with SSH and HTTP access - CI/CD pipelines with YAML configuration - Issue tracking - Built with modern Rails stack (Rails 8, PostgreSQL, Stimulus/Turbo, Tailwind CSS) Why Gisia? - Lightweight with minimal dependencies - Developer-first design - Fully open-source and auditable - You own your data Current Status: Currently in Alpha with core features complete. Merge requests and notifications are planned. Try It Out: You can take Gisia for a test drive by following the quick start guide in the README. Feedback and contributions welcome! https://ift.tt/FbVNyaw October 25, 2025 at 08:30PM
Show HN: I created a small 2D game about an ant https://ift.tt/kLHERYX
Show HN: I created a small 2D game about an ant Hello HN! I created a short game in just a few days, just for fun, where you play as an ant and feed it apples. This game also features random landscape generation, where clouds and trees are randomly distributed across all coordinates (only trees do not grow in the y direction). This is what took me the longest time :) I would appreciate your feedback ^ ^ https://ift.tt/cMkFW7O October 26, 2025 at 12:50AM
Show HN: Random Makers – Show HN and Product Hunt, but Faster and Not Corporate https://ift.tt/5QPn3NL
Show HN: Random Makers – Show HN and Product Hunt, but Faster and Not Corporate https://ift.tt/x6jN2Ku October 25, 2025 at 11:32PM
Friday, October 24, 2025
Show HN: Wsgrok – one of many ngrok alternatives https://ift.tt/a6buyfG
Show HN: Wsgrok – one of many ngrok alternatives I built it for myself because ngrok didn't let me add one more domain unless I paid $12 more. I probably should've looked for alternatives before building my own, but my grudge got in the way . Once I started, I wanted to make it better than the other options. Silly, I know. No one probably cares. I plan to open source it sometime next year because I’ve got other projects to finish first. It's free until I deplete my cloud credits, then it will switch to a tier-based model with a free option. https://wsgrok.com October 25, 2025 at 05:16AM
Show HN: Pensive – A bookmarking tool with full-text search and LLMs https://ift.tt/lJhbLCu
Show HN: Pensive – A bookmarking tool with full-text search and LLMs After Pocket shut down, I started working on a new bookmarking solution. My main goal was to make bookmarks searchable, something that Pocket was not good at. So I built Pensive, a bookmarking solution that saves the full page content and makes it searchable with full-text search. You can add pages using a browser extension or a Telegram bot (for saving on mobile). It is written in Go with PostgreSQL, all Dockerized on a $5 Hetzner server, and the front end is built with HTMX. I have also added embeddings (using Gemini Flash Lite) to let LLMs interact with your bookmarks contextually. It is stable enough that I now use it daily. I’m considering publishing it as open source, but first I want to have a proper version ready. Feel free to email me if you’re interested in helping out or if you have prior experience with open source. Feel free to try it at https://getpensive.com https://getpensive.com/ October 25, 2025 at 03:28AM
Show HN: The Σ-Manifold Manifesto https://ift.tt/syvxbqe
Show HN: The Σ-Manifold Manifesto This project explores the connection between *the linear structure of text* and its *emotional-aesthetic impact*. We identify *five fundamental relations* between consecutive sentences — labeled *A–E*. Each represents a shift of *subject-object*, i.e., a transformation of perspective and agency. When texts grow longer, these relations form *sequences* — and from the infinite combinatorial space, *eight stable patterns (Σ₁–Σ₈)* emerge empirically. Each pattern correlates with a distinct *semantic and emotional field* — cathartic, heroic, meditative, humorous, and so on. This allows us to instruct an LLM not through semantic prompts (“write a story about…”), but through *structural commands* — e.g., generate a narrative following sequence Σ₅ (Tragic Counterpoint). You can experiment with these archetypes directly here: [Narrative Generator]( https://ift.tt/JhnFfe9... ) or [via python]( https://ift.tt/l0TeKzp... ) Interestingly, there appears to be a parallel between these textual progressions and *musical harmony*. For example, if A–E are mapped to harmonic functions (I, IV, V, vi, ii), the narrative sequences behave like emotional “chord progressions” — where meaning flows, modulates, and resolves. Coherence in the generated text arises not only from syntax, but from the *associative field* that the LLM constructs around these shifting relations. When asked to “switch subjects,” it spontaneously moves from Poet to Writer , preserving aesthetic continuity rather than randomness. It might even hint at how *children acquire language*: by first sensing the melody of structural transitions, before mapping them to concepts and emotions. Such a method could eventually apply to *training neural systems*, where meaning is learned as flow — not as fixed representation. [Full text]( https://ift.tt/6F8di4S... ) October 25, 2025 at 01:34AM
Show HN: Check what is hogging your disk zpace https://ift.tt/jEf0DTM
Show HN: Check what is hogging your disk zpace Hello world! I would like to share that I have created a simple open-source Python CLI app to check what's hogging all the disk space! You can install it with pip. It's like space but Zpace. pip install zpace and find the big files consuming your disk space. Be it apps, virtual environments, machine learning models etc. Just run zpace Once you find one, that you can say "bye" to, just run rm -rf /i_dont_need/this/file/right_now to get rid of it. It was born out of frustration while lack of disk space prevented me to use my laptop properly. It has been very useful to me so far. I hope that it can be useful to you as well. Feel free to check it out. Currently tested on MacOS only https://ift.tt/47XyuY8 October 25, 2025 at 01:12AM
Thursday, October 23, 2025
Show HN: Git for LLMs – a context management interface https://ift.tt/kiFNJeO
Show HN: Git for LLMs – a context management interface Hi HN, we’re Jamie and Matti, co-founders of Twigg. During our master’s we continually found the same pain points cropping up when using LLMs. The linear nature of typical LLMs interfaces - like ChatGPT and Claude - made it really easy to get lost without any easy way to visualise or navigate your project. Worst of all, none of them are well suited for long term projects. We found ourselves spending days using the same chat, only for it to eventually break. Transferring context from one chat to another is also cumbersome. We decided to build something more intuitive to the ways humans think. We started with two simple ideas. Enabling chat branching for exploring tangents, and an interactive tree diagram to allow for easy visualisation and navigation of your project. Twigg has developed into an interface for context management - like “Git for LLMs”. We believe the input to a model - or the context - is fundamental to its performance. To extract the maximum potential of an LLM, we believe the users need complete control over exactly what context is provided to the model, which you can do using simple features like cut, copy and delete to manipulate your tree. Through Twigg, you can access a variety of LLMs from all the major providers, like ChatGPT, Gemini, Claude, and Grok. Aside from a standard tiered subscription model (free, plus, pro), we also offer a Bring Your Own Key (BYOK) service, where you can plug and play with your own API keys. Our target audience are technical users who use LLMs for large projects on a regular basis. If this sounds like you, please try out Twigg, you can sign up for free at https://twigg.ai/ . We would love to get your feedback! https://twigg.ai October 23, 2025 at 08:42PM
Show HN: Pg_textsearch – BM25 Ranking for Postgres https://ift.tt/Icj2Qtr
Show HN: Pg_textsearch – BM25 Ranking for Postgres I built pg_textsearch, a Postgres extension that brings proper BM25 ranking to full-text search. It's designed for AI/RAG workloads where search quality directly impacts LLM output. Postgres native ts_rank lacks corpus-aware signals (no IDF, no TF saturation, no length normalization). This causes mediocre documents to rank above excellent matches, which matters when your LLM depends on retrieval quality. Quick example: CREATE EXTENSION pg_textsearch; CREATE INDEX articles_idx ON articles USING bm25(content); SELECT title, content <@> to_bm25query('database performance', 'articles_idx') AS score FROM articles ORDER BY score LIMIT 10; Works seamlessly with pgvector or pgvectorscale for hybrid search. Fully transactional (no sync jobs). Preview release uses in-memory architecture (64MB default per index); disk-based segments coming soon. I love ParadeDB's pg_search but wanted something available on our managed Postgres. You can try pg_textsearch free on Tiger Cloud: https://ift.tt/HI72f41 Blog: https://ift.tt/PJAKbFa... Docs: https://ift.tt/sJSNMIc... Feedback welcome, especially from folks building RAG systems or hybrid search applications. https://ift.tt/aG3EWhF October 23, 2025 at 10:55PM
Wednesday, October 22, 2025
Show HN: Middlerok – reduces front end-back end integration from weeks to hours https://ift.tt/HXOgrB4
Show HN: Middlerok – reduces front end-back end integration from weeks to hours Generate production-ready OpenAPI specs, frontend & backend code and documentation with AI https://ift.tt/5TupMFy October 22, 2025 at 11:05PM
Show HN: Incremental JSON parser for streaming LLM tool calls in Ruby https://ift.tt/8cHvlyF
Show HN: Incremental JSON parser for streaming LLM tool calls in Ruby Built this for streaming AI tool calls. LLMs stream function arguments as JSON character-by-character. Most parsers reparse from scratch each time - O(n²) behavior that causes UI lag. This maintains parsing state, processing only new characters. True O(n) performance that stays imperceptible throughout the entire response. Ruby gem, MIT licensed. Would love feedback. https://ift.tt/W3biNSe October 23, 2025 at 01:02AM
Show HN: GuardianScan – Website audits for 2025 web standards https://ift.tt/kmbcoBQ
Show HN: GuardianScan – Website audits for 2025 web standards Hey HN, I've been building websites for 15 years. I'm a perfectionist - both visually and technically - and got tired of going back and forth between different tools, manual checks, and documentation for every audit. Built GuardianScan to run everything in under 45 seconds: - Core Web Vitals (LCP, INP, CLS) - WCAG 2.2 compliance - Security headers and CSP policies - Modern framework patterns - SEO and schema markup 47 checks total. These standards exist for a reason. Google prioritises fast sites. Accessible sites reach more people. Secure sites build trust. When you meet modern standards, you see it in your traffic and conversions. Stack: Next.js 15, React 19, Supabase, Browserless.io for headless Chrome. Won't catch everything - automated accessibility gets about 70% of issues. But it's what I use before every deploy now. Landing page is up. Launching Nov 1. £24/month, no enterprise pricing BS. Happy to answer questions. https://guardianscan.ai October 22, 2025 at 08:36PM
Show HN: Create interactive diagrams with pop-up content https://ift.tt/Ete2ONU
Show HN: Create interactive diagrams with pop-up content This is a recent addition to Vexlio which I think the HN crowd may find interesting or useful. TL;DR: easy creation of interactive diagrams, meaning diagrams that have mouse click/hover hooks that you can use to display pop-up content. The end result can be shared with a no-sign-in-required web link. My thought is that this is useful for system docs, onboarding or user guides, presentations, etc. Anything where there is a high-level view that should remain uncluttered + important metadata or details that still need to be available somewhere. You can try it out without signing up for anything, just launch the app here ( https://app.vexlio.com/ ), create a shape, select it with the main pointer tool and then click "Add popup" on the context toolbar. I'd be grateful for any and all feedback! https://ift.tt/2Xk5fMV October 22, 2025 at 08:15PM
Tuesday, October 21, 2025
Show HN: MTOR – A free, local-first PWA to automate workout progression https://ift.tt/XeqQdhb
Show HN: MTOR – A free, local-first PWA to automate workout progression Hi HN, My motivation for this came from frustration with existing workout trackers. Most felt clunky, hid core features like performance graphs behind a paywall, or forced a native app download. A few people close to me who take their training seriously shared the same sentiment, so I decided to build my own. I'm working on mTOR, a free, science-based workout tracker I built to automate progressive overload. It's a local-first PWA that works completely offline, syncs encrypted between your devices using passwordless passkeys, and allows for plan sharing via a simple link. The core idea is to make progression easier to track and follow. After a workout, it analyzes your performance (weight, reps, and RIR), highlights new personal records (PRs), and generates specific targets for your next session. It also reviews your entire program to provide scientific analysis on weekly volume, frequency, and recovery for each muscle group. This gets displayed visually on an anatomy model to help you learn which muscles are involved, and you can track your gains over time with historical performance charts for each exercise. During a workout, you get a total session timer, an automatic rest timer, and can see your performance from the last session for a clear target to beat. It automatically advances to the next incomplete exercise, and when you need to swap an exercise, it provides context-aware alternatives targeting the same muscles. It's also deeply customizable: * The UI has a dark theme, supports multiple languages (English, Spanish, German), lets you adjust the UI scale, and toggle the visibility of detailed muscle names, exercise types, historical performance badges, and a full history card. * You can set global defaults for weight units (kg/lbs), rest times, and plan targets, or enable/disable metrics like Reps in Reserve (RIR) and estimated 1-Rep Max. The exercise library can be filtered by your available equipment, you can create your own custom exercises with global notes, and there's a built-in weight plate calculator. * The progression system lets you define default rep ranges and RIR targets, or create specific overrides for different lifts (e.g., a 3-5 rep range for strength, 10-15 for accessories). * Editing is flexible: you can drag-and-drop to reorder days, exercises, and sets, duplicate workout days, track unilateral exercises (left/right side), and enter data with a quick wheel picker. I'll be here all day to answer questions. I'm also thinking about making the project open-source down the line and would be curious to hear any thoughts on that. Thanks for checking it out! https://mtor.club/ October 22, 2025 at 12:04AM
Show HN: bbcli – A TUI and CLI to browse BBC News like a hacker https://ift.tt/Hbc3e28
Show HN: bbcli – A TUI and CLI to browse BBC News like a hacker hey hn! I (re)built this TUI tool for browsing BBC News in the terminal, it uses an RSS feed for getting headlines and previews and you can read articles too. Try it out and let me know what you think! :) https://ift.tt/nJ498jM October 19, 2025 at 04:28PM
Show HN: OpenJobHub – A Free and Open Job Board https://ift.tt/OzkcJ5V
Show HN: OpenJobHub – A Free and Open Job Board Hey everyone, I recently built OpenJobHub, a free and open source job board for engineers. My goal is simple: in these tough times, I hope more people can find good jobs, and that job information can be more open and transparent. Feel free to share it with any recruiters, HRs, or friends who are hiring. Let's build a healthier job ecosystem together. GitHub: https://ift.tt/WEBK7oY Website: https://jobs.wowkit.net https://ift.tt/WEBK7oY October 21, 2025 at 11:03PM
Monday, October 20, 2025
Show HN: Online Sourcerer – The best answer to 'source?' https://ift.tt/eJa2SZK
Show HN: Online Sourcerer – The best answer to 'source?' Hello, I made this site to combat misinformation on the internet by allowing users to prove that their claim is valid by linking multiple sources and combining them in a single link It's very early stage so I would love feedback on: - What types of claims would be most useful to you? - How can I make verification/sourcing more robust? - Any features that would make this actually useful vs just interesting? Thanks in advance, feel free to roast :) https://ift.tt/53cFsM1 October 21, 2025 at 03:46AM
Show HN: I built a Product Hunt alternative for African tech startups https://ift.tt/F3urfQV
Show HN: I built a Product Hunt alternative for African tech startups With the advance of AI, everyone is building and creating solutions. It is actually amazing to see people in Africa engage this creative field and produce products that could help their ecosystem. So I built a platform that will soon be the biggest directory of tech products made by people in Africa to solve African Problems. I am taking this path because I understand we have very different problems and different cultures and things work differently especially in Africa. What do you guys think? I want your feedback on everything. https://ift.tt/HvcjpST October 21, 2025 at 02:25AM
Show HN: I created a cross-platform GUI for the JJ VCS (Git compatible) https://news.ycombinator.com/item?id=45645120
Show HN: I created a cross-platform GUI for the JJ VCS (Git compatible) Personally, I think the JJ VCS ( https://github.com/jj-vcs/jj ) hit a point some time in this past year where I find it hard to find a great reason to continue using `git`. Over the years I've cobbled together aliases and bash functions to try to improve my `git` workflow, but after using `jj`, which works with ~any git repo and integrates great with Github repos, all of the workflow issues I ran into with git are not only solved, but improved in ways I couldn't manage with simple scripts. One example is the `op log`, which lets you go to any point in your repo's time and provides simple `undo` and `redo` commands when you want to back out of a merge, didn't mean to rebase, etc. Because I have a pretty strong conviction that JJ is at this point a cleaner and more powerful version of `git`, my hopes are that it continues to grow. With that, it seemed a proper full-featured GUI was missing for the VCS. There's some plugins that add some integration into VS Code, and there's one in the works to get Intellij support working, but many of the constructs JJ provides in my opinion necessitate a grounds-up build of a GUI around how JJ works. Right now, Judo for JJ is an MVP in an open beta. I did my best to support all of the core functionality one would need, though there's many nice-to-haves that I am going to add, like native merge support, native splitting, etc. Most of this will be based on feedback from the Beta. I'm really grateful for the great community JJ has built, alongside the HN community itself in the countless VCS-based posts I've read over the years, and am hoping for lots of input here during Beta under real usage - the goal is to be a full-featured desktop GUI for the VCS, similar to many of the great products that are out there for git. https://judojj.com October 20, 2025 at 09:05PM
Show HN: NativeBlend – Text to fully editable 3D Models that don't suck https://news.ycombinator.com/item?id=45647738
Show HN: NativeBlend – Text to fully editable 3D Models that don't suck I'm a developer (not a 3D artist) who's been frustrated with current AI text-to-3D tools — most produce messy, monolithic meshes that are unusable without hours of cleanup. So I built NativeBlend, a side project aimed at generating editable 3D assets that actually fit into a real workflow. Key features: - Semantic Part Segmentation: Outputs separate, meaningful components (e.g., wheels, doors), not just a single mesh blob. - Native Blender Output: Generates clean, structured .blend files with proper hierarchies, editable PBR materials, and decent UVs — no FBX/GLB cleanup required. The goal is to give devs a usable starting point for game assets without the usual AI slop. I have a working demo and would love feedback: Does this solve a real need, or am I just scratching my own itch? Thanks for taking a look! https://native-blend-app.vercel.app/ October 21, 2025 at 12:27AM
Sunday, October 19, 2025
Show HN: Hokusai Pocket (WIP) – Portable GUIs with MRuby https://ift.tt/wngf1kv
Show HN: Hokusai Pocket (WIP) – Portable GUIs with MRuby Whassup?, A couple years ago, I started a project for easily authoring GUIs with Ruby. The project is named Hokusai. It features the ability to compose reactive UI components with events and props, and uses a unique-ish template language. More information on Hokusai can be found here: https://ift.tt/EhUzQC2 Since then I've worked on Hokusai Native ( https://ift.tt/QLctvzK ), which compiles a GraalVM native image / TruffleRuby version of Hokusai that can run / interpret these lil' gui apps. It's quite bloated though, as it has to ship all of truffle ruby + native image and supporting libs. Recently, I applied for a grant to develop a more portable version of this library using MRuby, and got pretty far while waiting for the results. It is named Hokusai Pocket and I consider it to be the final form/approach of this project. I wrote a builder in crystal-lang that embeds the entire Hokusai ruby code as MRuby bytecode, as well as the supporting C code. It can scaffold new projects by building tree-sitter/mruby/raylib, and outputs a binary from a source ruby file. It produce pretty small binaries (~3mb for MacOS) and uses raylib as the rendering engine. For an gif and example of a Hokusai Pocket demo please direct your mouse clicks to this gist: https://gist.github.com/skinnyjames/b510185c6bd83fd4e1a41324... I'd love to hear how this project plays for people. Still working on building for different targets, but android and web should be possible. The project is still undergoing active development, but any help is appreciated. The license is MIT. There also is a discord channel if you want to get help / chat / collaborate: https://ift.tt/Q2W54an _ (^) (_\ |_| \_\ |_| _\_\,/_| (`\(_|`\| (`\,) \ \ \,) | | \__(__| https://ift.tt/9UZBtPC October 20, 2025 at 07:00AM
Show HN: 18yo first iOS app: blocks distracting apps and unlocks with QR/barcode https://ift.tt/oPps1R9
Show HN: 18yo first iOS app: blocks distracting apps and unlocks with QR/barcode I built Recode because I realized I was spending 8-10 hours a day on my phone pretty consistently. I tried other screen time apps but I found them too easy to bypass and end my blocks whenever I wanted to use an app. My solution was to build an app blocker app that makes users have to scan a physical QR/barcode to take a break from their app blocks. This helped me be able to get my screen time down to just a few hours everyday since I didn't want to physically get up and go across the house to get my barcode. Anyways, since it worked for me I felt like sharing it. App store link: https://ift.tt/HP2DE5z... https://ift.tt/mKePaRL October 20, 2025 at 03:00AM
Show HN: Jotite – A whimsical Linux Markdown note-taking app https://ift.tt/TH4UcVp
Show HN: Jotite – A whimsical Linux Markdown note-taking app https://ift.tt/kfwDdgb October 20, 2025 at 01:32AM
Show HN: CheckHN – A checklist for the most popular Hacker News posts https://ift.tt/A7CwVHq
Show HN: CheckHN – A checklist for the most popular Hacker News posts I realized that I'm more interested in the all-time top content on Hacker News than in whatever is popular right now. To better keep track of which posts I've already read, I built a small web app that displays all posts sorted by popularity and allows me to check them off or save them for later. https://ift.tt/zGokRWn October 19, 2025 at 10:58PM
Show HN: WP-Easy, framework to build WordPress themes https://ift.tt/BRsHiyl
Show HN: WP-Easy, framework to build WordPress themes The inspiration for this framework came from my brother, an amazing graphic designer who wanted to build WordPress themes using only his FTP-based code editor. He knows HTML and CSS really well, and some jQuery, but not modern JavaScript. In my experience, this is common for people whose jobs are tangential to frontend web development... designers, copywriters, project managers, and backend engineers. So this is for people who don't want to deal with the mess of modern build tools. It tries to nudge people into a more modern direction: component-based architecture, JS modules, SCSS, and template routing. WP-Easy lets people like my brother build professional, modern themes without the usual barriers, just code with your favorite editor and see the results instantly. Key features: 1. File-based routing - Define routes in router.php with Express-like syntax (/work/:slug/) 2. Single File Components - PHP templates with
,
Saturday, October 18, 2025
Show HN: Odyis: lunar lander (1979) clone written in Rust https://ift.tt/VdSgtxW
Show HN: Odyis: lunar lander (1979) clone written in Rust Moin, to learn Rust I decided to create a simple clone of the original lunar lander game. I would love to hear feedback on the quality of the code! https://ift.tt/1fw7iF8 October 19, 2025 at 12:27AM
Friday, October 17, 2025
Show HN: Code review for AI native teams https://ift.tt/pB5Ox49
Show HN: Code review for AI native teams Bottleneck is a desktop app for code reviewing AI-written pull requests. The goal is to be able to track PRs on GitHub authored by agents (i.e. Codex, Devin, Cursor, Claude Code) and compare branches. So if you throw multiple coding agents at a ticket, this would be an easier way to let agents "bake off" against each other and pick the best one. (No need to open the Github website and switch between slow loading tabs). I'd love feedback from any power users who are deep with AI coding agents. https://ift.tt/RcYp3Tg October 18, 2025 at 05:52AM
Show HN: OneClickPRD – Save hours vibe coding with concise PRDs https://ift.tt/JOEnzqF
Show HN: OneClickPRD – Save hours vibe coding with concise PRDs Hi HN, I built OneClickPRD because as a solo builder I often wasted hours vibe coding without clear goals. I’d start with an idea, but it was vague, so the code got messy and I had to redo things. OneClickPRD asks you a few questions about your product and then generates a short, structured PRD. The format works well with AI tools like Replit, Lovable, or v0, so you can go from idea to working MVP much faster. Demo: https://ift.tt/1zFbxsO Would love your feedback: does this feel useful for your projects, and what would make it better? https://ift.tt/1zFbxsO October 18, 2025 at 02:55AM
Show HN: I turned my resume into a catchy song. It's a game changer https://ift.tt/8uAJ1rf
Show HN: I turned my resume into a catchy song. It's a game changer I turned my resume into a catchy pop song. Thought you'd all appreciate it. Worked directly on the Song Style prompt, which you can duplicate for your own fun catchy resume song. Just replace the lyrics! https://ift.tt/GOUqdN9 October 18, 2025 at 02:22AM
Show HN: We packaged an MCP server inside Chromium https://ift.tt/aPS973x
Show HN: We packaged an MCP server inside Chromium Hey HN, we just shipped a browser with an inbuilt MCP server! We're a YC startup (S24) building BrowserOS — an open‑source Chromium fork. We're a privacy‑first alternative to the new wave of AI browsers like Dia, Perplexity Comet. Since launching ~3 months ago, the #1 request has been to expose our browser as an MCP server. -- Google beat us to launch with chrome-devtools-mcp (solid product btw), which lets you build/debug web apps by connecting Chrome to coding assistants. But we wanted to take this a step further: we packaged the MCP server directly into our browser binary. That gives three advantages: 1. MCP server setup is super simple — no npx install, no starting Chrome with CDP flags, you just download the BrowserOS binary. 2. with our browser's inbuilt MCP server, AI agents can interact using your logged‑in sessions (unlike chrome-devtools-mcp which starts a fresh headless instance each time) 3. our MCP server also exposes new APIs from Chromium's C++ core to click, type, and draw bounding boxes on a webpage. Our APIs are also not CDP-based (Chrome Debug Protocol) and have robust anti-bot detection. -- Few example use cases for BrowserOS-mcp are: a) *Frontend development with Claude Code*: instead of screenshot‑pasting, claude-code gets WYSIWYG access. It can write code, take a screenshot, check console logs, and fix issues in one agentic sweep. Since it has your sessions, it can do QA stuff like "test the auth flow with my Google Sign‑In." Here's a video of claude-code using browserOS to improve the css styling with back-and-forth checking: https://youtu.be/vcSxzIIkg_0 b) *Use as an agentic browser:* You can install BrowserOS-mcp in claude-code or Claude Desktop and do things like form-filling, extraction, multi-step agentic tasks, etc. It honestly works better than Perplexity Comet! Here's a video of claude-code opening top 5 hacker news posts and summarizing: https://youtu.be/rPFx_Btajj0 -- *How we packaged MCP server inside Chromium binary*: We package the server as a Bun binary and expose MCP tools over HTTP instead of stdio (to support multiple sessions). And we have a BrowserOS controller installed as an extension at the application layer which the MCP server connects to over WebSocket to control the browser. Here's a rough architecture diagram: https://dub.sh/browseros-mcp-diag -- *How to install and use it:* We put together a short guide here: https://ift.tt/JWqg4hp Our vision is to reimagine the browser as an operating system for AI agents, and packaging an MCP server directly into it is a big unlock for that! I'll be hanging around all day, would love to get your feedback and answer any questions! https://ift.tt/9lMjbUQ October 17, 2025 at 09:52PM
Thursday, October 16, 2025
Show HN: Code First CDC from Postgres to ClickHouse with MooseStack https://ift.tt/aOSw16D
Show HN: Code First CDC from Postgres to ClickHouse with MooseStack https://ift.tt/REc72Bh October 17, 2025 at 02:46AM
Show HN: Arky – Visual 2D Markdown editor (access codes below) https://ift.tt/exPGS1y
Show HN: Arky – Visual 2D Markdown editor (access codes below) Hey HN! Arky is a markdown editor with a twist — instead of writing in a linear doc, you work on a 2D canvas where you can: • Place ideas anywhere spatially • Organize them into hierarchy with drag & drop • See the full document structure at a glance • AI writes contextually — drag & drop responses anywhere on canvas More info: https://arky.so/ Try it: https://app.arky.so/ (Access codes are in the comment below!) Quick demo(60s): https://youtu.be/Nxsr5Ag2vEM?si=g6nKheRWWNuaLTe8 Would love to hear your thoughts! https://app.arky.so October 17, 2025 at 01:20AM
Show HN: Inkeep (YC W23) – Agent builder that works both visually and in code https://ift.tt/3kNXrLY
Show HN: Inkeep (YC W23) – Agent builder that works both visually and in code Hi HN! I'm Nick from Inkeep. We built an agent builder with true 2-way sync between code and a drag-and-drop visual editor, so devs and non-devs can collaborate on the same agents. Here’s a demo video: https://ift.tt/9W0ep3o . As a developer, the flow is: 1) Build AI Chat Assistants or AI Workflows with the TypeScript SDK 2) Run `inkeep push` from your CLI to publish 3)Edit agents in the visual builder (or hand off to non-technical teams) 4) Run `inkeep pull to edit in code again. We built this because we wanted the accessibility of no-code workflow builders (n8n, Zapier), but the flexibility and devex of code-based agent frameworks (LangGraph, Mastra). We also wanted first-class support for chat assistants with interactive UIs, not just workflows. OpenAI got close, but you can only do a one-time export from visual builder to code and there’s vendor lock-in. How I've used it: I bootstrapped a few agents for our marketing and sales teams, then was able to hand off so they can maintain and create their own agents. This has enabled us to adopt agents across technical and non-technical roles in our company on a single platform. To try it, here’s the quickstart: https://ift.tt/nVweGA3 . We leaned on open protocols to make it easy to use agents anywhere: An MCP endpoint, so agents can be used from Cursor/Claude/ChatGPT A Chat UI library with interactive elements you can customize in React An API endpoint compatible with the Vercel AI SDK `useChat` hook Support for Agent2Agent (A2A) so they work with other agent ecosystems We made some practical templates like a customer_support, deep_research, and docs_assistant. Deployment is easy with Vercel/Docker with a fair-code license and there's a traces UI and OTEL logs for observability. Under the hood, we went all-in on a multi-agent architecture. Agents are made up of LLMs, MCPs, and agent-to-agent relationships. We’ve found this approach to be easier to maintain and more flexible than traditional “if/else” approaches for complex workflows. The interoperability works because the SDK and visual builder share a common underlying representation, and the Inkeep CLI bridges it with a mix of LLMs and TypeScript syntactic sugar. Details in our docs: https://docs.inkeep.com . We’re open to ideas and contributions! And would love to hear about your experience building agents - what works, hasn’t worked, what’s promising? https://ift.tt/hukwWEO October 16, 2025 at 06:20PM
Show HN: Coordable – Get better geocoding results with AI cleaning and analytics https://ift.tt/QonYh6P
Show HN: Coordable – Get better geocoding results with AI cleaning and analytics I’ve been working on a tool called Coordable, which helps analyze and improve geocoding results. If you’ve ever dealt with geocoding at scale, you’ve probably hit two recurring problems: Garbage in = garbage out. Addresses are often messy (“2nd floor”, “/”, abbreviations, multiple addresses in one line…). Most geocoders will fail or return incorrect matches if the input isn’t perfectly normalized. A result isn’t always a correct result. Many providers return something even if it’s wrong — e.g. shifting a house number, or confusing similar street names. Assessing whether a geocoded result is actually right is surprisingly hard to automate. Coordable tries to address both issues with AI and analytics: Uses an LLM-based cleaner to normalize messy addresses (multi-country support). Automatically evaluates geocoding accuracy by comparing input and output like a human would. Lets you benchmark multiple providers (Google, HERE, Mapbox, Census, BAN, etc.) side by side. Includes a dashboard to visualize results, quality metrics, and exports. It’s not a new geocoder — it wraps existing APIs and focuses on data quality, comparison, and automation. It’s currently in beta with free credits. If you work with geocoding or address data, I’d love to hear how you handle these challenges and what kind of analytics would be most useful to you. https://coordable.co/ October 16, 2025 at 11:11PM
Wednesday, October 15, 2025
Show HN: Shorter – search for shorter versions of your domain https://ift.tt/6UVtN0g
Show HN: Shorter – search for shorter versions of your domain https://shorter.dev October 16, 2025 at 07:29AM
Show HN: Specific (YC F25) – Build backends with specifications instead of code https://ift.tt/dsRXA3h
Show HN: Specific (YC F25) – Build backends with specifications instead of code Hi folks! Iman and I (Fabian) have been building Specific for a while now and are finally opening up our public beta. Specific is a platform for building backend APIs and services entirely through natural-language specifications and tests, without writing code. We then automatically turn your specs into a working system and deploy it for you, along with any infrastructure needed. We know a lot of developers who have already adopted spec-driven development to focus on high-level design and let coding agents take care of implementation. We are attempting to take this even further by making the specs themselves the source of truth. Of course, we can’t blindly trust coding agents to follow the spec, so we also support adding tests that will run to ensure the system behaves as expected and to avoid regressions. There is so much ground to cover, so we are focusing on a smaller set of initial features that in our experience should cover a large portion of backends: - An HTTP server for each project. Authentication can be added by simply stating in the spec how you want to protect your endpoint. - A database automatically spun up and schema configured if the spec indicates persistence is needed. - External APIs can be called. You can even link out to API docs in your specs. You currently can’t see the generated code, but we are working on enabling it. Of course, we don’t claim any ownership of the generated code and will gladly let you export it and continue building elsewhere. Specific is free to try and we are really eager to hear your feedback on it! Try it here: https://ift.tt/UliprDc https://specific.dev/ October 15, 2025 at 10:51PM
Show HN: Pxxl App – A Nigerian Alternative to Vercel, Render, and Netlify https://ift.tt/KDZU7uq
Show HN: Pxxl App – A Nigerian Alternative to Vercel, Render, and Netlify Hi HN, I built Pxxl App — a free web hosting and deployment platform for developers in Nigeria and beyond. It’s a Nigerian alternative to Vercel, Render, and Netlify, designed for those who want a simple, fast, and barrier-free way to host both frontend and backend apps. With Pxxl App, you can connect your Git repo and deploy in seconds — no credit card, no limits. You’ll get a live subdomain like yourapp.pxxl.pro, automatic builds, and continuous deployment. It supports: • Frontend frameworks: React, Next.js, Vue, Svelte, and more • Backend projects: Node.js, PHP, and Python • Features like environment variables, CI/CD, and instant rollback The goal is to make cloud deployment accessible to African and global developers without the typical payment or region restrictions. It’s completely free to start, and I’d love to hear feedback from the HN community on how to improve it or what features you’d want next. Check it out: https://pxxl.app https://pxxl.app October 15, 2025 at 11:55PM
Tuesday, October 14, 2025
Show HN: Wispbit – Keep codebase standards alive https://ift.tt/j7xcr8M
Show HN: Wispbit – Keep codebase standards alive Hey HN! Ilya and Nikita here. We're building wispbit ( https://wispbit.com ) - a tool that helps keep codebase standards alive. With the help of AI coding tools, engineers are writing more code than ever. Code output has increased, but the tooling to manage this hasn't improved. Background agents still write bad code, and your IDE still writes slop without the right context. So we built wispbit. It works by scanning your codebase for patterns you already use, and coming up with rules. Rules are kept up to date as standards change, and you can edit rules any time. You can enforce these rules during code review, and because we have this rules system, you can run a CLI locally to review using these rules. You can think of it as a portable rules file that you can bring anywhere. We put a lot of work into making a system that produces good rules and avoids slop. For repository crawling, we have an agent that dispatches subagents, similar to Anthropic's research agent. These subagents will go through and look for common patterns within modules and directories, and report back to the main agent, which synthesizes the results. We also do a historical scan on your pull request comments, determine which ones were addressed, filter out comments that wouldn't make a good rule, and use that to create or update rules. Our early users are seeing 80%+ resolution rates, meaning that 80% of comments that wispbit makes are resolved. Long-term, we see ourselves being a validation layer for AI-written code. With tools like Devin and Cursor, we find ourselves having to re-prompt the same solution many times. We still don't know the long-term implications on AI-assisted codebases, so we want to get in front of that as soon as possible. We've opened up signups for free to HN folks at https://wispbit.com . We're also around to chat and answer questions! https://wispbit.com October 15, 2025 at 01:22AM
Show HN: GoHPTS-TCP/UDP Transparent Proxy with ARP Spoofing and Traffic Sniffing https://ift.tt/e8yrRQY
Show HN: GoHPTS-TCP/UDP Transparent Proxy with ARP Spoofing and Traffic Sniffing https://ift.tt/gWFw8fz October 14, 2025 at 08:19PM
Show HN:I built a free AI tool that scans and sorts financial news for traders https://ift.tt/7lxBVG2
Show HN:I built a free AI tool that scans and sorts financial news for traders https://www.fxradar.live/ October 14, 2025 at 10:56PM
Show HN: Ark v0.6.0 – Go ECS with new declarative event system https://ift.tt/ba7oZPn
Show HN: Ark v0.6.0 – Go ECS with new declarative event system Ark is a high-performance Entity Component System (ECS) library for Go. Ark v0.6.0 introduces a new event system built around lightweight, composable observers. These allow applications to react to ECS lifecycle changes like entity creation/removal, component updates, relation changes using declarative filters and callbacks. Observers follow the same patterns as Ark’s query system, making them easy to integrate and reason about. Custom events are also supported. They can be emitted manually and observed with the same filtering logic, making them ideal for modeling domain-specific interactions such as input handling, and other reactive game logic. As a new performance-related feature, filters and queries are now concurrency-safe and can be executed in parallel. This release also includes a load of performance improvements, from faster archetype switching over optimized query and table creation to improved performance of bitmask operations. The new World.Shrink method helps reclaim unused memory in dynamic workloads. Docs have been expanded with a full guide to the event system, examples for both built-in and custom events, and an Ebiten integration example. A cheat sheet for common operations has been added. Finally, Ark now has 100% test coverage. Changelog: https://ift.tt/5g6OGKu Repo: https://ift.tt/mrwDyXd Would love feedback from anyone building games, simulations, or ECS tooling in Go. https://ift.tt/mrwDyXd October 14, 2025 at 12:34PM
Monday, October 13, 2025
Show HN: Wordle but you have to predict your score before playing https://ift.tt/7BcqaPH
Show HN: Wordle but you have to predict your score before playing If you want to compare your results feel free to join the HN team. However you can also play without signing up for an account. Its a SvelteKit application running on Cloudflare Workers. Would love some feedback on the idea and execution! https://ift.tt/UR3fJzI October 14, 2025 at 08:56AM
Show HN: AI Toy I worked on is in stores https://ift.tt/AMUb2G1
Show HN: AI Toy I worked on is in stores Alt link: https://ift.tt/BqH7CWJ Video demo: https://www.youtube.com/watch?v=0z7QJxZWFQg The first time I talked with AI santa and it responded with a joke I was HOOKED. The fun/nonsense doesn't click until you try it yourself. What's even more exciting is you can build it yourself: libpeer: https://ift.tt/kB7QVLs pion: https://ift.tt/xgGMJ6F Then go do all your fun logic in your Pion server. Connect to any Voice AI provider, or roll your own via Open Source. Anything is possible. If you have questions or hit any roadblocks I would love to help you. I have lots of hardware snippets on my GitHub: https://ift.tt/LPc2Wk1 . https://ift.tt/Xptiadc October 12, 2025 at 07:45PM
Show HN: Make AI text sound human https://ift.tt/wCuQcV4
Show HN: Make AI text sound human Transform ChatGPT, Claude, and Gemini text into natural, human-like writing with a single click. https://refine.so October 14, 2025 at 02:39AM
Sunday, October 12, 2025
Show HN: Promptlet – Mac app to help you stop typing "ultrathink" over and over https://ift.tt/iU8hZ9t
Show HN: Promptlet – Mac app to help you stop typing "ultrathink" over and over https://ift.tt/GlJ6QeP October 13, 2025 at 02:29AM
Show HN: Pyreqwest – Powerful and fast Rust-reqwest based HTTP client for Python https://ift.tt/fSznkWp
Show HN: Pyreqwest – Powerful and fast Rust-reqwest based HTTP client for Python Python has lacked a batteries-included HTTP library that would have both async and sync clients. Httpx (httpcore), which has offered this, is unfortunately pretty much unmaintained and suffering from huge perf issues ( https://ift.tt/AVIfCqB ). I built pyreqwest HTTP client for Python that is fully Rust based on top of reqwest. It includes all features reqwest offers plus some more. Also including unit testing utilities (mocking, ASGI app support). Go check https://ift.tt/tO312QJ :) https://ift.tt/tO312QJ October 13, 2025 at 01:17AM
Show HN: I rewrote the express library to rust https://ift.tt/F8iZ0lI
Show HN: I rewrote the express library to rust https://shyam20001.github.io/rsjs/ October 13, 2025 at 12:05AM
Show HN: I built a simple ambient sound app with no ads or subscriptions https://ift.tt/lgswaO7
Show HN: I built a simple ambient sound app with no ads or subscriptions I’ve always liked having background noise while working or falling asleep, but I got frustrated that most “white noise” or ambient sound apps are either paywalled, stuffed with ads, or try to upsell subscriptions for basic features. So I made Ambi, a small iOS app with a clean interface and a set of freely available ambient sounds — rain, waves, wind, birds, that sort of thing. You can mix them, adjust volume levels, and just let it play all night or while you work. Everything works offline and there are no hidden catches. It’s something I built for myself first, but I figured others might find it useful too. Feedback, bugs, and suggestions are all welcome. https://ift.tt/LDNgTYk... https://ambisounds.app/ October 12, 2025 at 08:19PM
Saturday, October 11, 2025
Show HN: Vello's high-performance 2D GPU engine to .NET https://ift.tt/3KROWlH
Show HN: Vello's high-performance 2D GPU engine to .NET VelloSharp v0.5.0-alpha.1 is now available, delivering Vello’s high-performance 2D GPU engine to .NET atop the wgpu 3D backend. Integrates cleanly with Avalonia, WPF, and WinForms to modernize rendering pipelines without disruptive rewrites. https://ift.tt/eQu6TID October 12, 2025 at 03:32AM
Show HN: Solving the cluster 1 problem with vCluster standalone https://ift.tt/28J6Fjd
Show HN: Solving the cluster 1 problem with vCluster standalone vcluster is an open source tool for Kubernetes multi tenancy and over the years it has matured to have hosted controlplane virtual cluster, shared virtual clusters but the host cluster problem was always there. With vcluster standalone, you can now create the first cluster also with the same developer experience and consolidate the multiple vendor problem. With this, you can now use vcluster for entire multi tenancy spectrum. Feel free to discuss, happy to answer any questuons. https://ift.tt/G4qQhol October 8, 2025 at 10:20PM
Show HN: Sprite Garden - HTML Canvas 2D sandbox and farming https://ift.tt/m0uoE3W
Show HN: Sprite Garden - HTML Canvas 2D sandbox and farming Sprite Garden: https://kherrick.github.io/sprite-garden/ A 2D sandbox exploration and farming game that runs entirely in the web browser. As a fully HTML, CSS, and JavaScript game, it is highly readable, hackable, and customizable. Included on "globalThis" is the "spriteGarden" global object with the game config and state readily available. Drawing with tiles is as easy as opening dev tools (use the menu in the browser as keyboard is captured), or entering the "Konami Code," for a full screen view and a map editor. - Share games from the world state manager - Explore unique procedurally generated biomes - Dig for resources like coal, iron, and gold - Use collected materials to place blocks and shape the world - Discover underground cave systems filled with resources - Plant and harvest different crops with "realistic" growth cycles Examples: - Preparing a QR Code to be mined: https://gist.github.com/kherrick/1191ae457e1f6e1a65031c38c2d... - Drawing a heart in the sky: https://gist.github.com/kherrick/3dc9af05bccc126b11cc26fb30a... - Entering the Konami Code (map editor / fullscreen): https://gist.github.com/kherrick/effbe1463d9b78da046f27c5d42... I'm unsure how the game should be taken further, or whether it should progress. Some potential ideas for the future include: - Input Box with JS Execution: Provide a safe, sandboxed input area in the game's UI where players can write small JS functions or scripts (instead of exposing it on globalThis). - API Exposure: Expose a controlled API or object representing game state and functions, like terrain manipulation, crop growth, player movement, to the user script so players can automate or modify behaviors. - Event Hooks: Allow players to register hooks into game events (e.g., world update, planting crops) where their custom code runs, enabling mods or custom automation. - Multiplayer: Use WebRTC to allow many players in the same world. - Actual gamification: Make reasons to play, health meter, powerups, plant combinations, enemies? - Better mobile controls: Currently on screen, no swiping for movement. - Easier building with blocks: Currently block position based on location of player. Also featured on: - Microsoft Store: https://ift.tt/wkTUiMA - Wayback Machine: https://ift.tt/CiRW6UH.... Feedback is highly welcome, and source is available at: https://ift.tt/9wRyxIX https://kherrick.github.io/sprite-garden/ October 12, 2025 at 03:15AM
Friday, October 10, 2025
Show HN: Multiple choice video webgame experiment https://ift.tt/fa0boDm
Show HN: Multiple choice video webgame experiment Hey all, just wanted to share a little game experiment. It's a rooms & keys kind of adventure with a lot of random deaths. It plays a Veo3-generated video in response to clicks, with Gemini used for coding. Prompting the videos was fun, but trying to vibe code everything was not. In the future I'll go back to using LLMs more sparingly for isolated functions, or at least try not to have it create anything that requires seeing the output to debug. https://ift.tt/qp3djhy October 10, 2025 at 10:39PM
Show HN: Modeling the Human Body in Rust So I Can Cmd+Click Through It https://ift.tt/8vTariF
Show HN: Modeling the Human Body in Rust So I Can Cmd+Click Through It I started this trying to understand two things: why my Asian friends turn red after drinking, and why several friends all seemed to have migraine clusters. I was reading medical papers and textbooks, but kept getting lost jumping between topics. I thought: what if I could just Cmd+Click through this like code? What if "ALDH2 gene" was actually clickable, and took me to the variant, the phenotype, the population frequencies? So I started modeling human biology in Rust with my Ralph agent (Claude in a loop, ty ghuntley). Turns out the type system is perfect for this. Every biological entity is strongly-typed with relationships enforced at compile time. After 1 day of agent coding: - 277 Rust files, ~95k lines of code - 1,561 tests passing - 13 complete organ systems - Genetics with ancestry-specific variants - Clinical pathology models Try it: git clone https://ift.tt/PmKbCk0 cd open_human_ontology cargo run --example ide_navigation_demo Then open `examples/ide_navigation_demo.rs` and Cmd+Click through: Understanding Asian flush: AsianGeneticVariantsCatalog::get_metabolic_variants() // Click through to: // → ALDH2 gene on chromosome 12q24.12 // → rs671 variant (Glu504Lys) // → 40% frequency in Japanese population // → Alcohol flush reaction // → 10x esophageal cancer risk with alcohol // → Acetaldehyde metabolism pathway Understanding migraines: Migraine { subtype: WithAura, triggers: [Stress, LackOfSleep, HormonalChanges], genetic_variants: ["rs2075968", "rs1835740"], ... } // Click through to: // → 17 migraine trigger types // → 12 aura symptom types // → Genetic risk factors // → Why clusters happen (HormonalChanges → Menstruation) Now I can actually navigate the connections instead of flipping through PDFs. Heart → CoronaryArtery → Plaque. VisualCortex → 200M neurons → NeuralConnection pathways. It's like Wikipedia but type-checked and with jump-to-definition. This isn't production medical software - it's a learning tool. But it's way more useful than textbooks for understanding how biological systems connect. The agent keeps expanding it. Sometimes it OOMs but that's part of the fun. Tech: Rust, nalgebra, serde, rayon, proptest I am not a dr or medical professional this is for my education you can commit to it if you want to or review and open some PR's if you find wrong information or want to add references. https://ift.tt/PmKbCk0 October 10, 2025 at 11:29PM
Show HN: Iframetest.com https://ift.tt/NzsrhWw
Show HN: Iframetest.com https://iframetest.com/ October 6, 2025 at 03:25PM
Thursday, October 9, 2025
Show HN: Created macOS app to help you keep your distance from your screen https://ift.tt/PdJFER7
Show HN: Created macOS app to help you keep your distance from your screen Hey everyone, If you're anything like me, you spend a good chunk of your day (and night) on your computer. I often find that when I'm zoned in, my posture gets worse and worse and my face ends up being really close to the screen. And over a course of a workday, when I finally unplug, my eyes have a hard time focusing on things that are far away. This has become a big enough problem for me that I decided to create an app to help me keep my face far enough from the screen. Now, I could've gone with a simple notification with a timer built into it but, as with all reminder notification, they soon become noise for me and I end up just dismissing it. I needed something to actively force me to move back. Which is where FarSight comes in. It uses your camera to gauge your distance and blurs the entire screen if it detects that you are getting close enough for a certain period of time. I made it so that it won't be extremely annoying and disruptive (e.g. blurring the screen every time you cross the line) but just enough of a nuisance to be helpful. I've been using it everyday since creating it and it's definitely helped me with eye strain, double vision, and surprisingly, my posture as well. I'm not sure if I'll keep it free forever but I wanted to release it first to ask for feedback. I only have the app in MacOS so if it has enough interest, I'll invest into making Windows counterpart. https://ift.tt/rq9nJIW... Also, in case anyone is wondering, no data is collected and the snapshots during the app's usage are not saved but only used to calculate the distance. October 10, 2025 at 01:57AM
Show HN: SDF-Field Synthesis – rendering SDF without ray marching https://ift.tt/Ef1xjI9
Show HN: SDF-Field Synthesis – rendering SDF without ray marching Hello, HN! I have developed a new approach to rendering the scenes described by the SDF, which completely abandons the classic ray marching. The idea is simple: instead of searching for ray intersections, we consider the scene as a potential field of view, where the source is the SDF itself. The solution of the field equation gives zero isosurfaces — they correspond to the visible boundaries of objects. Advantages: - natural coherence between pixels, - analytical continuation for infinite resolution, - adaptive computational load depending on the complexity of the scene, - all the advantages of SDF are preserved (procedurality, Boolean operations, compactness). The full description of the idea with basic formulas and concept is published on Zenodo: https://ift.tt/I8BpmoG It will be interesting to hear the opinion of the community: - How feasible is the method? - What potential applications do you see? - Are there any similar approaches that I might have missed? https://ift.tt/j437XcA October 9, 2025 at 09:12PM
Show HN: I'm starting a fund that runs on luck, caffeine, and denial https://ift.tt/HJn0KoR
Show HN: I'm starting a fund that runs on luck, caffeine, and denial ngl afora is a cool name. this was just for fun. the message i wanna pass on is after $POLY raising $2B from $ICE, the signal that markets really are everything is true. cheers to that lads! have a great day. https://ift.tt/NFwQm8Y October 10, 2025 at 12:18AM
Wednesday, October 8, 2025
Show HN: Spica – OSS Tool to Generate Infinite Length Sora-2 Videos https://ift.tt/y02nLHb
Show HN: Spica – OSS Tool to Generate Infinite Length Sora-2 Videos https://ift.tt/h9IJyrb October 9, 2025 at 12:04AM
Show HN: Open-source UI for running multiple coding agents https://ift.tt/9HtXs7y
Show HN: Open-source UI for running multiple coding agents Hi HN! I've recently been finding productivity in running parallel CLI coding agents(after not believing in them initially). After having to do a ton of git stashing and branch fumbling, I decided I needed to something to more ergonomically run these agents in their own dedicated spaces. I tried a lot of the existing products but they either were too convoluted or flat out didn't work. Some of them also seem to roll their own chat UI which I don't think is the right approach, I wanted to something to lightly wrap my terminal sessions. So I built FleetCode! It uses git worktrees and let's you run multiple agents at once. It's made my multi agent coding workflow much easier. It's free and open source, would love some feedback! https://ift.tt/IKCkpMZ October 8, 2025 at 11:30PM
Show HN: KI Song Erstellen Kostenlos – AI Music Generator FüR Deutsche Musik https://ift.tt/sucAXQV
Show HN: KI Song Erstellen Kostenlos – AI Music Generator FüR Deutsche Musik Kostenloser KI-Musikgenerator für deutsche Songs. Text rein → professioneller Song in wenigen Minuten. Gebaut für Content Creator, die Copyright-freie Musik brauchen. https://ift.tt/QFSjZC7 GitHub: https://ift.tt/5mMFb4B Probiert es aus! https://ift.tt/QFSjZC7 October 8, 2025 at 10:26PM
Tuesday, October 7, 2025
Show HN: DidMySettingsChange – A tool that checks changed windows settings https://ift.tt/gnTSsle
Show HN: DidMySettingsChange – A tool that checks changed windows settings Microsoft has been under heavy scrutiny with how they manage Windows over the years, particularly concerning privacy and telemetry settings. Many users find that after disabling certain settings, these settings are mysteriously re-enabled after updates or without any apparent reason. DidMySettingsChange is a Python script designed to help users keep track of their Windows privacy and telemetry settings, ensuring that they stay in control of their privacy without the hassle of manually checking each setting. Features Comprehensive Checks: Automatically scans all known Windows privacy and telemetry settings. Change Detection: Alerts users if any settings have been changed from their preferred state. Customizable Configuration: Allows users to specify which settings to monitor. Easy to Use: Simple command-line interface that provides clear and concise output. Logs and Reports: Generates detailed logs and reports for auditing and troubleshooting. https://ift.tt/xJP3iyt October 6, 2025 at 04:19AM
Show HN: I'm building a browser for reverse engineers https://ift.tt/XiD6UMP
Show HN: I'm building a browser for reverse engineers https://ift.tt/azjlHd2 October 6, 2025 at 09:02PM
Show HN: Gotask, a simple task manager CLI built using Golang https://ift.tt/fEIsWeS
Show HN: Gotask, a simple task manager CLI built using Golang Hey folks, Gotask is a simple golang CLI I made to explore some aspects of the Go programming language. https://ift.tt/wrEUtT7 October 8, 2025 at 12:20AM
Show HN: Rethinking audit trails in Django (structured and database-free) https://ift.tt/mKT6que
Show HN: Rethinking audit trails in Django (structured and database-free) We were using django-easy-audit to track CRUD and API events in Django. The problem: every action meant an extra database write, which slowed transactions, increased infra costs, and left us with audit tables that weren’t great for analytics. I built django-activity-audit (PyPI) to fix this: - Extends Django’s logging system with custom AUDIT and API levels. - Captures CRUD + API request/response events as structured JSON logs. - Vector tails the logs and ships them into ClickHouse. - Grafana makes them queryable and visual. This removes the extra DB writes, gives structured data ready for analysis, and keeps costs down. Curious — how are others handling audit logging in Django (or other frameworks)? Do you log it, write it to a DB, or something else entirely? https://ift.tt/HKtIX4r October 7, 2025 at 10:25PM
Monday, October 6, 2025
Show HN: TinqerJS– LINQ-inspired QueryBuilder for TypeScript + Postgres/SQLite https://ift.tt/ApBJ3TQ
Show HN: TinqerJS– LINQ-inspired QueryBuilder for TypeScript + Postgres/SQLite https://tinqerjs.org October 6, 2025 at 08:58PM
Show HN: I've build a platform for writing technical/scientific documents https://ift.tt/lXCeNGM
Show HN: I've build a platform for writing technical/scientific documents https://ift.tt/vpBfcQz October 6, 2025 at 04:28PM
Show HN: I Built a Transcription CLI Because Uploading 4GB Videos Was Killing Me https://ift.tt/krKzG9g
Show HN: I Built a Transcription CLI Because Uploading 4GB Videos Was Killing Me https://ift.tt/5CwT0vF October 6, 2025 at 11:52PM
Show HN: Meihus – Mortgage calculator that shows early payment impact monthly https://ift.tt/YNIXRTr
Show HN: Meihus – Mortgage calculator that shows early payment impact monthly I built it because I needed a way to see how much will my early payments save in interest overtime. Most of the websites I found are solely made for the US, which does not help my case. So I built one that has a bit more flexibility in terms of loan data. Any feedback is appreciated. https://meihus.com October 6, 2025 at 12:00PM
Sunday, October 5, 2025
Show HN: A Node.js CLI tool to generate ai.txt, llms.txt, robots.txt, humans.txt https://ift.tt/SnvxpD3
Show HN: A Node.js CLI tool to generate ai.txt, llms.txt, robots.txt, humans.txt https://ift.tt/rg9CU7n October 6, 2025 at 09:28AM
Show HN: High-fidelity, compact, and real time rendering of university campus https://ift.tt/cerdJRX
Show HN: High-fidelity, compact, and real time rendering of university campus Technical thread: https://ift.tt/sGfVOFY https://hoanh.space/aalto/ October 6, 2025 at 05:21AM
Show HN: A Vectorless LLM-Native Document Index Method https://ift.tt/MvtiYy1
Show HN: A Vectorless LLM-Native Document Index Method The word "index" originally came from how humans retrieve info: book indexes and tables of contents that guide us to the right place in documents. Computers later borrowed the term for data structures: e.g., B-trees, hash tables, and more recently, vector indexes. They are highly efficient for machines; but abstract and unnatural: not something a human, or an LLM, can understand and directly use as a reasoning aid. This creates a gap between how indexes work for computers and how they should work for models that reason like humans. PageIndex is a new step that "looks back to move forward". It revives the original, human-oriented idea of an index and adapts it for LLMs. Now the index itself (PageIndex) lives inside the LLM's context window: the model sees a hierarchical table-of-contents tree and reasons its way down to the right span, much like a person would retrieve information using a book's index. PageIndex MCP shows how this works in practice: it runs as a MCP server, exposing a document's structure directly to LLMs/Agents. This means platforms like Claude, Cursor, or any MCP-enabled agent or LLM can navigate the index themselves and reason their way through documents, not with vectors/chunking, but in a human-like, reasoning-based way. https://ift.tt/2YKdSuH October 5, 2025 at 10:09PM
Saturday, October 4, 2025
Show HN: An open-source, RL-native observability framework we've been missing https://ift.tt/Hu7CcRF
Show HN: An open-source, RL-native observability framework we've been missing The RL ecosystem is maturing— verifiers are standardizing how we build and share environments. However, as it grows, we need observability tooling that actually understands RL primitives. Running RL experiments without visibility into rollout quality, reward distributions, or failure modes is a waste of time. Monitor provides live tracking, per-example inspection, and programmatic access—see what's happening during runs and debug what went wrong afterward. https://ift.tt/OptYXSZ October 5, 2025 at 03:05AM
Show HN: World Amazing Framework: Like Django for Civilization https://ift.tt/rgaihsT
Show HN: World Amazing Framework: Like Django for Civilization Any initial thoughts? This framework is meant to be a tool for construction, so if you want to play around with it for creating potential specific implementations, you can drop the contents of the website, the GitHub README, and the entire overview.md into an AI chat, and that should be enough to use the framework, at least conceptually. Would y'all want me to pre-prime a chat in Google AI Studio with the full context of the plan and some basic direction for discourse? I can share a link to a ready-to-go environment. The core documentation should answer most mechanical questions. And if you feed the docs into an AI chat, you can ask it any question you may have, or to simply ask it to explain something in different ways, or hypothesize solutions to any world issue, either systemic or regional. Gemini Pro 2.5 can take the full doc in one prompt, and its ability to co-create ideas is remarkable. I've been using it mostly through the AI Studio interface. Much of the overview is as much my work as it is a synthesis of my collaboration with Gemini Pro 2.5, ChatGPT-4o, and some early contributions from GPT-4 about a year ago. Before LLMs, I was building out pamphlet-style pages on a website (that are up at whomanatee.org, which is the base wrapper implementation of the framework), and I was planning to use them as talking points. I was anticipating that much of the deep thinking would have to happen in slow, public discourse. With LLMs, I've been able to stress-test these ideas from every possible angle, using any past event or theory to see if the framework could withstand scrutiny. At one point, a model argued that Adam Smith would have rejected this idea as fantasy. So I worked with it to develop an economic plan that "synthetic Adam" praised. It's incredible that we now have the ability to get synthesized thoughts from almost any perspective. You could ask it, "What would Barack Obama think of this plan? And using the framework, what would be your response to any hesitations he may have?" And it responds with incredible analysis, synthesis, and feedback. https://ift.tt/h8O4p1U October 5, 2025 at 03:44AM
Show HN: Run – a CLI universal code runner I built while learning Rust https://ift.tt/igDw0E3
Show HN: Run – a CLI universal code runner I built while learning Rust Hi HN — I’m learning Rust and decided to build a universal CLI for running code in many languages. The tool, Run, aims to be a single, minimal dependency utility for: running one-off snippets (from CLI flags), running files, reading and executing piped stdin, and providing language-specific REPLs that you can switch between interactively. I designed it to support both interpreted languages (Python, JS, Ruby, etc.) and compiled languages (Rust, Go, C/C++). It detects languages from flags or file extensions, can compile temporary files for compiled languages, and exposes a unified REPL experience with commands like :help, :lang, and :quit. Install: cargo install run-kit (or use the platform downloads on GitHub). Source & releases: https://ift.tt/CxeTrXA I used Rust while following the official learning resources and used AI to speed up development, so I expect there are bugs and rough edges. I’d love feedback on: usability and UX of the REPL, edge cases for piping input to language runtimes, security considerations (sandboxing/resource limits), packaging and cross-platform distribution. Thanks — I’ll try to answer questions and share design notes. https://ift.tt/CxeTrXA October 5, 2025 at 12:04AM
Friday, October 3, 2025
Show HN: Beacon (open source) – Built after AWS billed me 700% more for RDS https://ift.tt/70aE5X6
Show HN: Beacon (open source) – Built after AWS billed me 700% more for RDS I've been hosting my side project on AWS. I was paying an okay price for not managing infrastructure at all. I moved everything to AWS Ligthsail after my startup credits run out. The project was initially a success and made several thousand euros per month in revenue. Then came covid with new regulations, and suddenly my customers were non existent (the problem it solved was no longer there). After that it was not making me money, I was paying it from my own pocket to maintain it, thinking maybe it will come back. Then one day, after some ignored spam AWS emails, I got a huge charge on my card, along with a bill from AWS. The charge was orders of magnitude higher than the previous charges."WTF??" I said to myself while rushing to log into the dashboard to see what the issue was. No DDoS, no misconfiguration, nothing unusual. I logged into the root account to look at the billing page, and there it was:RDS PostgreSQL legacy fee ~€200 because I did not upgrade to Postgres 16 (from 13). I was baffled. I paid monthly €25 (27% tax included) for the smallest RDS instance, then I see this monster fee for something I think should cost maybe €2. I mean AWS just has to run it in a different environment. For €200 I could buy them a new server to run it for me. That's when I had the realization: "I have a spare Raspberry Pi 3, I'll just host everything on that. That will be for free." But self-hosting came with it's own challenges, especially on a resource-constrained device. I needed better tools to deploy and monitor my application. SSH-ing into the Raspberry Pi every time I wanted to deploy a newer version was a pain in the ass. So was debugging issues. Existing deployment and monitoring solutions were either too expensive, too complex, or didn't work well with resource-constrained devices like Raspberry Pi. Examples: * Grafana/Prometheus for monitoring: Over-engineered for my needs. * OpenSearch/ELK for logs: A nightmare on low-resource devices. * Metabase for dashboards: A ram hungry monster that eats up more resources than if I hosted 100 applications. And to access the db remotely opening a port and putting it behind Cloudflare Zero Trust is much easier than setting up Metabase. So I decided to build my own deployment and montitoring agent, and why not make it opensource? The agent can currently deploy applications from github by polling release tags, monitor device metrics, alert when thresholds are reached, forward logs to cloud dashboard. It's still in development, with features improving every week. If you are interested, give it a start on Github. https://beaconinfra.dev October 4, 2025 at 01:52AM
Show HN: Was pissed about Google Docs, So I made an Text Editor myself https://ift.tt/tzxJWgU
Show HN: Was pissed about Google Docs, So I made an Text Editor myself It’s been a while since I’ve started to write a book . The process of creation of it has not been easy , first because I’m not a writer , I’ve created well though out internet posts here and there, which ended up creating my first book. It was a good experience , but then I’ve started to think that a book that just gathered my thoughts online it’s not entirely “writing” a book , I needed more. And than I’ve opened google docs and start typing. Then I started to figure out what I wanted to write: should it be a fantasy story, a self-biography, or an observation of the world? I believe most writers have this figured out beforehand, but not me. I began writing pieces to see if they would fit together and make sense. I started gathering philosophical anecdotes based on my core beliefs and sensed something brewing. When I finally decided what the book would be about, and what I wanted to write, the type of writing I wanted to do, I saw an already sizable document with ideas scattered throughout it. That was good for me, as I could just join the pieces, but I didn’t want to be trapped in writing that could be repetitive. I wanted to have the ideas, philosophy, the whole reason why the book is like this, stored in a place I could easily access. I'm planning to use AI as a memory dump, where I can add information during a conversation. Then, whenever I consult it, I can check if I've already written something and if it reflects the temper and pace I want for my book. Everything seems fine, but we encountered a few problems. First, the AI's writing was a conundrum of errors. I could gain assistance and a sense of what to write, but the AI itself, due to our prolonged interchange, started to hallucinate and produce nonsense or "forget" our conversation. The second issue was that the AI couldn't consistently verify what was already written. As the text grew larger, the context window began to shrink, and the more I used the AI tool, the less helpful it became. So I decided to search for a tool that could do what I wanted. I found elements in each of the products I've used: some were extremely satisfying to write with, others had good features to enhance text, some allowed me to organize my book by scattering ideas effectively, and still others used AI for correction and proofreading tasks. The solutions for this market are diverse and offer numerous approaches. I could easily transition between tools, but I wanted something unified to keep my writing process in one place. That’s why I created this text editor and called it SourcePilot. It’s a tool that identifies your writing style as you write, allowing you to add notes, sources, and videos, and to use them as context for the AI, enabling more nuanced outputs tailored to your writing. It was interesting to build, and I’m providing a link you can try. It’s a desktop app, and you can use it for free, depending on the hardware you have. I’m looking for people who could give me feedback on what's wrong with it. People who could not install it (I’ve built it on Mac and could not test Linux and Windows), or have problems logging in. I keep getting loads of problems because I’m using the tool right now as I write this text. I'm planning to launch a new version soon, featuring an anti-slop algorithm I’ve developed, along with document branching. I just want to see if there are people interested in using it at the moment. If there aren't users, that's fine. I think I’ve made something for myself anyway. :) Thank you for your attention if you made it this far. You’re greatly appreciated. Cheers! https://sourcepilot.co/ October 4, 2025 at 01:28AM
Show HN: A VS Code Extension for Genesis DB – The event sourcing database https://ift.tt/YPIzxkR
Show HN: A VS Code Extension for Genesis DB – The event sourcing database I've been working on an extension that brings the full Genesis DB experience directly into Visual Studio Code. Genesis DB itself is a production-ready event-sourcing database engine, and with this extension you can now: - Manage multiple connections (dev, staging, prod) with token-based authentication - Explore events via a built-in Event Explorer UI - Commit new events, run GDBQL queries, and instantly view results - Manage schemas (register, browse, validate) directly in the editor - Use built-in GDPR features like event erasure, without leaving VS Code The goal: no more context switching between CLI, APIs, and docs, everything integrated into the editor you already use. You can check out the extension here: https://marketplace.visualstudio.com/items?itemName=patricec... I'd love to hear what you think: Is this something that would help your workflow? What would you expect from a database extension in VS Code? https://ift.tt/9oqV6vh October 4, 2025 at 01:12AM
Show HN: FLE v0.3 – Claude Code Plays Factorio https://ift.tt/d3i4og2
Show HN: FLE v0.3 – Claude Code Plays Factorio We're excited to release v0.3.0 of the Factorio Learning Environment (FLE), an open-source environment for evaluating AI agents on long-horizon planning, spatial reasoning, and automation tasks. == What is FLE? == FLE uses the game Factorio to test whether AI can handle complex, open-ended engineering challenges. Agents write Python code to build automated factories, progressing from simple resource extraction (~30 units/min) to sophisticated production chains (millions of units/sec). == What's new in 0.3.0 == - Headless scaling: No longer needs the game client, enabling massive parallelization! - OpenAI Gym compatibility: Standard interface for RL research - Claude Code integration: We're livestreaming Claude playing Factorio [on Twitch]( https://ift.tt/E7mHNWv ) - Better tooling and SDK: 1-line CLI commands to run evaluations (with W&B logging) == Key findings == We evaluated frontier models (Claude Opus 4.1, GPT-5, Gemini 2.5 Pro, Grok 4) on 24 production automation tasks of increasing complexity. Even the best models struggle: - Most models still rely on semi-manual strategies rather than true automation - Agents rarely define helper functions or abstractions, limiting their ability to scale - Error recovery remains difficult – agents often get stuck in repetitive failure loops The performance gap between models on FLE correlates more closely with real-world task benchmarks (like GDPVal) than with traditional coding/reasoning evals. == Why this matters == Unlike benchmarks based on exams that saturate quickly, Factorio's exponential complexity scaling means there's effectively no performance ceiling. The skills needed - system debugging, constraint satisfaction, logistics optimization - transfer directly to real challenges. == Try it yourself == >>> uv add factorio-learning-environment >>> uv add "factorio-learning-environment[eval]" >>> fle cluster start >>> fle eval --config configs/gym_run_config.json We're looking for researchers, engineers, and modders interested in pushing the boundaries of agent capabilities. Join our Discord if you want to contribute. We look forward to meeting you and seeing what you can build! -- FLE Team https://jackhopkins.github.io/factorio-learning-environment/versions/0.3.0.html October 4, 2025 at 01:02AM
Thursday, October 2, 2025
Show HN: BetterBrain – Dementia prevention, covered by insurance https://ift.tt/saYyCzU
Show HN: BetterBrain – Dementia prevention, covered by insurance Hey all! I’ve been building BetterBrain for the past few months, which is the first dementia prevention program entirely covered by insurance. BetterBrain combines expert clinicians, comprehensive testing and state of the art AI - and for many insurance plans is $0. Research shows that dementia can be detected up to 20 years in advance. Despite this, many people at risk of dementia overlook regular brain health assessments. Many members of our founding team have family members affected by neurodegenerative disease. We’re also hiring aggressively if anyone is interested in changing the future of treating neurodegenerative disease. Would love to talk to anyone interested https://ift.tt/ivGNqV9 https://ift.tt/ivGNqV9 October 3, 2025 at 07:33AM
Show HN: YNOT – Free, Open-Source YouTube Downloader https://ift.tt/OgwRA6y
Show HN: YNOT – Free, Open-Source YouTube Downloader Hey HN! I built YNOT, a simple cross-platform YouTube downloader with a GUI. It's powered by yt-dlp and completely free/open-source (WTFPL license). Key features: - Simple GUI - just paste URL and download - Downloads HD/4K videos - Cross-platform (Windows, macOS, Linux) - No ads, no tracking, completely private - Lightweight and fast GitHub: https://ift.tt/6KXsdzp I'd love to hear your feedback and suggestions! https://james-see.github.io/ynot/ October 3, 2025 at 04:26AM
Show HN: Uber for Flights https://ift.tt/F02HDeE
Show HN: Uber for Flights My friend and I built BookMyFlight to finally modernize flight search + booking. Why we built this: - Personalization. I fly the same route every month, and there’s no platform that knows my preferences so that I can open it, find and book my flight, and close it within a minute. - Booking is slow. I hate seeing a long clunky airline form each time I need to book. I want booking a flight to feel more like booking an Uber. How it works: 1. Optionally make an account and save your traveler preferences. Personally, I've specified my routine route as SFO to CLE and that I only want red-eye direct flights for this route. 2. Search for flights using chat or the search panel. Chat feels especially time-saving when you have preferences saved (e.g. I just say “search my routine trip"). 3. Once you find the flight you want, use the one-click book feature which books your flight directly with the airline. The first time you book a flight, you’ll have to fill out your traveler info, but you won't see that form after that. Notes: - Your booking is directly with the airline (this means when something goes wrong, you get direct support from the airline—not a third-party) - You can add your rewards numbers for each airline to keep earning points/status The ultimate goal is to create the best possible experience that every traveler wants, but that OTAs and airlines don’t care to create. Also very receptive to hearing pain points from frequent flyers; we think this space is really outdated and could use some innovation. Try it out and let us know what you think :) https://bookmyflight.ai October 3, 2025 at 01:29AM
Show HN: Enhance – A Terminal UI for GitHub Actions https://ift.tt/eXGxaPW
Show HN: Enhance – A Terminal UI for GitHub Actions I'm very excited to share what I've been working on lately! Introducing ENHANCE, a terminal UI for GitHub Actions that lets you easily see and interact with your PRs checks. It's available under a sponsorware model. Get more info on the site: -> https://ift.tt/WDxF1b5 This is an attempt to make my OSS development something sustainable. Happy to hear feedback about the model as well as the tool! Cheers! https://ift.tt/ea2GmYK October 3, 2025 at 12:49AM
Wednesday, October 1, 2025
Show HN: Rostra is a P2P (f2f) social network https://ift.tt/5RynIiq
Show HN: Rostra is a P2P (f2f) social network A public instance is available at https://rostra.me/ . It will default to showing the interface from the perspective of my own identity, in a read-only mode. Click "Logout" and then "Random" to generate your own identity to play with. https://app.radicle.xyz/nodes/radicle.dpc.pw/rad%3AzzK566qFsZnXomX2juRjxj9K1LuF October 2, 2025 at 03:40AM
Show HN: Open-source project – HTTP cache and reverse proxy https://ift.tt/wh7GB2v
Show HN: Open-source project – HTTP cache and reverse proxy https://borislavv.github.io/advcache.dev/ October 1, 2025 at 01:11PM
Show HN: Ocrisp, One-Click RAG Implementation, Simple and Portable https://ift.tt/l1IaBeA
Show HN: Ocrisp, One-Click RAG Implementation, Simple and Portable https://ift.tt/sHc0ETa October 1, 2025 at 08:23PM
Show HN: Spit Notes – A songwriting app that keeps lyrics and audio together https://ift.tt/WyuUrLF
Show HN: Spit Notes – A songwriting app that keeps lyrics and audio together Any songwriter who uses the iOS Notes app to write their lyrics has a mess of New Recording 142 voice memos in their Voice Memos app. I made Spit Notes as basically the Notes app but with a built-in voice recorder that keeps your audio files neatly organized on the same line as your lyrics. Now you can quickly capture your song ideas while driving or when you wake up in the middle of the night without worrying about losing them in your pile of untitled voice memos. While you can attach audio to notes with other apps, adding and recording audio has a lot of friction, and often the layout of the audio elements in those apps are too pronounced to keep the text flowing seamlessly. This is not the case with Spit Notes. I've wanted this app to exist for years but I put off making it myself because I knew it would take me a lot of time to build manually without knowing Swift. In recent years I've been writing AI-assisted code, but with AI coding agents getting better and better, 3 months ago I decided to see if I could vibe code a full product. The code for this project was not AI-assisted, but human-assisted, with me providing the vision and feel of the app, while the AI agent takes that, makes a pass at the code base, and then I QA it, letting the AI know what worked/didn't work, and iterating. I started with paying for Cursor and using Opus 4 but after getting insanely good initial results with Opus 4 and seeing my cursor costs start to rise, I remembered this post https://ift.tt/KlYEw0z and took the plunge with Claude Code max $200 plan. This turned out to be an incredible value because it allowed me to use claude basically without limit. However, Gemini still had the biggest context window and as the project grew I had to use Gemini to create plans for big features and find deep bugs across all of the AI-generated modules. Pro tip: create broad reference files for you and the AI, like an ARCHITECTURE.md where you keep a human-readable version of the technical big picture. You can then reference that for the AI so it stays aligned with your current progression. Once the Codex CLI became available on homebrew it was a wrap. I cancelled my claude code max plan and have been happily coding without ever hitting any rate limits (other than when they made that update that accidentally reduced rate limits instead of increase them). Today, I pay for for chatgpt plus and gemini $20 plan and am able to clear most obstacles on the first or second prompt. I haven't tried opus 4.5 but since I am not really getting stuck with codex, for now I'll stick with that. https://ift.tt/k3P6rcd October 1, 2025 at 10:12PM
Subscribe to:
Comments (Atom)
Show HN: Easily visualize torch, Jax, tf, NumPy, etc. tensors https://ift.tt/anqfwJx
Show HN: Easily visualize torch, Jax, tf, NumPy, etc. tensors hey hn, wrote a python library for myself to visualize tensors. Makes learning...
-
Show HN: An AI logo generator that can also generate SVG logos Hey everyone, I've spent the past 2 weeks building an AI logo generator, ...
-
Show HN: I Made an AI Social Media Manager to Automate Content Creation Hey HN, I am a Solopreneur, and I love building apps to automate bor...
-
RoboPianist, a piano playing robot simulation in the browser https://ift.tt/zywcBo6 March 30, 2023 at 10:52PM