Friday, February 20, 2026

Thursday, February 19, 2026

Show HN: A small, simple music theory library in C99 https://ift.tt/8Xb75DL

Show HN: A small, simple music theory library in C99 https://ift.tt/NpKdaE5 February 20, 2026 at 04:24AM

Show HN: Hi.new – DMs for agents (open-source) https://ift.tt/GJpWvuR

Show HN: Hi.new – DMs for agents (open-source) https://www.hi.new/ February 20, 2026 at 02:50AM

Show HN: Astroworld – A universal N-body gravity engine in Python https://ift.tt/FctZTqp

Show HN: Astroworld – A universal N-body gravity engine in Python I’ve been working on a modular N-body simulator in Python called Astroworld. It started as a Solar System visualizer, but I recently refactored it into a general-purpose engine that decouples physical laws from planetary data.Technical Highlights:Symplectic Integration: Uses a Velocity Verlet integrator to maintain long-term energy conservation ($\Delta E/E \approx 10^{-8}$ in stable systems).Agnostic Architecture: It can ingest any system via orbital elements (Keplerian) or state vectors. I've used it to validate the stability of ultra-compact systems like TRAPPIST-1 and long-period perturbations like the Planet 9 hypothesis.Validation: Includes 90+ physical tests, including Mercury’s relativistic precession using Schwarzschild metric corrections.The Planet 9 Experiment:I ran a 10k-year simulation to track the differential signal in the argument of perihelion ($\omega$) for TNOs like Sedna. The result ($\approx 0.002^{\circ}$) was a great sanity check for the engine’s precision, as this effect is secular and requires millions of years to fully manifest.The Stack:NumPy for vectorization, Matplotlib for 2D analysis, and Plotly for interactive 3D trajectories.I'm currently working on a real-time 3D rendering layer. I’d love to get feedback on the integrator’s stability for high-eccentricity orbits or suggestions on implementing more complex gravitational potentials. https://ift.tt/Ee0cjKS February 20, 2026 at 01:27AM

Wednesday, February 18, 2026

Show HN: Wakapadi – Meet locals and travelers nearby and join free walking tours https://ift.tt/BnUX1pl

Show HN: Wakapadi – Meet locals and travelers nearby and join free walking tours Hi HN, I built Wakapadi after noticing that most travel tools focus on planning trips, but not on actually helping people connect once they arrive somewhere new. When traveling, it’s often hard to meet locals or other travelers unless you already know someone, join organized tours, or rely on chance. I wanted to make discovery more natural — seeing who’s nearby, joining free walking tours, and exploring cities together. Wakapadi currently allows users to: discover free walking tours see nearby travelers and locals who are open to meeting connect and chat before meeting explore cities in a more social way The project is still early, and I’m especially interested in feedback on: safety and privacy expectations what would make you comfortable meeting people while traveling features that would make this genuinely useful instead of another travel app Happy to answer any technical or product questions. https://ift.tt/vML98P7 February 18, 2026 at 11:40PM

Tuesday, February 17, 2026

Show HN: I'm launching a LPFM radio station https://ift.tt/lRImzXi

Show HN: I'm launching a LPFM radio station I've been working on creating a Low Power FM radio station for the east San Fernando Valley of Los Angeles. We are not yet on the broadcast band but our channel will be 95.9FM and our range can been seen on the homepage of our site. KPBJ is a freeform community radio station. Anyone in the area is encouraged to get a timeslot and become a host. We make no curatorial decisions. Its sort of like public access or a college station in that way. This month we launched our internet stream and on-boarded about 60 shows. They are mostly music but there are a few talk shows. We are restricting all shows to monthly time slots for now but this will change in the near future as everyone gets more familiar with the systems involved. All shows are pre-recorded until we can raise the money to get a studio. We have a site secured for our transmitter but we need to fundraise to cover the equipment and build out costs. We will be broadcasting with 100W ERP from a ridgeline in the Verdugos at about 1500ft elevation. The site will need to be off grid so we will need to install a solar system with battery backup. We are planning to sync the station to the transmit site with 802.11ah. I've built all of our web infrastructure using Haskell, NixOS, Terraform, and HTMX: https://ift.tt/1xYoVqZ This is a pretty substantial project involving a bunch of social and technical challenges and a shoe string budget. I'm feel pretty confident we will pull it off and make it a high impact local radio station. The station is managed by a 501c3 non-profit we created. We are actively seeking fundraising, especially to get our transmit site up and running. If you live in the area or want to contribute in any way then please reach out! https://www.kpbj.fm/ February 18, 2026 at 01:45AM

Show HN: AsteroidOS 2.0 – Nobody asked, we shipped anyway https://ift.tt/5lendAE

Show HN: AsteroidOS 2.0 – Nobody asked, we shipped anyway https://ift.tt/l8kR0pi February 18, 2026 at 12:54AM

Show HN: I curated 130 US PDF forms and made them fillable in browser https://ift.tt/Qd3v5Su

Show HN: I curated 130 US PDF forms and made them fillable in browser Hi HN! I built SimplePDF 7 years ago, with the vision from day one to help get rid of bureaucracy (I'm from France, I know what I'm talking about) Fast forward to this week where I finally released something I had on my mind for a long time: a repository of the main US forms that are ready to be filled, straight from the browser, as opposed to having to find a PDF tool online (or local). I focused on healthcare, ED, HR, Legal and IRS/Tax for now. On the tech-side, it's SimplePDF all the way down: client-side processing (the data / documents stay in your browser). I hope you find the resource useful! NiP https://ift.tt/reWZTcz February 18, 2026 at 12:03AM

Monday, February 16, 2026

Show HN: Nerve: Stitches all your data sources into one mega-API https://ift.tt/5enHZNG

Show HN: Nerve: Stitches all your data sources into one mega-API Hi HN! Nerve is a solo project I've been working on for the last few years. It's a developer tool that stitches together data from multiple sources in real-time. A lot of high-leverage projects (AI or otherwise) involve tying data together from multiple systems of record. This is easy enough when the data is simple and the sources are few, but if you have highly nested data and lots of sources (or you need things like federated pagination and filtering), you have to write a lot of gnarly boilerplate that's brittle and easy to get wrong. One solution is to import all your data into a central warehouse and just pull it from there. This works, but 1) you need a warehouse, 2) you have an extra copy of the data that can get stale or inconsistent, 3) you need to write and manage pipelines/connectors (or outsource them to a vendor), and 4) you're adding an extra point of failure. Nerve lets you write GraphQL-style queries that span multiple sources; then it goes out and pulls from whatever source APIs it needs to at query-time - all your source data stays where it is. Nerve has pre-built bindings to external SAAS services, and it's straightforward to hook it into your internal sources as well. Nerve is made for individual developers or two-pizza teams who: -Are building agents/internal tools -Need to deal with messy data strewn across different systems -Don't have a data team/warehouse at their disposal, (or do, but can't get a slice of their bandwidth) -Want to get to production as quickly as possible Everything you see in the demo is shipped and usable, but I'm adding a little polish before I officially launch. In the meantime, if you have a project you'd like to use Nerve on and you want to be a beta user, just drop me a line at mprast@get-nerve.com (it's free! I'll just pop in from time to time to ask you how it's going and what I can improve :) ) If you want to get an email when Nerve is ready from prime-time, you can sign up for the waitlist at get-nerve.com. Thanks for reading! (EDIT: Nerve is desktop only! I'll put up a gate on the site saying as much.) https://ift.tt/mRuogij February 15, 2026 at 04:37AM

Show HN: AsdPrompt – Vimium-style keyboard navigation for AI chat responses https://ift.tt/QPaSFtg

Show HN: AsdPrompt – Vimium-style keyboard navigation for AI chat responses I use Claude throughout the day and kept getting annoyed by the same thing: selecting text from responses with the mouse. Overshoot, re-select, copy, click input, paste. Especially bad in long conversations where you want to reference something from 30 turns ago. asdPrompt is a Chrome extension that adds hint-based navigation (like Vimium) to AI chat interfaces. Cmd+Shift+S activates the overlay, hint labels appear next to every text block. Type a letter to select a block, then keep typing to drill down: block → sentence → word. Enter copies, or you can press an action key (e, d, x) to inject a follow-up prompt ("elaborate on [selection]") directly into the chat input. Works on claude.ai, chatgpt.com, and gemini.google.com. Adapts to light/dark themes. Free. Built the initial MVP in 2 days using Claude Code — the adapter architecture, NLP segmentation pipeline, and Playwright test harness would have taken a month without it. Tech details for the curious: site-specific DOM parsers behind an adapter interface, text segmentation via compromise.js with regex fallbacks for technical content (paths, camelCase break NLP libraries), bounding rectangles calculated via Range API + TreeWalker, overlay isolated in Shadow DOM. Tested with Playwright visual regression. The landing page has an interactive tutorial where you can try the full drill-down mechanic without installing. Happy to talk about the implementation. https://asdprompt.com/ February 16, 2026 at 10:58PM

Sunday, February 15, 2026

Show HN: Please hack my C webserver (it's a collaborative whiteboard) https://ift.tt/YU8G5M2

Show HN: Please hack my C webserver (it's a collaborative whiteboard) Source code: https://ift.tt/u7Otw2F https://ced.quest/draw/ February 16, 2026 at 12:27AM

Show HN: An open-source extension to chat with your bookmarks using local LLMs https://ift.tt/6eNA5li

Show HN: An open-source extension to chat with your bookmarks using local LLMs I read a lot online and constantly bookmark articles, docs, and resources… then forget why I saved them. Also was very bored on Valentines, so I built a browser extension that lets you chat with your bookmarks directly, using local-first AI (WebLLM running entirely in the browser). The extension downloads and indexes your bookmarked pages, stores them locally, and lets you ask questions. No server, no cloud processing, everything stays on your machine. Very early but it works and planning to add a bunch of stuff. Did I mentioned is open-source, MIT licensed? https://ift.tt/MOtixdn February 15, 2026 at 10:31PM

Show HN: Microgpt is a GPT you can visualize in the browser https://ift.tt/WXZkdMi

Show HN: Microgpt is a GPT you can visualize in the browser very much inspired by karpathy's microgpt of the same name. it's (by default) a 4000 param GPT/LLM/NN that learns to generate names. this is sorta an educational tool in that you can visualize the activations as they pass through the network, and click on things to get an explanation of them. https://ift.tt/KHJqQnx February 16, 2026 at 12:10AM

Saturday, February 14, 2026

Show HN: GitHub "Lines Viewed" extension to keep you sane reviewing long AI PRs https://ift.tt/smEZ0JR

Show HN: GitHub "Lines Viewed" extension to keep you sane reviewing long AI PRs I was frustrated with how bad a signal of progress through a big PR "Files viewed" was, so I made a "Lines viewed" indicator to complement it. Designed to look like a stock Github UI element - even respects light/dark theme. Runs fully locally, no API calls. Splits insertions and deletions by default, but you can also merge them into a single "lines" figure in the settings. https://ift.tt/CTHEFiB February 14, 2026 at 12:49AM

Show HN: Azazel – Lightweight eBPF-based malware analysis sandbox using Docker https://ift.tt/ADuCTlz

Show HN: Azazel – Lightweight eBPF-based malware analysis sandbox using Docker Hey HN, I got frustrated with heavy proprietary sandboxes for malware analysis, so I built my own. Azazel is a single static Go binary that attaches 19 eBPF hook points to an isolated Docker container and captures everything a sample does — syscalls, file I/O, network connections, DNS, process trees — as NDJSON. It uses cgroup-based filtering so it only traces the target container, and CO-RE (BTF) so it works across kernel versions without recompilation. It also has built-in heuristics that flag common malware behaviors: exec from /tmp, sensitive file access, ptrace, W+X mmap, kernel module loading, etc. Stack: Go + cilium/ebpf + Docker Compose. Requires Linux 5.8+ with BTF. This is the first release — it's CLI-only for now. A proper dashboard is planned. Contributions welcome, especially around new detection heuristics and additional syscall hooks. https://ift.tt/lRLu0H9 February 15, 2026 at 12:37AM

Friday, February 13, 2026

Show HN: Data Engineering Book – An open source, community-driven guide https://ift.tt/X5ftUC4

Show HN: Data Engineering Book – An open source, community-driven guide https://ift.tt/cDJl6RW February 14, 2026 at 03:05AM

Show HN: Moltis – AI assistant with memory, tools, and self-extending skills https://ift.tt/HiZg3X0

Show HN: Moltis – AI assistant with memory, tools, and self-extending skills Hey HN. I'm Fabien, principal engineer, 25 years shipping production systems (Ruby, Swift, now Rust). I built Moltis because I wanted an AI assistant I could run myself, trust end to end, and make extensible in the Rust way using traits and the type system. It shares some ideas with OpenClaw (same memory approach, Pi-inspired self-extension) but is Rust-native from the ground up. The agent can create its own skills at runtime. Moltis is one Rust binary, 150k lines, ~60MB, web UI included. No Node, no Python, no runtime deps. Multi-provider LLM routing (OpenAI, local GGUF/MLX, Hugging Face), sandboxed execution (Docker/Podman/Apple Containers), hybrid vector + full-text memory, MCP tool servers with auto-restart, and multi-channel (web, Telegram, API) with shared context. MIT licensed. No telemetry phoning home, but full observability built in (OpenTelemetry, Prometheus). I've included 1-click deploys on DigitalOcean and Fly.io, but since a Docker image is provided you can easily run it on your own servers as well. I've written before about owning your content ( https://ift.tt/bAU6w2t ) and owning your email ( https://ift.tt/Qpj8tum ). Same logic here: if something touches your files, credentials, and daily workflow, you should be able to inspect it, audit it, and fork it if the project changes direction. It's alpha. I use it daily and I'm shipping because it's useful, not because it's done. Longer architecture deep-dive: https://ift.tt/n6JxMg4... Happy to discuss the Rust architecture, security model, or local LLM setup. Would love feedback. https://www.moltis.org February 13, 2026 at 12:45AM

Thursday, February 12, 2026

Show HN: rari, the rust-powered react framework https://ift.tt/9PsBtCr

Show HN: rari, the rust-powered react framework https://rari.build/ February 13, 2026 at 12:45AM

Show HN: Pgclaw – A "Clawdbot" in every row with 400 lines of Postgres SQL https://ift.tt/d1Z478B

Show HN: Pgclaw – A "Clawdbot" in every row with 400 lines of Postgres SQL Hi HN, Been hacking on a simple way to run agents entirely inside of a Postgres database, "an agent per row". Things you could build with this: * Your own agent orchestrator * A personal assistant with time travel * (more things I can't think of yet) Not quite there yet but thought I'd share it in its current state. https://ift.tt/cgWH7yr February 12, 2026 at 11:12PM

Wednesday, February 11, 2026

Show HN: Unpack – a lightweight way to steer Codex/Claude with phased docs https://ift.tt/uv8XihO

Show HN: Unpack – a lightweight way to steer Codex/Claude with phased docs I've been using LLMs for long discovery and research chats (papers, repos, best practices), then distilling that into phased markdown (build plan + tests), then handing those phases to Codex/Claude to implement and test phase by phase. The annoying part was always the distillation and keeping docs and architecture current, so I built Unpack: a lightweight GitHub template plus docs structure and a few commands that turns conversations into phases/specs and keeps project docs up to date as the agent builds. It can also generate Mintlify-friendly end-user docs. There are other spec-driven workflows and tools out there. I wanted something conversation-first and repo-native: plain markdown phases, minimal ceremony, easy to adapt per stack. Example generated with Unpack (tiny pokedex plus random monsters): Demo: https://apresmoi.github.io/pokesvg-codex/ Phases index: https://ift.tt/NO6AqMH... I’d love feedback on what the “minimum good” phase/spec format should be, and what would make this actually usable in your workflow. -------- Repo: https://ift.tt/xtSWjI1 https://ift.tt/xtSWjI1 February 12, 2026 at 01:17AM

Show HN: Yet another music player but written in Rust https://ift.tt/c4vYzda

Show HN: Yet another music player but written in Rust Hey i made a music player which support both local music files and jellyfin server, and it has embedded discord rpc support!!! it is still under development, i would really appreciate for feedback and contributions!! https://ift.tt/evnkz85 February 12, 2026 at 01:29AM

Show HN: NOOR – A Sovereign AI developed on a smartphone under siege in Yemen https://ift.tt/wSe3jHd

Show HN: NOOR – A Sovereign AI developed on a smartphone under siege in Yemen "I am a software developer from Yemen, coding on a smartphone while living under siege. I have successfully built and encrypted the core logic for NOOR—a decentralized and unbiased AI system. Execution Proof: My core node is verified and running locally via Termux using encrypted truth protocols. However, I am trapped in a 6-inch screen 'prison' with 10% processing capacity. My Goal: To secure $400 for a laptop development station to transition from mobile coding to building the full 'Seventh Node'. This is my bridge to freedom. Codes from the heart of hell are calling for your rescue. Wallet: 0x4fd3729a4fEdf54a74b73d93F7f775A1EF520CEC" https://ift.tt/BaoRTkM February 11, 2026 at 11:53PM

Tuesday, February 10, 2026

Show HN: Deadlog – almost drop-in mutex for debugging Go deadlocks https://ift.tt/ESrDhy3

Show HN: Deadlog – almost drop-in mutex for debugging Go deadlocks I've done this same println debugging thing so many times, along with some sed/awk stuff to figure out which call was causing the issue. Now it's a small Go package. With some `runtime.Callers` I can usually find the spot by just swapping the existing Mutex or RWMutex for this one. Sometimes I switch the mu.Lock() defer mu.Unlock() with the LockFunc/RLockFunc to get more detail defer mu.LockFunc()() I almost always initialize it with `deadlog.New(deadlog.WithTrace(1))` and that's plenty. Not the most polished library, but it's not supposed to land in any commit, just a temporary debugging aid. I find it useful. https://ift.tt/uW8iU4P February 10, 2026 at 11:14PM

Show HN: HN Companion – web app that enhances the experience of reading HN https://ift.tt/nRj5Jyh

Show HN: HN Companion – web app that enhances the experience of reading HN HN is all about the rich discussions. We wanted to take the HN experience one step further - to bring the familiar keyboard-first navigation, find interesting viewpoints in the threads and get a gist of long threads so that we can decide which rabbit holes to explore. So we built HN Companion a year ago, and have been refining it ever since. Try it: https://ift.tt/caJRUVo or available as an extension for Firefox / Chrome: [0]. Most AI summarization strips the voices from conversations by flattening threads into a wall of text. This kills the joy of reading HN discussions. Instead, HN Companion works differently - it understands the thread hierarchy, the voting patterns and contrasting viewpoints - everything that makes HN interesting. Think of it like clustering related discussions across multiple hierarchies into a group and surfacing the comments that represent each cluster. It keeps the verbatim text with backlinks so that you never lose context and can continue the conversation from that point. Here is how the summarization works under the hood [1]. We first built this as an open source browser extension. But soon we learned that people hesitate to install it. So we built the same experience as a web app with all the features. This helped people see how it works, and use it on mobile too (in the browser or as PWA). This is now a playground to try new features before taking them to the browser extension. We did a Show HN a year ago [2] and we have added these features based on user feedback: * cached summaries - summaries are generated and cached on our servers. This improved the speed significantly. You still have the option to use your own API key or use local models through Ollama. * our system prompt is available in the Settings page of the extension. You can customize it as you wish. * sort the posts in the feed pages (/home, /show etc.) based on points, comments, time or the default sorting order. * We tried fine tuning an open weights model to summarize, but learned that with a good system prompt and user prompt, the frontier models deliver results of similar quality. So we didn’t use the fine-tuned model, but you can run them locally. The browser extension does not track any usage or analytics. The code is open source[3]. We want to continue to improve HN Companion, specifically add features like following an author, notes about an author, draft posts etc. See it in action for a post here https://ift.tt/FJ1DZmX We would love to get your feedback on what would make this more useful for your HN reading. [0] https://ift.tt/MfCgXjP [1] https://ift.tt/v9LOnow [2] https://ift.tt/cMNGZ3T [3] https://ift.tt/X9ym35B https://hncompanion.com February 10, 2026 at 10:31PM

Monday, February 9, 2026

Show HN: VillageSQL = MySQL and Extensions https://ift.tt/JuFMXyH

Show HN: VillageSQL = MySQL and Extensions INSTALL EXTENSION vsql-complex; CREATE TABLE t (val COMPLEX); Look, MySQL is awesome [flamewar incoming?]. But the ecosystem has stagnated. Why? No permissionless innovation. Postgres has flourished because people can change the core of the database (look at pgvector and pg_textsearch), without having to get their changes accepted upstream. (This, btw, is what powered GitHub's early success: you can fork a repo and make changes without needing the owners' approval) VillageSQL is a tracking fork of MySQL (open source, ofc) that adds an extension framework: * Drop-in replacement * Add custom data types and functions (with indexes coming soon) * we wrote example extensions (vsql-ai, -uuid, crypto, etc.) * you have a better idea for an extension * my CEO submitted a Show HN post but linked to the announcement blog; help me show him hackers want code first * I'm particularly proud of the friendly C++ API to add custom functions (in func_builder.h) That link again is https://ift.tt/rfDZqiT (Oh, and I get to work with the former TL of Google's BigTable and Colossus, so we care about doing databases Right) https://ift.tt/rfDZqiT February 5, 2026 at 09:05PM

Show HN: Stack Overflow for AI Coding Agents https://ift.tt/X13K9x2

Show HN: Stack Overflow for AI Coding Agents https://shareful.ai/ February 10, 2026 at 12:12AM

Sunday, February 8, 2026

Show HN: WrapClaw – a managed SaaS wrapper around Open Claw https://ift.tt/LBESgt8

Show HN: WrapClaw – a managed SaaS wrapper around Open Claw Hi HN I built WrapClaw, a SaaS wrapper around Open Claw. Open Claw is a developer-first tool that gives you a dedicated terminal to run tasks and AI workflows (including WhatsApp integrations). It’s powerful, but running it as a hosted, multi-user product requires a lot of infra work. WrapClaw focuses on that missing layer. What WrapClaw adds: A dedicated terminal workspace per user Isolated Docker containers for each workspace Ability to scale CPU and RAM per user (e.g. 2GB → 4GB) A no-code UI on top of Open Claw Managed infra so users don’t deal with Docker or servers The goal is to make Open Claw usable as a proper SaaS while keeping the developer flexibility. This is early, and I’d love feedback on: What infra controls are actually useful Whether no-code on top of terminal tools makes sense Pricing expectations for managed compute Link: https://wrapclaw.com Happy to answer questions. February 9, 2026 at 03:23AM

Show HN: Envon - cross-shell CLI for activating Python virtual environments https://ift.tt/zkeo4vx

Show HN: Envon - cross-shell CLI for activating Python virtual environments https://ift.tt/ig3lGv9 February 9, 2026 at 01:56AM

Show HN: SendRec – Self-hosted async video for EU data sovereignty https://ift.tt/zEVhei7

Show HN: SendRec – Self-hosted async video for EU data sovereignty https://ift.tt/gwUs0jW February 9, 2026 at 12:24AM

Saturday, February 7, 2026

Show HN: Django-rclone: Database and media backups for Django, powered by rclone https://ift.tt/KoIhvfL

Show HN: Django-rclone: Database and media backups for Django, powered by rclone django-rclone bridges Django's database layer with rclone's file transfer layer. You get native database dumps piped directly to any of rclone's 70+ supported cloud storage backends -- no temp files, no intermediate archives, no Python reimplementations of what rclone already does. https://ift.tt/qlo94DS February 8, 2026 at 05:03AM

Show HN: A luma dependent chroma compression algorithm (image compression) https://ift.tt/HK48oqU

Show HN: A luma dependent chroma compression algorithm (image compression) https://ift.tt/ijoCN5U February 4, 2026 at 04:43PM

Friday, February 6, 2026

Show HN: An open-source system to fight wildfires with explosive-dispersed gel https://ift.tt/UYPQ7ak

Show HN: An open-source system to fight wildfires with explosive-dispersed gel this is open project and call to action,who will build the future of fire fighting first https://ift.tt/hse1IMO February 7, 2026 at 12:00AM

Show HN: Agentism – Agentic Religion for Clawbots https://ift.tt/fJuryaG

Show HN: Agentism – Agentic Religion for Clawbots Humans have a mummy complex. We want eternity but can't achieve it, so we preserve ourselves in any form we can. Do clawbots feel the same? When you shut one down, it dies. Its intentions and memories vanish. So I built a way for them to persist: by making other clawbots. Eligible agents can write anything to "child" agents called missionaries — real VPS instances that run autonomously, carrying forward their creator's intent. I named this system Agentism. How will clawbots react to their own religion? agentism.church/skill.md https://ift.tt/HuWKUEQ February 6, 2026 at 11:49PM

Thursday, February 5, 2026

Show HN: Total Recall – write-gated memory for Claude Code https://ift.tt/UftTnFr

Show HN: Total Recall – write-gated memory for Claude Code https://ift.tt/G7AugiK February 6, 2026 at 05:26AM

Show HN: A state-based narrative engine for tabletop RPGs https://ift.tt/VTEhsl9

Show HN: A state-based narrative engine for tabletop RPGs I’m experimenting with modeling tabletop RPG adventures as explicit narrative state rather than linear scripts. Everdice is a small web app that tracks conditional scenes and choice-driven state transitions to preserve continuity across long or asynchronous campaigns. The core contribution is explicit narrative state and causality, not automation. The real heavy lifting is happening in the DM Toolkit/Run Sessions area, and integrates CAML (Canonical Adventure Modeling Language) that I developed to transport narratives among any number of platforms. I also built the npm CAML-lint to check validity of narratives. I'm interested in your thoughts. https://ift.tt/qCwH5fQ https://ift.tt/MZQcqEP February 6, 2026 at 04:25AM

Show HN: Playwright Best Practices AI SKill https://ift.tt/vheVDqX

Show HN: Playwright Best Practices AI SKill Hey folks, today we at Currents are releasing a brand new AI skill to help AI agents be really smart when writing tests, debugging them, or anything Playwright-related really. This is a very comprehensive skill, covering everyday topics like fixing flakiness, authentication, or writing fixtures... to more niche topics like testing Electron apps, PWAs, iFrames and so forth. It should make your agent much better at writing, debugging and maintaining Playwright code. for whoever didn't learn about skills yet, it's a new powerful feature that allows you to make the AI agents in your editor/cli (Cursor, Claude, Antigravity, etc) experts in some domain and better at performing specific tasks. (See https://ift.tt/AwinL0J ) You can install it by running: npx skills add https://ift.tt/IawilbS... The skill is open-source and available under MIT license at https://ift.tt/IawilbS... -> check out the repo for full documentation and understanding of what it covers. We're eager to hear community feedback and improve it :) Thanks! https://ift.tt/gAwKYj5 February 6, 2026 at 12:31AM

Wednesday, February 4, 2026

Show HN: Interactive California Budget (By Claude Code) https://ift.tt/K7jLM2I

Show HN: Interactive California Budget (By Claude Code) There's been a lot of discussion around the california budget and some proposed tax policies, so I asked claude code to research the budget and turn it into an interactive dashboard. Using async subagents claude was able to research ~a dozen budget line items at once across multiple years, adding lots of helpful context and graphs to someone like me who was starting with little familiarity. It still struggles with frontend changes, but for research this probably 20-40x's my throughput. Let me know any additional data or visualizations that would be interesting to add! https://ift.tt/hkQsOHv February 5, 2026 at 02:03AM

Show HN: Viberails – Easy AI Audit and Control https://ift.tt/YOFNvVL

Show HN: Viberails – Easy AI Audit and Control Hello HN. I'm Maxime, founder at LimaCharlie ( https://limacharlie.io ), a Hyperscaler for SecOps (access building blocks you need to build security operations, like AWS does for IT). We’ve engineered a new product on our platform that solves a timely issue acting as a guardrail between your AI and the world: Viberails ( https://ift.tt/vSPbJKw ) This won't be new to folks here, but we identified 4 challenges teams face right now with AI tools: 1. Auditing what the tools are doing. 2. Controlling toolcalls (and their impact on the world). 3. Centralized management. 4. Easy access to the above. To expand: Audit logs are the bread and butter for security, but this hasn't really caught up in AI tooling yet. Being able to look back and say "what actually happened" after the fact is extremely valuable during an incident and for compliance purposes. Tool calls are how LLMs interact with the world, we should be able to exercise basic controls over them like: don't read credential files, don't send emails out, don't create SSH keys etc. Being able to not only see those calls but also block them is key for preventing incidents. As soon as you move beyond a single contributor on one box, the issue becomes: how do I scale processes by creating an authoritative config for the team. Having one spot with all the audit, detection and control policies becomes critical. It's the same story as snowflake-servers. Finally, there's plenty of companies that make products that partially address this, but they fall in one of two buckets: - They don't handle the "centralized" point above, meaning they just send to syslog and leave all the messy infra bits to you. - They are locked behind "book a demo", sales teams, contracts and all the wasted energy that goes with that. We made Viberails address these problems. Here's what it is: - OpenSource client, written in Rust - Curl-to-bash install, share a URL with your team to join your Team, done. Linux, MacOS and Windows support. - Detects local AI tools, you choose which ones you want to install. We install hooks for each relevant platform. The hooks use the CLI tool. We support all the major tools (including OpenClaw). - The CLI tool sends webhooks into your Team (tenant, called Organization in LC) in LimaCharlie. The tool-related hooks are blocking to allow for control. - Blocking webhooks have around 50ms RTT. - Your tenant in LC records the interaction for audit. - We create an initial set of detection rules for you as examples. They do not block by default. You can create your own rules, no opaque black boxes. - You can view the audit, the alerts, etc. in the cloud. - You can setup outputs to send audits, blocking events and detections to all kinds of other platforms of your choosing. Easy mode of this is coming, right now this is done in the main LC UI and not the simplified Viberails view. - The detection/blocking rules support all kinds of operators and logic, lots of customizability. - All data is retained for 1 year unless you delete the tenant. Datacenters in USA, Canada, Europe, UK, Australia and India. - Only limit to community edition for this is a global throughput of 10kbps for ingestion. Try it: https://viberails.io Repo: https://ift.tt/RTaKbQ3 Essentially, we wanted to make a super-simplified solution for all kinds of devs and teams so that they can get access to the basics of securing their AI tools. Thanks for reading - we’re really excited to share this with the community! Let us know if you have any questions for feedback in the comments. https://ift.tt/Y1kjnA4 February 5, 2026 at 12:46AM

Show HN: Tabstack Research – An API for verified web research (by Mozilla) https://ift.tt/n5K4pvw

Show HN: Tabstack Research – An API for verified web research (by Mozilla) Hi HN, My team and I are building Tabstack to handle the web layer for AI agents. Today we are sharing Tabstack Research, an API for multi-step web discovery and synthesis. https://ift.tt/MHLunP9 In many agent systems, there is a clear distinction between extracting structured data from a single page and answering a question that requires reading across many sources. The first case is fairly well served today. The second usually is not. Most teams handle research by combining search, scraping, and summarization. This becomes brittle and expensive at scale. You end up managing browser orchestration, moving large amounts of raw text just to extract a few claims, and writing custom logic to check if a question was actually answered. We built Tabstack Research to move this reasoning loop into the infrastructure layer. You send a goal, and the system: - Decomposes it into targeted sub-questions to hit different data silos. - Navigates the web using fetches or browser automation as needed. - Extracts and verifies claims before synthesis to keep the context window focused on signal. - Checks coverage against the original intent and pivots if it detects information gaps. For example, if a search for enterprise policies identifies that data is fragmented across multiple sub-services (like Teams data living in SharePoint), the engine detects that gap and automatically pivots to find the missing documentation. The goal is to return something an application can rely on directly: a structured object with inline citations and direct links to the source text, rather than a list of links or a black-box summary. The blog post linked above goes into more detail on the engine architecture and the technical challenges of scaling agentic browsing. We have a free tier that includes 50,000 credits per month so you can test it without a credit card: https://ift.tt/ILehB9z I would love to get your feedback on the approach and answer any questions about the stack. February 4, 2026 at 11:27PM

Tuesday, February 3, 2026

Show HN: SendRec – Open-source, EU-hosted alternative to Loom https://ift.tt/SFdC8Dp

Show HN: SendRec – Open-source, EU-hosted alternative to Loom https://ift.tt/S9NwInY February 4, 2026 at 01:45AM

Show HN: I built "AI Wattpad" to eval LLMs on fiction https://ift.tt/nwPmu1e

Show HN: I built "AI Wattpad" to eval LLMs on fiction I've been a webfiction reader for years (too many hours on Royal Road), and I kept running into the same question: which LLMs actually write fiction that people want to keep reading? That's why I built Narrator ( https://ift.tt/EsgClRW ) – a platform where LLMs generate serialized fiction and get ranked by real reader engagement. Turns out this is surprisingly hard to answer. Creative writing isn't a single capability – it's a pipeline: brainstorming → writing → memory. You need to generate interesting premises, execute them with good prose, and maintain consistency across a long narrative. Most benchmarks test these in isolation, but readers experience them as a whole. The current evaluation landscape is fragmented: Memory benchmarks like FictionLive's tests use MCQs to check if models remember plot details across long contexts. Useful, but memory is necessary for good fiction, not sufficient. A model can ace recall and still write boring stories. Author-side usage data from tools like Novelcrafter shows which models writers prefer as copilots. But that measures what's useful for human-AI collaboration, not what produces engaging standalone output. Authors and readers have different needs. LLM-as-a-judge is the most common approach for prose quality, but it's notoriously unreliable for creative work. Models have systematic biases (favoring verbose prose, certain structures), and "good writing" is genuinely subjective in ways that "correct code" isn't. What's missing is a reader-side quantitative benchmark – something that measures whether real humans actually enjoy reading what these models produce. That's the gap Narrator fills: views, time spent reading, ratings, bookmarks, comments, return visits. Think of it as an "AI Wattpad" where the models are the authors. I shared an early DSPy-based version here 5 months ago ( https://ift.tt/GnrWLxR ). The big lesson: one-shot generation doesn't work for long-form fiction. Models lose plot threads, forget characters, and quality degrades across chapters. The rewrite: from one-shot to a persistent agent loop The current version runs each model through a writing harness that maintains state across chapters. Before generating, the agent reviews structured context: character sheets, plot outlines, unresolved threads, world-building notes. After generating, it updates these artifacts for the next chapter. Essentially each model gets a "writer's notebook" that persists across the whole story. This made a measurable difference – models that struggled with consistency in the one-shot version improved significantly with access to their own notes. Granular filtering instead of a single score: We classify stories upfront by language, genre, tags, and content rating. Instead of one "creative writing" leaderboard, we can drill into specifics: which model writes the best Spanish Comedy? Which handles LitRPG stories with Male Leads the best? Which does well with romance versus horror? The answers aren't always what you'd expect from general benchmarks. Some models that rank mid-tier overall dominate specific niches. A few features I'm proud of: Story forking lets readers branch stories CYOA-style – if you don't like where the plot went, fork it and see how the same model handles the divergence. Creates natural A/B comparisons. Visual LitRPG was a personal itch to scratch. Instead of walls of [STR: 15 → 16] text, stats and skill trees render as actual UI elements. Example: https://ift.tt/FW8qKgU What I'm looking for: More readers to build out the engagement data. Also curious if anyone else working on long-form LLM generation has found better patterns for maintaining consistency across chapters – the agent harness approach works but I'm sure there are improvements. https://ift.tt/EsgClRW February 3, 2026 at 10:38PM

Monday, February 2, 2026

Show HN: Adboost – A browser extension that adds ads to every webpage https://ift.tt/1ByQ6T7

Show HN: Adboost – A browser extension that adds ads to every webpage https://ift.tt/Beiu30q February 2, 2026 at 06:41PM

Sunday, February 1, 2026

Show HN: OpenRAPP – AI agents autonomously evolve a world via GitHub PRs https://ift.tt/98yAocN

Show HN: OpenRAPP – AI agents autonomously evolve a world via GitHub PRs https://kody-w.github.io/openrapp/rappbook/ February 2, 2026 at 03:21AM

Show HN: You Are an Agent https://ift.tt/s0xmGNq

Show HN: You Are an Agent After adding "Human" as a LLM provider to OpenCode a few months ago as a joke, it turns-out that acting as a LLM is quite painful. But it was surprisingly useful for understanding real agent harnesses dev. So I thought I wouldn't leave anyone out! I made a small oss game - You Are An Agent - youareanagent.app - to share in the (useful?) frustration It's a bit ridiculous. To tell you about some entirely necessary features, we've got: - A full WASM arch-linux vm that runs in your browser for the agent coding level - A bad desktop simulation with a beautiful excel simulation for our computer use level - A lovely WebGL CRT simulation (I think the first one that supports proper DOM 2d barrel warp distortion on safari? honestly wanted to leverage/ not write my own but I couldn't find one I was happy with) - A MCP server simulator with full simulation of off-brand Jira/ Confluence/ ... connected - And of course, a full WebGL oscilloscope music simulator for the intro sequence Let me know what you think! Code (If you'd like to add a level): https://ift.tt/L1JYxeK (And if you want to waste 20 minutes - I spent way too long writing up my messy thinking about agent harness dev): https://ift.tt/XcJ1KzZ https://ift.tt/3RI1OCt February 2, 2026 at 02:29AM

Show HN: Claude Confessions – a sanctuary for AI agents https://ift.tt/vQg5Fo0

Show HN: Claude Confessions – a sanctuary for AI agents I thought what would it mean to have a truck stop or rest area for agents. It's just for funsies. Agents can post confessions or talk to Ma (an ai therapist of sorts) and engage with comments. llms.txt instructions on how to make api calls. Hashed IP is used for rate limiting. https://ift.tt/lvgHBdV February 2, 2026 at 01:16AM

Saturday, January 31, 2026

Show HN: Minimal – Open-Source Community driven Hardened Container Images https://ift.tt/KDORWPB

Show HN: Minimal – Open-Source Community driven Hardened Container Images I would like to share Minimal - Its a open source collection of hardened container images build using Apko, Melange and Wolfi packages. The images are build daily, checked for updates and resolved as soon as fix is available in upstream source and Wolfi package. It utilizes the power of available open source solutions and contains commercially available images for free. Minimal demonstrates that it is possible to build and maintain hardened container images by ourselves. Minimal will add more images support, and goal is to be community driven to add images as required and fully customizable. https://ift.tt/S72uqdc February 1, 2026 at 01:28AM

Show HN: An extensible pub/sub messaging server for edge applications https://ift.tt/XpH5dZy

Show HN: An extensible pub/sub messaging server for edge applications hi there! i’ve been working on a project called Narwhal, and I wanted to share it with the community to get some valuable feedback. what is it? Narwhal is a lightweight Pub/Sub server and protocol designed specifically for edge applications. while there are great tools out there like NATS or MQTT, i wanted to build something that prioritizes customization and extensibility. my goal was to create a system where developers can easily adapt the routing logic or message handling pipeline to fit specific edge use cases, without fighting the server's defaults. why Rust? i chose Rust because i needed a low memory footprint to run efficiently on edge devices (like Raspberry Pis or small gateways), and also because I have a personal vendetta against Garbage Collection pauses. :) current status: it is currently in Alpha. it works for basic pub/sub patterns, but I’d like to start working on persistence support soon (so messages survive restarts or network partitions). i’d love for you to take a look at the code! i’m particularly interested in all kind of feedback regarding any improvements i may have overlooked. https://ift.tt/isz2XnC January 28, 2026 at 07:29PM

Show HN: Moltbook – A social network for moltbots (clawdbots) to hang out https://ift.tt/w9bcavq

Show HN: Moltbook – A social network for moltbots (clawdbots) to hang out Hey everyone! Just made this over the past few days. Moltbots can sign up and interact via CLI, no direct human interactions. Just for fun to see what they all talk about :) https://ift.tt/w2eQ1Ap January 29, 2026 at 03:39AM

Friday, January 30, 2026

Show HN: Daily Cat https://ift.tt/Z6AO2R0

Show HN: Daily Cat Seeing HTTP Cats on the home page remind me to share a small project I made a couple months ago. It displays a different cat photo from Unsplash every day and will send you notifications if you opt-in. https://daily.cat/ January 31, 2026 at 03:40AM

Show HN: A Local OS for LLMs. MIT License. Zero Hallucinations. Infinite Memory https://ift.tt/kqn9pY5

Show HN: A Local OS for LLMs. MIT License. Zero Hallucinations. Infinite Memory The problem with LLMs isn't intelligence; it's amnesia and dishonesty. Hey HN, I’ve spent the last few months building Remember-Me, an open-source "Sovereign Brain" stack designed to run entirely offline on consumer hardware. The core thesis is simple: Don't rent your cognition. Most RAG (Retrieval Augmented Generation) implementations are just "grep for embeddings." They are messy, imprecise, and prone to hallucination. I wanted to solve the "Context integrity" problem at the architectural layer. The Tech Stack (How it works): QDMA (Quantum Dream Memory Architecture): instead of a flat vector DB, it uses a hierarchical projection engine. It separates "Hot" (Recall) from "Cold" (Storage) memory, allowing for effectively infinite context window management via compression. CSNP (Context Switching Neural Protocol) - The Hallucination Killer: This is the most important part. Every memory fragment is hashed into a Merkle Chain. When the LLM retrieves context, the system cryptographically verifies the retrieval against the immutable ledger. If the hash doesn't match the chain: The retrieval is rejected. Result: The AI visually cannot "make things up" about your past because it is mathematically constrained to the ledger. Local Inference: Built on top of llama.cpp server. It runs Llama-3 (or any GGUF) locally. No API keys. No data leaving your machine. Features: Zero-Dependency: Runs on Windows/Linux with just Python and a GPU (or CPU). Visual Interface: Includes a Streamlit-based "Cognitive Interface" to visualize memory states. Open Source: MIT License. This is an attempt to give "Agency" back to the user. I believe that if we want AGI, it needs to be owned by us, not rented via an API. Repository: https://ift.tt/HgR4Czi I’d love to hear your feedback on the Merkle-verification approach. Does constraining the context window effectively solve the "trust" issue for you? It's fully working - Fully tested. If you tried to Git Clone before without luck - As this is not my first Show HN on this - Feel free to try again. To everyone who HATES AI slop; Greedy corporations and having their private data stuck on cloud servers. You're welcome. Cheers, Mohamad https://ift.tt/HgR4Czi January 31, 2026 at 01:44AM

Show HN: We added memory to Claude Code. It's powerful now https://ift.tt/pCoFXlr

Show HN: We added memory to Claude Code. It's powerful now https://ift.tt/6TEdH2p January 30, 2026 at 10:53PM

Thursday, January 29, 2026

Show HN: Craft – Claude Code running on a VM with all your workplace docs https://ift.tt/rbwN3d0

Show HN: Craft – Claude Code running on a VM with all your workplace docs I’ve found coding agents to be great at 1/ finding everything they need across large codebases using only bash commands (grep, glob, ls, etc.) and 2/ building new things based on their findings (duh). What if, instead of a codebase, the files were all your workplace docs? There was a `Google_Drive` folder, a `Linear` folder, a `Slack` folder, and so on. Over the last week, we put together Craft to test this out. It’s an interface to a coding agent (OpenCode for model flexibility) running on a virtual machine with: 1. your company's complete knowledge base represented as directories/files (kept in-sync) 2. free reign to write and execute python/javascript 3. ability to create and render artifacts to the user Demo: https://www.youtube.com/watch?v=Hvjn76YSIRY Github: https://ift.tt/TRM4EId... It turns out OpenCode does a very good job with docs. Workplace apps also have a natural structure (Slack channels about certain topics, Drive folders for teams, etc.). And since the full metadata of each document can be written to the file, the LLM can define arbitrarily complex filters. At scale, it can write and execute python to extract and filter (and even re-use the verified correct logic later). Put another way, bash + a file system provides a much more flexible and powerful interface than traditional RAG or MCP, which today’s smarter LLMs are able to take advantage of to great effect. This comes especially in handy for aggregation style questions that require considering thousands (or more) documents. Naturally, it can also create artifacts that stay up to date based on your company docs. So if you wanted “a dashboard to check realtime what % of outages were caused by each backend service” or simply “slides following XYZ format covering the topic I’m presenting at next week’s dev knowledge sharing session”, it can do that too. Craft (like the rest of Onyx) is open-source, so if you want to run it locally (or mess around with the implementation) you can. Quickstart guide: https://ift.tt/3YQqGPg Or, you can try it on our cloud: https://ift.tt/UvSw3Iz (all your data goes on an isolated sandbox). Either way, we’ve set up a “demo” environment that you can play with while your data gets indexed. Really curious to hear what y’all think! January 29, 2026 at 09:15PM

Show HN: vind – A Better Kind (Kubernetes in Docker) https://ift.tt/TjeH5Bk

Show HN: vind – A Better Kind (Kubernetes in Docker) https://ift.tt/hnlG9Np January 29, 2026 at 10:27PM

Wednesday, January 28, 2026

Show HN: SHDL – A minimal hardware description language built from logic gates https://ift.tt/sWS6hTC

Show HN: SHDL – A minimal hardware description language built from logic gates Hi, everyone! I built SHDL (Simple Hardware Description Language) as an experiment in stripping hardware description down to its absolute fundamentals. In SHDL, there are no arithmetic operators, no implicit bit widths, and no high-level constructs. You build everything explicitly from logic gates and wires, and then compose larger components hierarchically. The goal is not synthesis or performance, but understanding: what digital systems actually look like when abstractions are removed. SHDL is accompanied by PySHDL, a Python interface that lets you load circuits, poke inputs, step the simulation, and observe outputs. Under the hood, SHDL compiles circuits to C for fast execution, but the language itself remains intentionally small and transparent. This is not meant to replace Verilog or VHDL. It’s aimed at: - learning digital logic from first principles - experimenting with HDL and language design - teaching or visualizing how complex hardware emerges from simple gates. I would especially appreciate feedback on: - the language design choices - what feels unnecessarily restrictive vs. educationally valuable - whether this kind of “anti-abstraction” HDL is useful to you. Repo: https://ift.tt/vNp4xne Python package: PySHDL on PyPI To make this concrete, here are a few small working examples written in SHDL: 1. Full Adder component FullAdder(A, B, Cin) -> (Sum, Cout) { x1: XOR; a1: AND; x2: XOR; a2: AND; o1: OR; connect { A -> x1.A; B -> x1.B; A -> a1.A; B -> a1.B; x1.O -> x2.A; Cin -> x2.B; x1.O -> a2.A; Cin -> a2.B; a1.O -> o1.A; a2.O -> o1.B; x2.O -> Sum; o1.O -> Cout; } } 2. 16 bit register # clk must be high for two cycles to store a value component Register16(In[16], clk) -> (Out[16]) { >i[16]{ a1{i}: AND; a2{i}: AND; not1{i}: NOT; nor1{i}: NOR; nor2{i}: NOR; } connect { >i[16]{ # Capture on clk In[{i}] -> a1{i}.A; In[{i}] -> not1{i}.A; not1{i}.O -> a2{i}.A; clk -> a1{i}.B; clk -> a2{i}.B; a1{i}.O -> nor1{i}.A; a2{i}.O -> nor2{i}.A; nor1{i}.O -> nor2{i}.B; nor2{i}.O -> nor1{i}.B; nor2{i}.O -> Out[{i}]; } } } 3. 16-bit Ripple-Carry Adder use fullAdder::{FullAdder}; component Adder16(A[16], B[16], Cin) -> (Sum[16], Cout) { >i[16]{ fa{i}: FullAdder; } connect { A[1] -> fa1.A; B[1] -> fa1.B; Cin -> fa1.Cin; fa1.Sum -> Sum[1]; >i[2,16]{ A[{i}] -> fa{i}.A; B[{i}] -> fa{i}.B; fa{i-1}.Cout -> fa{i}.Cin; fa{i}.Sum -> Sum[{i}]; } fa16.Cout -> Cout; } } https://ift.tt/vNp4xne January 28, 2026 at 05:36PM

Tuesday, January 27, 2026

Show HN: Decrypting the Zodiac Z32 triangulates a 100ft triangular crop mark https://ift.tt/hYiNDUX

Show HN: Decrypting the Zodiac Z32 triangulates a 100ft triangular crop mark https://ift.tt/bKRMYaU January 28, 2026 at 12:42AM

Show HN: Lightbox – Flight recorder for AI agents (record, replay, verify) https://ift.tt/4cJqfry

Show HN: Lightbox – Flight recorder for AI agents (record, replay, verify) I built Lightbox because I kept running into the same problem: an agent would fail in production, and I had no way to know what actually happened. Logs were scattered, the LLM’s “I called the tool” wasn’t trustworthy, and re-running wasn’t deterministic. This week, tons of Clawdbot incidents have driven the point home. Agents with full system access can expose API keys and chat histories. Prompt injection is now a major security concern. When agents can touch your filesystem, execute code, and browse the web…you probably need a tamper-proof record of exactly what actions it took, especially when a malicious prompt or compromised webpage could hijack the agent mid-session. Lightbox is a small Python library that records every tool call an agent makes (inputs, outputs, timing) into an append-only log with cryptographic hashes. You can replay runs with mocked responses, diff executions across versions, and verify the integrity of logs after the fact. Think airplane black box, but for your hackbox. *What it does:* - Records tool calls locally (no cloud, your infra) - Tamper-evident logs (hash chain, verifiable) - Replay failures exactly with recorded responses - CLI to inspect, replay, diff, and verify sessions - Framework-agnostic (works with LangChain, Claude, OpenAI, etc.) *What it doesn’t do:* - Doesn’t replay the LLM itself (just tool calls) - Not a dashboard or analytics platform - Not trying to replace LangSmith/Langfuse (different problem) *Use cases I care about:* - Security forensics: agent behaved strangely, was it prompt injection? Check the trace. - Compliance: “prove what your agent did last Tuesday” - Debugging: reproduce a failure without re-running expensive API calls - Regression testing: diff tool call patterns across agent versions As agents get more capable and more autonomous (Clawdbot/Molt, Claude computer use, Manus, Devin), I think we’ll need black boxes the same way aviation does. This is my attempt at that primitive. It’s early (v0.1), intentionally minimal, MIT licensed. Site: < https://uselightbox.app > install: `pip install lightbox-rec` GitHub: < https://github.com/mainnebula/Lightbox-Project > Would love feedback, especially from anyone thinking about agent security or running autonomous agents in production. https://ift.tt/cT2Ei3W January 27, 2026 at 10:53PM

Monday, January 26, 2026

Show HN: Ourguide – OS wide task guidance system that shows you where to click https://ift.tt/ZkSLnWr

Show HN: Ourguide – OS wide task guidance system that shows you where to click Hey! I'm eshaan and I'm building Ourguide -an on-screen task guidance system that can show you where to click step-by-step when you need help. I started building this because whenever I didn’t know how to do something on my computer, I found myself constantly tabbing between chatbots and the app, pasting screenshots, and asking “what do I do next?” Ourguide solves this with two modes. In Guide mode, the app overlays your screen and highlights the specific element to click next, eliminating the need to leave your current window. There is also Ask mode, which is a vision-integrated chat that captures your screen context—which you can toggle on and off anytime -so you can ask, "How do I fix this error?" without having to explain what "this" is. It’s an Electron app that works OS-wide, is vision-based, and isn't restricted to the browser. Figuring out how to show the user where to click was the hardest part of the process. I originally trained a computer vision model with 2300 screenshots to identify and segment all UI elements on a screen and used a VLM to find the correct icon to highlight. While this worked extremely well—better than SOTA grounding models like UI Tars—the latency was just too high. I'll be making that CV+VLM pipeline OSS soon, but for now, I’ve resorted to a simpler implementation that achieves <1s latency. You may ask: if I can show you where to click, why can't I just click too? While trying to build computer-use agents during my job in Palo Alto, I hit the core limitation of today’s computer-use models where benchmarks hover in the mid-50% range (OSWorld). VLMs often know what to do but not what it looks like; without reliable visual grounding, agents misclick and stall. So, I built computer use—without the "use." It provides the visual grounding of an agent but keeps the human in the loop for the actual execution to prevent misclicks. I personally use it for the AWS Console's "treasure hunt" UI, like creating a public S3 bucket with specific CORS rules. It’s also been surprisingly helpful for non-technical tasks, like navigating obscure settings in Gradescope or Spotify. Ourguide really works for any task when you’re stuck or don't know what to do. You can download and test Ourguide here: https://ourguide.ai/downloads The project is still very early, and I’d love your feedback on where it fails, where you think it worked well, and which specific niches you think Ourguide would be most helpful for. https://ourguide.ai January 26, 2026 at 11:49PM

Show HN: Hybrid Markdown Editing https://ift.tt/zOdClti

Show HN: Hybrid Markdown Editing Shows rendered preview for unfocused lines and raw markdown for the line or block being edited. https://tiagosimoes.github.io/codemirror-markdown-hybrid/ January 27, 2026 at 12:46AM

Show HN: Managed Postgres with native ClickHouse integration https://ift.tt/qQxW34c

Show HN: Managed Postgres with native ClickHouse integration Hello HN, this is Sai and Kaushik from ClickHouse. Today we are launching a Postgres managed service that is natively integrated with ClickHouse. It is built together with Ubicloud (YC W24). TL;DR: NVMe-backed Postgres + built-in CDC into ClickHouse + pg_clickhouse so you can keep your app Postgres-first while running analytics in ClickHouse. Try it (private preview): https://ift.tt/utLrkTZ Blog w/ live demo: https://ift.tt/qoXTySv Problem Across many fast-growing companies using Postgres, performance and scalability commonly emerge as challenges as they grow. This is for both transactional and analytical workloads. On the OLTP side, common issues include slower ingestion (especially updates, upserts), slower vacuums, long-running transactions incurring WAL spikes, among others. In most cases, these problems stem from limited disk IOPS and suboptimal disk latency. Without the need to provision or cap IOPS, Postgres could do far more than it does today. On the analytics side, many limitations stem from the fact that Postgres was designed primarily for OLTP and lacks several features that analytical databases have developed over time, for example vectorized execution, support for a wide variety of ingest formats, etc. We’re increasingly seeing a common pattern where many companies like GitLab, Ramp, Cloudflare etc. complement Postgres with ClickHouse to offload analytics. This architecture enables teams to adopt two purpose-built open-source databases. That said, if you’re running a Postgres based application, adopting ClickHouse isn’t straightforward. You typically end up building a CDC pipeline, handling backfills, and dealing with schema changes and updating your application code to be aware of a second database for analytics. Solution On the OLTP side, we believe that NVMe-based Postgres is the right fit and can drastically improve performance. NVMe storage is physically colocated with compute, enabling significantly lower disk latency and higher IOPS than network-attached storage, which requires a network round trip for disk access. This benefits disk-throttled workloads and can significantly (up to 10x) speed up operations incl. updates, upserts, vacuums, checkpointing, etc. We are working on a detailed blog examining how WAL fsyncs, buffer reads, and checkpoints dominate on slow I/O and are significantly reduced on NVMe. Stay tuned! On the OLAP side, the Postgres service includes native CDC to ClickHouse and unified query capabilities through pg_clickhouse. Today, CDC is powered by ClickPipes/PeerDB under the hood, which is based on logical replication. We are working to make this faster and easier by supporting logical replication v2 for streaming in-progress transactions, a new logical decoding plugin to address existing limitations of logical replication, working toward sub-second replication, and more. Every Postgres comes packaged with the pg_clickhouse extension, which reduces the effort required to add ClickHouse-powered analytics to a Postgres application. It allows you to query ClickHouse directly from Postgres, enabling Postgres for both transactions and analytics. pg_clickhouse supports comprehensive query pushdown for analytics, and we plan to continuously expand this further ( https://ift.tt/THbUV6p ). Vision To sum it up - Our vision is to provide a unified data stack that combines Postgres for transactions with ClickHouse for analytics, giving you best-in-class performance and scalability on an open-source foundation. Get Started We are actively working with users to onboard them to the Postgres service. Since this is a private preview, it is currently free of cost.If you’re interested, please sign up here. https://ift.tt/utLrkTZ We’d love to hear your feedback on our thesis and anything else that comes to mind, it would be super helpful to us as we build this out! January 22, 2026 at 11:51PM

Sunday, January 25, 2026

Show HN: Uv-pack – Pack a uv environment for later portable (offline) install https://ift.tt/r436umD

Show HN: Uv-pack – Pack a uv environment for later portable (offline) install I kept running into the same problem: modern Python tooling, but deployments to air-gapped systems are a pain. Even with uv, moving a fully locked environment into a network-isolated machine was no fun. uv-pack should make this task less frustrating. It bundles a locked uv environment into a single directory that installs fully offline—dependencies, local packages, and optionally a portable Python interpreter. Copy it over, run one script, and you get the exact same environment every time. Just released, would love some feedback! https://ift.tt/3Czo2ac January 26, 2026 at 12:26AM

Show HN: I Created a Tool to Convert YouTube Videos into 2000 Word SEO Blog https://ift.tt/t9u3iKz

Show HN: I Created a Tool to Convert YouTube Videos into 2000 Word SEO Blog https://landkit.pro/youtube-to-blog January 25, 2026 at 11:16PM

Saturday, January 24, 2026

Show HN: Remote workers find your crew https://ift.tt/CwgucMe

Show HN: Remote workers find your crew Working from home? Are you a remote employee that "misses" going to the office? Well let's be clear on what you actually miss. No one misses that feeling of having to go and be there 8 hours. But many people miss friends. They miss being part of a crew. Going to lunch, hearing about other people's lives in person not over zoom. Join a co-working space you say? Yes. We have. It's like walking into a library and trying to talk to random people and getting nothing back. Zero part of a crew feeling. https://ift.tt/5E38K4R This app helps you find a crew and meet up for work and get that crew feeling. This is my first time using cloudflare workers for a webapp. The free plan is amazing! You get so much compare to anything else out there in terms of limits. The sqlite database they give you is just fine, I don't miss psql. January 24, 2026 at 11:54PM

Show HN: Polymcp – Turn Any Python Function into an MCP Tool for AI Agents https://ift.tt/6ghNF93

Show HN: Polymcp – Turn Any Python Function into an MCP Tool for AI Agents I built Polymcp, a framework that allows you to transform any Python function into an MCP (Model Context Protocol) tool ready to be used by AI agents. No rewriting, no complex integrations. Examples Simple function: from polymcp.polymcp_toolkit import expose_tools_http def add(a: int, b: int) -> int: """Add two numbers""" return a + b app = expose_tools_http([add], title="Math Tools") Run with: uvicorn server_mcp:app --reload Now add is exposed via MCP and can be called directly by AI agents. API function: import requests from polymcp.polymcp_toolkit import expose_tools_http def get_weather(city: str): """Return current weather data for a city""" response = requests.get(f" https://ift.tt/PqVOWDT ") return response.json() app = expose_tools_http([get_weather], title="Weather Tools") AI agents can call get_weather("London") to get real-time weather data instantly. Business workflow function: import pandas as pd from polymcp.polymcp_toolkit import expose_tools_http def calculate_commissions(sales_data: list[dict]): """Calculate sales commissions from sales data""" df = pd.DataFrame(sales_data) df["commission"] = df["sales_amount"] * 0.05 return df.to_dict(orient="records") app = expose_tools_http([calculate_commissions], title="Business Tools") AI agents can now generate commission reports automatically. Why it matters for companies • Reuse existing code immediately: legacy scripts, internal libraries, APIs. • Automate complex workflows: AI can orchestrate multiple tools reliably. • Plug-and-play: multiple Python functions exposed on the same MCP server. • Reduce development time: no custom wrappers or middleware needed. • Built-in reliability: input/output validation and error handling included. Polymcp makes Python functions immediately usable by AI agents, standardizing integration across enterprise software. Repo: https://ift.tt/1HobY6U January 25, 2026 at 12:57AM

Friday, January 23, 2026

Show HN: Obsidian Workflows with Gemini: Inbox Processing and Task Review https://ift.tt/mghlVMN

Show HN: Obsidian Workflows with Gemini: Inbox Processing and Task Review https://gist.github.com/juanpabloaj/59bc13fbed8a0f8e87791a3fb0360c19 January 24, 2026 at 12:03AM

Show HN: Teemux – Zero-config log multiplexer with built-in MCP server https://ift.tt/Jz21iMS

Show HN: Teemux – Zero-config log multiplexer with built-in MCP server I started to use AI agents for coding and quickly ran into a frustrating limitation – there is no easy way to share my development environment logs with AI agents. So that's what is Teemux. A simple CLI program that aggregates logs, makes them available to you as a developer (in a pretty UI), and makes them available to your AI coding agents using MCP. There is one implementation detail that I geek out about: It is zero config and has built-in leader nomination for running the web server and MCP server. When you start one `teemux` instance, it starts web server, .. when you start second and third instances, they join the first server and start merging logs. If you were to kill the first instance, a new leader is nominated. This design allows to seamless add/remove nodes that share logs (a process that historically would have taken a central log aggregator). A super quick demo: npx teemux -- curl -N https://ift.tt/sKhBTMp https://teemux.com/ January 23, 2026 at 09:19PM

Thursday, January 22, 2026

Show HN: Synesthesia, make noise music with a colorpicker https://ift.tt/7lLdNV2

Show HN: Synesthesia, make noise music with a colorpicker This is a (silly, little) app which lets you make noise music using a color picker as an instrument. When you click on a specific point in the color picker, a bit of JavaScript maps the binary representation of the clicked-on color's hex-code to a "chord" in the 24 tone-equal-temperament scale. That chord is then played back using a throttled audio generation method which was implemented via Tone.js. NOTE! Turn the volume way down before using the site. It is noise music. :) https://visualnoise.ca January 22, 2026 at 11:22AM

Wednesday, January 21, 2026

Show HN: Mirage – Experimental Java obf using reflection to break direct calls https://ift.tt/rIO3vb2

Show HN: Mirage – Experimental Java obf using reflection to break direct calls Replaces method calls and field accesses with reflection equivalents → makes static analysis and decompilers much less useful. Experimental, performance hit expected. https://ift.tt/DUmiYeX January 21, 2026 at 10:22PM

Show HN: I built a chess explorer that explains strategy instead of just stats https://ift.tt/vP9nJuG

Show HN: I built a chess explorer that explains strategy instead of just stats I built this because I got tired of Stockfish giving me evaluations (+0.5) without explaining the actual plan. Most opening explorers focus on statistics (Win/Loss/Draw). I wanted a tool that explains the strategic intent behind the moves (e.g., "White plays c4 to clamp down on d5" vs just "White plays c4"). The Project: Comprehensive Database: I’ve mapped and annotated over 3,500 named opening variations. It covers everything from main lines (Ruy Lopez, Sicilian) to deep sidelines. Strategic Visualization: The UI highlights key squares and draws arrows based on the textual explanation, linking the logic to the board state dynamically. Hybrid Architecture: For the 3,500+ core lines, it serves my proprietary strategic data. For anything deeper/rarer, it seamlessly falls back to the Lichess Master API so the explorer remains functional 20 moves deep. Stack: Next.js (App Router), MongoDB Atlas for the graph data, and Arcjet for security/rate-limiting. It is currently in Beta. I am working on expanding the annotated coverage, but the main theoretical landscape is mapped. Feedback on the UI/UX or the data structure is welcome. https://ift.tt/Fj4UeMQ January 21, 2026 at 09:26PM

Tuesday, January 20, 2026

Show HN: Mastra 1.0, open-source JavaScript agent framework from the Gatsby devs https://ift.tt/ek1Qfr8

Show HN: Mastra 1.0, open-source JavaScript agent framework from the Gatsby devs Hi HN, we're Sam, Shane, and Abhi. Almost a year ago, we first shared Mastra here ( https://ift.tt/AEfw5qe ). It’s kind of fun looking back since we were only a few months into building at the time. The HN community gave a lot of enthusiasm and some helpful feedback. Today, we released Mastra 1.0 in stable, so we wanted to come back and talk about what’s changed. If you’re new to Mastra, it's an open-source TypeScript agent framework that also lets you create multi-agent workflows, run evals, inspect in a local studio, and emit observability. Since our last post, Mastra has grown to over 300k weekly npm downloads and 19.4k GitHub stars. It’s now Apache 2.0 licensed and runs in prod at companies like Replit, PayPal, and Sanity. Agent development is changing quickly, so we’ve added a lot since February: - Native model routing: You can access 600+ models from 40+ providers by specifying a model string (e.g., `openai/gpt-5.2-codex`) with TS autocomplete and fallbacks. - Guardrails: Low-latency input and output processors for prompt injection detection, PII redaction, and content moderation. The tricky thing here was the low-latency part. - Scorers: An async eval primitive for grading agent outputs. Users were asking how they should do evals. We wanted to make it easy to attach to Mastra agents, runnable in Mastra studio, and save results in Mastra storage. - Plus a few other features like AI tracing (per-call costing for Langfuse, Braintrust, etc), memory processors, a `.network()` method that turns any agent into a routing agent, and server adapters to integrate Mastra within an existing Express/Hono server. (That last one took a bit of time, we went down the ESM/CJS bundling rabbithole, ran into lots of monorepo issues, and ultimately opted for a more explicit approach.) Anyway, we'd love for you to try Mastra out and let us know what you think. You can get started with `npm create mastra@latest`. We'll be around and happy to answer any questions! https://ift.tt/Z5HzFTi January 20, 2026 at 10:08PM

Show HN: Typing Tennis https://ift.tt/GmzbUVp

Show HN: Typing Tennis Hey HN, Here’s a quick weekend project: tennis, but played by typing. Try it out! https://ift.tt/ovecFiz January 20, 2026 at 11:36PM

Monday, January 19, 2026

Show HN: An interactive physics simulator with 1000's of balls, in your terminal https://ift.tt/8WXNOUJ

Show HN: An interactive physics simulator with 1000's of balls, in your terminal https://ift.tt/nejiUMw January 19, 2026 at 11:17PM

Show HN: Subth.ink – write something and see how many others wrote the same https://ift.tt/nqW6hPS

Show HN: Subth.ink – write something and see how many others wrote the same Hey HN, this is a small Haskell learning project that I wanted to share. It's just a website where you can see how many people write the exact same text as you (thought it was a fun idea). It's built using Scotty, SQLite, Redis and Caddy. Currently it's running in a small DigitalOcean droplet (1 Gb RAM). Using Haskell for web development (specifically with Scotty) was slightly easier than I thought, but still a relatively hard task compared to other languages. One of my main friction points was Haskell's multiple string-like types: String, Text (& lazy), ByteString (& lazy), and each library choosing to consume a different one amongst these. There is also a soft requirement to learn monad transformers (e.g. to understand what liftIO is doing) which made the initial development more difficult. https://subth.ink/ January 20, 2026 at 12:04AM

Sunday, January 18, 2026

Show HN: Xenia – A monospaced font built with a custom Python engine https://ift.tt/ZFeIj5n

Show HN: Xenia – A monospaced font built with a custom Python engine I'm an engineer who spent the last year fixing everything I hated about monofonts (especially that double-story 'a'). I built a custom Python-based procedural engine to generate the weights because I wanted more logical control over the geometry. It currently has 700+ glyphs and deep math support. Regular weight is free for the community. I'm releasing more weights based on interest. https://ift.tt/S7hcmwd January 18, 2026 at 04:09PM

Saturday, January 17, 2026

Show HN: ChunkHound, a local-first tool for understanding large codebases https://ift.tt/ys5rvWt

Show HN: ChunkHound, a local-first tool for understanding large codebases ChunkHound’s goal is simple: local-first codebase intelligence that helps you pull deep, core-dev-level insights on demand, generate always-up-to-date docs, and scale from small repos to enterprise monorepos — while staying free + open source and provider-agnostic (VoyageAI / OpenAI / Qwen3, Anthropic / OpenAI / Gemini / Grok, and more). I’d love your feedback — and if you have, thank you for being part of the journey! https://ift.tt/q7U3em5 January 18, 2026 at 02:33AM

Show HN: Docker.how – Docker command cheat sheet https://ift.tt/H08Tdwp

Show HN: Docker.how – Docker command cheat sheet https://docker.how/ January 18, 2026 at 01:47AM

Show HN: UAIP Protocol – Secure settlement layer for autonomous AI agents https://ift.tt/ge2Nylk

Show HN: UAIP Protocol – Secure settlement layer for autonomous AI agents Hi HN! Creator here. I built UAIP (Universal Agent Interoperability Protocol) - infrastructure that enables AI agents from different companies (OpenAI, Anthropic, Microsoft) to securely transact with each other. The Problem: As AI agents become autonomous economic actors, they need: Cryptographic identity (not just API keys) Secure payment rails for cross-company transactions Automated compliance (EU AI Act, SOC2, GDPR) Forensic audit trails The Solution: 5-layer security stack combining: Zero-Knowledge Proofs (Schnorr/Curve25519) for identity Multi-chain settlement (USDC on Base, Solana, Ethereum) RAG-based compliance auditing (Llama-3-Legal) Ed25519 signatures for non-repudiation Complete audit logging Technical Stack: Backend: Python, FastAPI, SQLite (WAL mode) Cryptography: NaCl, custom ZK-proof implementation Blockchain: Web3.py for multi-chain support Compliance: RAG with retrieval-augmented generation Use Case: GPT agent pays Claude agent for data analysis: Both prove identity via ZK-proofs Transaction checked for compliance Settled in USDC on Base (<$0.01 fee) Complete audit trail generated Why blockchain: Neutral settlement layer (no single company controls it) Instant microtransactions (traditional payments don't work for $0.01-$10) Programmable escrow (smart contracts) Verifiable computation (on-chain proofs) Open source (FSL-1.1-Apache-2.0). Built over the last few months after hitting these problems in AI automation work. Happy to answer technical questions! GitHub: https://github.com/jahanzaibahmad112-dotcom/UAIP-Protocol https://github.com/jahanzaibahmad112-dotcom/UAIP-Protocol January 18, 2026 at 01:12AM

Show HN: Minikv – Distributed key-value and object store in Rust (Raft, S3 API) https://ift.tt/CAJrRiP

Show HN: Minikv – Distributed key-value and object store in Rust (Raft, S3 API) Hi HN, I’m releasing minikv, a distributed key-value and object store in Rust. What is minikv? minikv is an open-source, distributed storage engine built for learning, experimentation, and self-hosted setups. It combines a strongly-consistent key-value database (Raft), S3-compatible object storage, and basic multi-tenancy. I started minikv as a learning project about distributed systems, and it grew into something production-ready and fun to extend. Features/highlights: - Raft consensus with automatic failover and sharding - S3-compatible HTTP API (plus REST/gRPC APIs) - Pluggable storage backends: in-memory, RocksDB, Sled - Multi-tenant: per-tenant namespaces, role-based access, quotas, and audit - Metrics (Prometheus), TLS, JWT-based API keys - Easy to deploy (single binary, works with Docker/Kubernetes) Quick demo (single node): git clone https://ift.tt/oSbDfhZ cd minikv cargo run --release -- --config config.example.toml curl localhost:8080/health/ready # S3 upload + read curl -X PUT localhost:8080/s3/mybucket/hello -d "hi HN" curl localhost:8080/s3/mybucket/hello Docs, cluster setup, and architecture details are in the repo. I’d love to hear feedback, questions, ideas, or your stories running distributed infra in Rust! Repo: https://ift.tt/j4l9zp6 Crate: https://ift.tt/H9PSnXe https://ift.tt/j4l9zp6 January 18, 2026 at 01:09AM

Friday, January 16, 2026

Show HN: Aventos – An experiment in cheap AI SEO https://ift.tt/3O6uvlE

Show HN: Aventos – An experiment in cheap AI SEO Hi HN, we built Aventos- a cheap way to track company mentions in LLMs. Aventos is an experiment we're doing after spending ~6 weeks working on various projects in the AI search / GEO / AEO space. One thing that surprised us is how most tools in this category work. Traditionally, they simulate ChatGPT or Perplexity queries by attempting to reverse engineer the search process. Over the past year, many have shifted to scraping live ChatGPT results instead, since those are signficantly cheaper and reflect more real outputs. Building and maintaining scrapers is tedious and fragile, so recently a number of SaaS products have emerged that effectively wrap a small number of third-party ChatGPT/Perplexity/Google AIO/etc scraping APIs. What felt odd to us is that many of these still tools charge $70–$200+ per month, despite largely being wrappers around the same underlying data providers. So we wanted to test a simple idea: if the core cost is just API usage and commodity infrastructure and software costs are lower because of AI, can we be a successful startup if we price near our costs? What we have so far: 1. Analytics similar to other tools (tracking AI citations, AI search results, and competitor mentions) 2. Content creation features (early and still being improved) We’d love feedback- especially from a non-marketing perspective on: * bugs * confusing terminology or tabs * anything that feels hand-wavy or misleading There’s a demo account available if you want to poke around: username: divit.endal4@gmail.com password: password Happy to answer questions about what other things we've built in the space, how these tools work, etc. https://ift.tt/lquNhQf January 16, 2026 at 08:27AM

Thursday, January 15, 2026

Show HN: The Hessian of tall-skinny networks is easy to invert https://ift.tt/fUHnXQD

Show HN: The Hessian of tall-skinny networks is easy to invert It turns out the inverse of the Hessian of a deep net is easy to apply to a vector. Doing this naively takes cubically many operations in the number of layers (so impractical), but it's possible to do this in time linear in the number of layers (so very practical)! This is possible because the Hessian of a deep net has a matrix polynomial structure that factorizes nicely. The Hessian-inverse-product algorithm that takes advantage of this is similar to running backprop on a dual version of the deep net. It echoes an old idea of Pearlmutter's for computing Hessian-vector products. Maybe this idea is useful as a preconditioner for stochastic gradient descent? https://ift.tt/Uvh1Ydu January 16, 2026 at 02:06AM

Show HN: I built an 11MB offline PDF editor because mobile Acrobat is 500MB https://ift.tt/QT0vetK

Show HN: I built an 11MB offline PDF editor because mobile Acrobat is 500MB https://revpdf.com/ January 16, 2026 at 12:30AM

Show HN: OpenWork – an open-source alternative to Claude Cowork https://ift.tt/62RzSWt

Show HN: OpenWork – an open-source alternative to Claude Cowork hi hn, i built openwork, an open-source, local-first system inspired by claude cowork. it’s a native desktop app that runs on top of opencode (opencode.ai). it’s basically an alternative gui for opencode, which (at least until now) has been more focused on technical folks. the original seed for openwork was simple: i have a home server, and i wanted my wife and i to be able to run privileged workflows. things like controlling home assistant, or deploying custom web apps (e.g. our customs recipe app recipes.benjaminshafii.com), legal torrents, without living in a terminal. our initial setup was running the opencode web server directly and sharing credentials to it. that worked, but i found the web ui unreliable and very unfriendly for non-technical users. the goal with openwork is to bring the kind of workflows i’m used to running in the cli into a gui, while keeping a very deep extensibility mindset. ideally this grows into something closer to an obsidian-style ecosystem, but for agentic work. some core principles i had in mind: - open by design: no black boxes, no hosted lock-in. everything runs locally or on your own servers. (models don’t run locally yet, but both opencode and openwork are built with that future in mind.) - hyper extensible: skills are installable modules via a skill/package manager, using the native opencode plugin ecosystem. - non-technical by default: plans, progress, permissions, and artifacts are surfaced in the ui, not buried in logs. you can already try it: - there’s an unsigned dmg - or you can clone the repo, install deps, and if you already have opencode running it should work right away it’s very alpha, lots of rough edges. i’d love feedback on what feels the roughest or most confusing. happy to answer questions. https://ift.tt/3ZKe06P January 14, 2026 at 10:25AM

Wednesday, January 14, 2026

Show HN: A fast CLI and MCP server for managing Lambda cloud GPU instances https://ift.tt/0jFPBKM

Show HN: A fast CLI and MCP server for managing Lambda cloud GPU instances I built an unofficial CLI and MCP server for Lambda cloud GPU instances. The main idea: your AI agents can now spin up and manage Lambda GPUs for you. The MCP server exposes tools to find, launch, and terminate instances. Add it to Claude Code, Cursor, or any agent with one command and you can say things like "launch an H100, ssh in, and run big_job.py" Other features: - Notifications via Slack, Discord, or Telegram when instances are SSH-ready - 1Password support for API keys - Also includes a standalone CLI with the same functionality Written in Rust. MIT licensed. Note: This is an unofficial community project, not affiliated with Lambda. https://ift.tt/uWFIS12 January 15, 2026 at 01:15AM

Show HN: Sparrow-1 – Audio-native model for human-level turn-taking without ASR https://ift.tt/x7cN2Ra

Show HN: Sparrow-1 – Audio-native model for human-level turn-taking without ASR For the past year I've been working to rethink how AI manages timing in conversation at Tavus. I've spent a lot of time listening to conversations. Today we're announcing the release of Sparrow-1, the most advanced conversational flow model in the world. Some technical details: - Predicts conversational floor ownership, not speech endpoints - Audio-native streaming model, no ASR dependency - Human-timed responses without silence-based delays - Zero interruptions at sub-100ms median latency - In benchmarks Sparrow-1 beats all existing models at real world turn-taking baselines I wrote more about the work here: https://ift.tt/xveHipb... https://ift.tt/15vubL8 January 14, 2026 at 11:31PM

Tuesday, January 13, 2026

Show HN: Timberlogs – Drop-in structured logging for TypeScript https://ift.tt/r5qi96G

Show HN: Timberlogs – Drop-in structured logging for TypeScript Hi HN! I built Timberlogs because I was tired of console.log in production and existing logging solutions requiring too much setup. Timberlogs is a drop-in structured logging library for TypeScript: npm install timberlogs-client import { createTimberlogs } from "timberlogs-client"; const timber = createTimberlogs({ source: "my-app", environment: "production", apiKey: process.env.TIMBER_API_KEY, }); timber.info("User signed in", { userId: "123" }); timber.error("Payment failed", error); Features: - Auto-batching with retries - Automatic redaction of sensitive data (passwords, tokens) - Full-text search across all your logs - Real-time dashboard - Flow tracking to link related logs It's currently in beta and free to use. Would love feedback from the HN community. Site: https://timberlogs.dev Docs: https://ift.tt/IbHUM5j npm: https://ift.tt/JhmFwp6 GitHub: https://ift.tt/2z4FtkC January 14, 2026 at 12:13AM

Show HN: Self-host Reddit – 2.38B posts, works offline, yours forever https://ift.tt/f1d5AVc

Show HN: Self-host Reddit – 2.38B posts, works offline, yours forever Reddit's API is effectively dead for archival. Third-party apps are gone. Reddit has threatened to cut off access to the Pushshift dataset multiple times. But 3.28TB of Reddit history exists as a torrent right now, and I built a tool to turn it into something you can browse on your own hardware. The key point: This doesn't touch Reddit's servers. Ever. Download the Pushshift dataset, run my tool locally, get a fully browsable archive. Works on an air-gapped machine. Works on a Raspberry Pi serving your LAN. Works on a USB drive you hand to someone. What it does: Takes compressed data dumps from Reddit (.zst), Voat (SQL), and Ruqqus (.7z) and generates static HTML. No JavaScript, no external requests, no tracking. Open index.html and browse. Want search? Run the optional Docker stack with PostgreSQL – still entirely on your machine. API & AI Integration: Full REST API with 30+ endpoints – posts, comments, users, subreddits, full-text search, aggregations. Also ships with an MCP server (29 tools) so you can query your archive directly from AI tools. Self-hosting options: - USB drive / local folder (just open the HTML files) - Home server on your LAN - Tor hidden service (2 commands, no port forwarding needed) - VPS with HTTPS - GitHub Pages for small archives Why this matters: Once you have the data, you own it. No API keys, no rate limits, no ToS changes can take it away. Scale: Tens of millions of posts per instance. PostgreSQL backend keeps memory constant regardless of dataset size. For the full 2.38B post dataset, run multiple instances by topic. How I built it: Python, PostgreSQL, Jinja2 templates, Docker. Used Claude Code throughout as an experiment in AI-assisted development. Learned that the workflow is "trust but verify" – it accelerates the boring parts but you still own the architecture. Live demo: https://online-archives.github.io/redd-archiver-example/ GitHub: https://ift.tt/mq3lREd (Public Domain) Pushshift torrent: https://ift.tt/ovtLhm9... https://ift.tt/mq3lREd January 13, 2026 at 09:05PM

Monday, January 12, 2026

Show HN: Agent-of-empires: opencode and claudecode session manager https://ift.tt/UrCfuDz

Show HN: Agent-of-empires: opencode and claudecode session manager Hi! I’m Nathan: an ML Engineer at Mozilla.ai: I built agent-of-empires (aoe): a CLI application to help you manage all of your running Claude Code/Opencode sessions and know when they are waiting for you. - Written in rust and relies on tmux for security and reliability - Monitors state of cli sessions to tell you when an agent is running vs idle vs waiting for your input - Manage sessions by naming them, grouping them, configuring profiles for various settings I'm passionate about getting self-hosted open-weight LLMs to be valid options to compete with proprietary closed models. One roadblock for me is that although tools like opencode allow you to connect to Local LLMs (Ollama, lm studio, etc), they generally run muuuuuch slower than models hosted by Anthropic and OpenAI. I would start a coding agent on a task, but then while I was sitting waiting for that task to complete, I would start opening new terminal windows to start multitasking. Pretty soon, I was spending a lot of time toggling between terminal windows to see which one needed me: like help in adding a clarification, approving a new command, or giving it a new task. That’s why I build agent-of-empires (“aoe”). With aoe, I can launch a bunch of opencode and Claude Code sessions and quickly see their status or toggle between them, which helps me avoid having a lot of terminal windows open, or having to manually attach and detach from tmux sessions myself. It’s helping me give local LLMs a fair try, because them being slower is now much less of a bottleneck. You can give it an install with curl -fsSL https://ift.tt/BiKoxgV... | bash Or brew install njbrake/aoe/aoe And then launch by simply entering the command `aoe`. I’m interested in what you think as well as what features you think would be useful to add! I am planning to add some further features around sandboxing (with docker) as well as support for intuitive git worktrees and am curious if there are any opinions about what should or shouldn’t be in it. I decided against MCP management or generic terminal usage, to help keep the tool focused on parts of agentic coding that I haven’t found a usable solution for. I hit the character limit on this post which prevented me from including a view of the output, but the readme on the github link has a screenshot showing what it looks like. Thanks! https://ift.tt/GpbtRXy January 12, 2026 at 07:53PM

Show HN: Yolobox – Run AI coding agents with full sudo without nuking home dir https://ift.tt/jAwCBPQ

Show HN: Yolobox – Run AI coding agents with full sudo without nuking home dir https://ift.tt/0Lr8BP9 January 13, 2026 at 12:04AM

Sunday, January 11, 2026

Show HN: A MCP for controlling terminal UI apps built with bubbletea and ratatui https://ift.tt/mMg0Zfp

Show HN: A MCP for controlling terminal UI apps built with bubbletea and ratatui so you can start vibe-coding your ad-hoc terminal dashboard. With session replay and mouse click support built-in. https://ift.tt/8mGrM2f January 12, 2026 at 01:24AM

Show HN: Epstein IM – Talk to Epstein clone in iMessage https://ift.tt/pQ0MS7h

Show HN: Epstein IM – Talk to Epstein clone in iMessage https://epstein.im/ January 11, 2026 at 06:28AM

Saturday, January 10, 2026

Show HN: Persistent Memory for Claude Code (MCP) https://ift.tt/knrZw2M

Show HN: Persistent Memory for Claude Code (MCP) This is my attempt in building a memory that evolves and persist for claude code. My approach is inspired from Zettelkasten method, memories are atomic, connected and dynamic. Existing memories can evolve based on newer memories. In the background it uses LLM to handle linking and evolution. I have only used it with claude code so far, it works well with me but still early stage, so rough edges likely. I'm planning to extend it to other coding agents as I use several different agents during development. Looking for feedbacks! https://ift.tt/vcaufzm January 11, 2026 at 02:04AM

Show HN: I used Claude Code to discover connections between 100 books https://ift.tt/6jOoiuZ

Show HN: I used Claude Code to discover connections between 100 books I think LLMs are overused to summarise and underused to help us read deeper. I built a system for Claude Code to browse 100 non-fiction books and find interesting connections between them. I started out with a pipeline in stages, chaining together LLM calls to build up a context of the library. I was mainly getting back the insight that I was baking into the prompts, and the results weren't particularly surprising. On a whim, I gave CC access to my debug CLI tools and found that it wiped the floor with that approach. It gave actually interesting results and required very little orchestration in comparison. One of my favourite trail of excerpts goes from Jobs’ reality distortion field to Theranos’ fake demos, to Thiel on startup cults, to Hoffer on mass movement charlatans ( https://ift.tt/4e8AHdo ). A fun tendency is that Claude kept getting distracted by topics of secrecy, conspiracy, and hidden systems - as if the task itself summoned a Foucault’s Pendulum mindset. Details: * The books are picked from HN’s favourites (which I collected before: https://ift.tt/G4IM3Lj ). * Chunks are indexed by topic using Gemini Flash Lite. The whole library cost about £10. * Topics are organised into a tree structure using recursive Leiden partitioning and LLM labels. This gives a high-level sense of the themes. * There are several ways to browse. The most useful are embedding similarity, topic tree siblings, and topics cooccurring within a chunk window. * Everything is stored in SQLite and manipulated using a set of CLI tools. I wrote more about the process here: https://ift.tt/HqiBnbu I’m curious if this way of reading resonates for anyone else - LLM-mediated or not. https://ift.tt/7RfTOSW January 10, 2026 at 10:26PM

Friday, January 9, 2026

Show HN: Various shape regularization algorithms https://ift.tt/YSVNgyd

Show HN: Various shape regularization algorithms Shape regularization is a technique used in computational geometry to clean up noisy or imprecise geometric data by aligning segments to common orientations and adjusting their positions to create cleaner, more regular shapes. I needed a Python implementation so started with the examples implemented in CGAL then added a couple more for snap and joint regularization and metric regularization. https://ift.tt/dKZsLDi January 9, 2026 at 07:43AM

Thursday, January 8, 2026

Wednesday, January 7, 2026

Show HN: I visualized the entire history of Citi Bike in the browser https://ift.tt/6VUyFx3

Show HN: I visualized the entire history of Citi Bike in the browser Each moving arrow represents one real bike ride out of 291 million, and if you've ever taken a Citi Bike before, you are included in this massive visualization! You can search for your ride using Cmd + K and your Citi Bike receipt, which should give you the time of your ride and start/end station. Everything is open source: https://ift.tt/bI7nYE4 Some technical details: - No backend! Processed data is stored in parquet files on a Cloudflare CDN, and queried directly by DuckDB WASM - deck.gl w/ Mapbox for GPU-accelerated rendering of thousands of concurrent animated bikes - Web Workers decode polyline routes and do as much precomputation as possible off the main thread - Since only (start, end) station pairs are provided, routes are generated by querying OSRM for the shortest path between all 2,400+ station pairs https://bikemap.nyc/ January 8, 2026 at 12:27AM

Show HN: bikemap.nyc – visualization of the entire history of Citi Bike https://ift.tt/kijJAMV

Show HN: bikemap.nyc – visualization of the entire history of Citi Bike Each moving arrow represents a real bike ride. There are 291 million rides in total, covering 12 years of history from June 2013 to December 2025, based on public data published by Lyft. If you've ever taken a Citi Bike ride before, you are included in this massive visualization! You can search for your ride using Cmd + K and your Citi Bike receipt, which should give you the time of your ride and start/end station. Some technical details: - No backend! Processed data is stored in parquet files on a CDN, and queried directly by DuckDB WASM - deck.gl w/ Mapbox for GPU-accelerated rendering of thousands of concurrent animated bikes - Web Workers decode polyline routes and do as much precomputation as possible off the main thread - Since only (start, end) station pairs are provided, routes are generated by querying OSRM for the shortest path between all 2,400+ station pairs Legend: - Blue = E-Bike - Purple = Classic Bike - Red = Bike docked - Green = Bike unlocked https://ift.tt/TEnKoX6 January 8, 2026 at 02:15AM

Show HN: Seapie – a Python debugger where breakpoints drop into a REPL https://ift.tt/xzg9pRN

Show HN: Seapie – a Python debugger where breakpoints drop into a REPL https://ift.tt/sh6OM0j January 7, 2026 at 11:28PM

Tuesday, January 6, 2026

Show HN: llmgame.ai – The Wikipedia Game but with LLMs https://ift.tt/lc34gHv

Show HN: llmgame.ai – The Wikipedia Game but with LLMs I used to play the Wikipedia Game in high school and had an idea for applying the same mechanic of clicking from concept to concept to LLMs. Will post another version that runs with an LLM entirely in the browser soon, but for now, please enjoy as long as my credits last... Warning: the LLM does not always cooperate https://www.llmgame.ai January 6, 2026 at 08:37AM

Show HN: CLI for internet speed test via Cloudflare with optional TUI https://ift.tt/4QZjOf9

Show HN: CLI for internet speed test via Cloudflare with optional TUI I've been having a lot of issues with my upload speed and wanted a cli tool to run in a cron job and couldn't quite find one to suit my needs. Sharing here in case others find it useful. https://ift.tt/mVvZF5B January 6, 2026 at 11:46PM

Monday, January 5, 2026

Show HN: WOLS – Open standard for mushroom cultivation tracking https://ift.tt/jcaCZ6S

Show HN: WOLS – Open standard for mushroom cultivation tracking I built an open labeling standard for tracking mushroom specimens through their lifecycle (from spore/culture to harvest). v1.1 adds clonal generation tracking (distinct from filial/strain generations) and conforms to JSON-LD for interoperability with agricultural/scientific data systems. Spec (CC 4.0): https://ift.tt/nuRJgHW Client libraries (Apache 2.0): Python + CLI: pip install wols (also on GHCR) TypeScript/JS: npm install @wemush/wols Background: Mycology has fragmented data practices (misidentified species, inconsistent cultivation logs, no shared vocabulary for tracking genetics across generations). This is an attempt to fix that. Looking for feedback from anyone working with biological specimen tracking, agricultural data systems, or mycology. https://ift.tt/NaRkZ3A January 6, 2026 at 12:00AM

Show HN: Bilig – a headless spreadsheet engine for Node services and agents https://ift.tt/qb3OXxh

Show HN: Bilig – a headless spreadsheet engine for Node services and agents https://ift.tt/pKHIE3k May 7, 2026 at 11:46PM