This is a autopost bolg frinds we are trying to all latest sports,news,all new update provide for you
Monday, May 4, 2026
Show HN: nfsdiag - a NFS diagnostic application https://ift.tt/4McOjar
Show HN: nfsdiag - a NFS diagnostic application https://ift.tt/sctymXx May 2, 2026 at 06:18PM
Sunday, May 3, 2026
Show HN: Tyche: An experimental distributed trading pipeline in Go Java https://ift.tt/6oX8b9W
Show HN: Tyche: An experimental distributed trading pipeline in Go Java https://ift.tt/JVDgLKY May 4, 2026 at 01:13AM
Show HN: Ableton Live MCP https://ift.tt/biVG5OL
Show HN: Ableton Live MCP https://ift.tt/Eg0h13Y May 3, 2026 at 11:35PM
Saturday, May 2, 2026
Show HN: Use an Android Phone as an HTTP Proxy https://ift.tt/agH6XVD
Show HN: Use an Android Phone as an HTTP Proxy I created a simple project to allow you to use a phone as a web proxy. This is not a proxy for the phone, its a way to proxy web traffic from elsewhere via the phone. One practical use case is accessing geo-restricted content. If you have a trusted contact in the country with an Android phone, this can serve as a simple alternative to a commercial VPN. To set it up you need to run a proxy server which can run as a docker container. You then need to install the app on the Android phone which will connect to the server. Finally you configure a browser to use the proxy server as the HTTP/HTTPS proxy. More details here: https://ift.tt/FpkSEhw Let me know how you go and if you run into any issues. https://ift.tt/lARwhIN May 3, 2026 at 05:44AM
Show HN: State of the Art of Coding Models, According to Hacker News Commenters https://ift.tt/GXO07JB
Show HN: State of the Art of Coding Models, According to Hacker News Commenters Hello HN, I was away from my computer for two weeks, and after coming back and reading the latest discussions on HN about coding assistants (models, harnesses), I felt very out of the loop. My normal process would have been to keep reading and figure out the latest and greatest from people's comments, but I wanted to try and automate this process. Basically the goal is to get a quick overview over which coding models are popular on HN. A next iteration could also scan for harnesses that people use, or info on self-hosting or hardware setups. I wrote a short intro on the page about the pipeline that collects and analyzes the data, but feel free to ask for more details or check the Google Sheet for more info. https://hnup.date/hn-sota https://hnup.date/hn-sota May 3, 2026 at 02:55AM
Show HN: Clipmon is a macOS clipboard manager on steroids https://ift.tt/2CUvF3Z
Show HN: Clipmon is a macOS clipboard manager on steroids https://ift.tt/mytUrvX May 3, 2026 at 01:59AM
Show HN: Rust library for Undo/Redo using deltas, snapshots or commands https://ift.tt/x0UVW3u
Show HN: Rust library for Undo/Redo using deltas, snapshots or commands https://ift.tt/VBRpf3b May 3, 2026 at 12:11AM
Friday, May 1, 2026
Show HN: Destiny – Claude Code's fortune Teller skill https://ift.tt/esyaEmx
Show HN: Destiny – Claude Code's fortune Teller skill Destiny is the Claude Code's plugin that gives you a real fortune reading. Type /destiny to see today's destiny! It uses the actual classical East Asian astrology system. You enter your birthday once, then /destiny gives you today's reading anytime. Two layers, kept honest: 1. The numbers (your eight-character birth chart, today's day pillar, the hexagram for the moment, five-element relationships) are computed by a Python script. Same person + same day = identical output. You can verify against any traditional calendar source. 2. The prose (today's stars, character sketch, life arc, advice) is written by Claude, applying centuries-old reading conventions to that fixed data. Not LLM-hallucinated horoscope. If you have fun with it, a star would mean a lot. https://ift.tt/NBtcVsn May 2, 2026 at 01:26AM
Show HN: GhostBox – Borrow a disposable little machine from the Global Free Tier https://ift.tt/SqLVRIc
Show HN: GhostBox – Borrow a disposable little machine from the Global Free Tier I built this because I was always creating machines on GH actions to test builds on different OS, and I wanted a tight CLI that could do it. I always saw Actions as this great resources and ephemeral machines you could do dev work in just were a natural way for me to work, so this grew out of that workflow. I didn't expect it to blow up, so it wasn't 100% finished when I posted it. But it should stabilize pretty quickly. Happy to know what you think and talk about it. https://ift.tt/E13Wfsx May 1, 2026 at 08:22PM
Show HN: Blotter, a live map of police radio activity https://ift.tt/NrWd8Qc
Show HN: Blotter, a live map of police radio activity https://blotter.fm May 1, 2026 at 11:33PM
Thursday, April 30, 2026
Show HN: Annotate your screen recordings and export to PDF online, no uploads https://ift.tt/Ui8HBRD
Show HN: Annotate your screen recordings and export to PDF online, no uploads https://framepin.com/ May 1, 2026 at 01:47AM
Show HN: Exploding Hamsters https://ift.tt/Ns5Jelh
Show HN: Exploding Hamsters https://ift.tt/rW8YtLH May 1, 2026 at 12:20AM
Wednesday, April 29, 2026
Show HN: A Multi User Multi Task Board MCP Server https://ift.tt/wMKFjJu
Show HN: A Multi User Multi Task Board MCP Server I built a simple multi user, multi board, Task/Kanban MCP server. I have been looking for something like this to manage development agents, but I wasn't seeing anything that felt like what I wanted. So I set down and decided to vibe code an alternative. While it was an experiment at first I have been using it daily for my personal development projects and I really think there are others who might be looking for exactly this. It's 100% a WIP, but it is also very usable. I have a demo instance running at https://mootasks.dev . If you find this interesting I'd appreciate a star. This is really the first thing I built that I felt would be of interest to others. The readme explains it, but if you have docker you can get this running in a couple minutes. It's helped my workflow a lot and I plan on continuing to add features / improve it. https://ift.tt/YXvrp3c April 30, 2026 at 01:11AM
Show HN: Generative UI Library for React https://ift.tt/ui7lsro
Show HN: Generative UI Library for React https://ift.tt/AJnP1xi April 30, 2026 at 12:58AM
Tuesday, April 28, 2026
Show HN: Drive any macOS app in the background without stealing the cursor https://ift.tt/db9OEcK
Show HN: Drive any macOS app in the background without stealing the cursor Hi HN, Francesco from Cua here. I hacked this project together last weekend, inspired by the Codex Computer-Use release and lessons learned from deploying GUI-operating agents for our customers. The main problem: when a UI automation process controls a desktop app today, it usually takes over the human’s session. Your cursor moves, keyboard focus gets stolen, windows jump to the front, and you have to stop working until the agent is done. That is why we have historically avoided encouraging users to run these processes directly on their host machine, instead relying on VMs or GUI containers for concurrency and background execution. But computer-use - the tools we give agents to operate computers like humans - does not scale cleanly that way. As models get smarter, agents need to share hosts safely, run in the background, and avoid collisions with the human or other agents using the same machine. We realized macOS has no first-class API for "drive this app without touching the cursor". CGEventPost routes through the hardware input stream, so it moves your cursor. CGEvent.postToPid avoids the cursor warp, but Chromium treats those events as untrusted and silently drops clicks at the renderer boundary. Activating the target app first raises the window and pulls focus, defeating the point of background execution. Cua Driver is our attempt at a real fix: a background computer-use driver for macOS that lets an agent click, type, scroll, and read native apps while your cursor, frontmost app, and Space stay where they are. The default interface is a CLI, so it is easy to script or call from any coding agent shell. Try it on macOS 14+: /bin/bash -c "$(curl -fsSL https://ift.tt/szqG6MU... )" The first internal use case was delegated demo recording. We ask Claude Code to drive an app while 'cua-driver recording start' captures the trajectory, screenshots, actions, and click markers. The result is an agent-generated product demo, Screen Studio inspired. Other things we have used it for: - Replacing Vercel’s agent-browser and other browser-use CLIs. With Claude Code and Cua Driver, you do not need Chrome DevTools Protocol at all. - A dev-loop QA agent that reproduces a visual bug, edits code, rebuilds, and verifies the UI while my editor stays frontmost. - Personal-assistant flows that use iMessage from Claude Code, Hermes, or other general-purpose agent CLIs. - Pulling visual context from Chrome, Figma, Preview, or YouTube windows I am not looking at, without relying on their APIs. What made this harder than expected: - CGEventPost warps the cursor because it goes through the HID stream. - CGEvent.postToPid does not warp the cursor, but Chromium drops it at the renderer IPC boundary. - Activating the target first raises the window and can drag you across Spaces. - Electron apps stop keeping useful AX trees alive when windows are occluded without a private remote-aware SPI. The unlock was SkyLight. SLEventPostToPid is a sibling of the public per-PID call, but it travels through a WindowServer channel Chromium accepts as trusted. Pair it with yabai’s focus-without-raise pattern, plus an off-screen primer click at (-1, -1), and the click lands without the window ever raising. One thing we learned: the right addressing mode depends on the app. Native macOS apps usually have rich AX trees, Chromium-family apps often need a hybrid of AX and screenshots, and apps like Blender or CAD tools may expose almost no useful AX surface. The mistake is defaulting to pixels everywhere - or defaulting to AX everywhere. Long technical writeup: https://ift.tt/48nvpVC... I would like feedback from people building Mac automation, agent harnesses, or accessibility tooling. If it breaks on an macOS app you care about, that is useful data for us. https://ift.tt/UMPAEOh April 28, 2026 at 09:33PM
Show HN: Open Bias – proxy that enforces agent behavior at runtime https://ift.tt/QuezU9l
Show HN: Open Bias – proxy that enforces agent behavior at runtime https://ift.tt/6vn0jwM April 29, 2026 at 12:02AM
Monday, April 27, 2026
Show HN: Waiting for LLMs Suck – Give your user a game https://ift.tt/lemYrBw
Show HN: Waiting for LLMs Suck – Give your user a game Give your user a game while they wait for the LLM to return a result. https://ift.tt/oD5GFjg April 28, 2026 at 08:15AM
Show HN: PrePrompt – rewrites vague prompts before they reach the LLM https://ift.tt/M8QPNl3
Show HN: PrePrompt – rewrites vague prompts before they reach the LLM https://preprompt.org/ April 28, 2026 at 08:19AM
Show HN: AgentSwift – Open-source iOS builder agent https://ift.tt/jpQbazJ
Show HN: AgentSwift – Open-source iOS builder agent I'm working on a coding agent for building iOS apps. It's built on openspec and xcodebuildmcp. It's free and open source. https://ift.tt/ayU1vRp April 28, 2026 at 06:44AM
Show HN: 49Agents – Infinite canvas IDE for AI agents https://ift.tt/G2oJXO4
Show HN: 49Agents – Infinite canvas IDE for AI agents https://ift.tt/snRje1p April 28, 2026 at 06:06AM
Sunday, April 26, 2026
Show HN: WaveletLM – wavelet-based, attention-free model with O(n log n) scaling https://ift.tt/gSdUNJ4
Show HN: WaveletLM – wavelet-based, attention-free model with O(n log n) scaling WaveletLM is a wavelet-based, attention-free architecture that replaces self-attention with learned lifting wavelet decomposition, a Fast Walsh-Hadamard Transform, per-scale gated spectral mixing with SwiGLU activation, an inverse FWHT, and wavelet reconstruction. Combined with expanded MLPs and sparse product-key memory, this yields a model with O(n log n) scaling in sequence length. With 23.8 PPL on WikiText-103, WaveletLM beats both GPT-2 Medium, which was trained on 80× more data, and Transformer-XL Standard, which uses recurrence to extend its effective context. It is undertrained and underregularized due to budget constraints, so there is much room for development and improvement. I invite anyone who is curious to examine the model, test it out, and extend its capabilities further. All code and weights are fully open source, and a PG-19 run will be completed in 2-3 days. Generations can be done in 4-5 GB VRAM at 28.8 tokens/second, and the model is trainable in 16.25 hours with 20 GB of VRAM, both on a 5090. README for comparison tables, instructions, logs, and future plans: https://ift.tt/jY0IGMQ Weights: https://ift.tt/CWcy6kh Generations: https://ift.tt/ywvfZTE... The following samples were chosen for coherence, not factual accuracy. Factuality will require scaling and downstream techniques such as RAG and instruction tuning. > The history of the city is reflected in its architecture, which includes the historic Old Town and New Castle County Courthouse Square Historic District. The building was designed by John H. Stevens, who also designed the Albany-Fulton Celebration in 1906 and built a steel-hulled shipyard on the lake shore. > The album was released on August 25, 2007 by Sony Music Entertainment and features several songs from the record including "Never Say Die", "The Show", "Don't Cry for Me Argentina" and a cover of "I Can Only Imagine (But You Are Not Alone)". > The species was first described by Swedish zoologist Carl Linnaeus in 1758 as Agaricus adustus. The genus name is derived from the Latin words perma "to tie", and pous ("like") means "with a large head". In 1821, French mycologists Jean-Baptiste de Lacaille placed it in section Cricetae of the order Carnivora. He later renamed it Spongiforma punctata after the Greek kribensis. https://ift.tt/jY0IGMQ April 26, 2026 at 11:18PM
Saturday, April 25, 2026
Show HN: Odozi – open-source iOS journaling app https://ift.tt/dUHzleo
Show HN: Odozi – open-source iOS journaling app Yeah I know I hate the name too but I wasn't about to pay up for odyssey.app. It's an open source project so feel free to poke around with it / fork it. I talk about it more on the marketing website, but a few of us have been using it for the past month and kind of fun. Obviously there will be a slew of issues / feedback / nits that come from this, but c'est la vie. GH is here: https://ift.tt/ZwM6qTV https://odozi.app April 25, 2026 at 09:22PM
Show HN: Quay – Menu-bar Git sync https://ift.tt/tRCBDVd
Show HN: Quay – Menu-bar Git sync I write Astro blog posts in a text editor; when I'm done I want them pushed to GitHub so Cloudflare deploys the site. To make it comfortable, I built Quay for the menu bar. Also useful for Obsidian vault syncing. Point it at a folder, connect a GitHub repo, and it stages/commits/pushes/pulls. Multiple repos, editable commit messages, branch switching, merges with conflict detection. Shows open issue and PR counts per repo. But it's is not a full Git client (no diffs, blame, cherry-pick, or rebase) and it doesn't create remote repos. Native macOS app (Swift/SwiftUI). Wraps the local git binary (prompts to install Xcode Command Line Tools if missing). No custom Git implementation. Sandboxed, no telemetry, GitHub-only. macOS. 7-day trial, €9 one-time on the App Store. https://ift.tt/Q3OdJbp April 25, 2026 at 11:53PM
Friday, April 24, 2026
Show HN: #1 On This Day https://ift.tt/OxKgplA
Show HN: #1 On This Day https://onthisday-theta.vercel.app April 24, 2026 at 09:42PM
Show HN: TurbineFi – Build, Backtest, Deploy Prediction Market Strategies https://ift.tt/EyeQ0H5
Show HN: TurbineFi – Build, Backtest, Deploy Prediction Market Strategies Hey HN! We just finished our first major build of TurbineFi, an AI-assisted workflow for building, backtesting, and running prediction market strategies. There are over 1,000 community strategies you can try out, there's a backtesting engine integrated in the workflow, and you get your own sandbox to execute the trades 24/7. Currently live for Kalshi, Polymarket coming soon. We developed a custom DSL to make compiling AI-assisted strategies more deterministic than raw python generation, so creating a strategy takes seconds even on low-tier models (thinking of migrating to a self-hosted model soon to reduce costs). We also worked with Locus (YCF25) to do the sandbox provisioning, so that we never manage keys for users. When a user signs up with their email, Privy creates a wallet for them, and then that wallet uses the X402 agent payment protocol to pay for their own server. We created a deployment harness around it that accepts and runs new code via a hosted API, so once it's up, every deployment is authorized by EIP-712 signatures. It keeps everything non-custodial, and code deployments happen in seconds. And users don't really realize they're using crypto rails. Turbine also includes weather and crypto historical information, so you can do things like fading the BTC-15min UP markets when it's cold in NYC, and backtest and run it in seconds. Adding sports data soon. There's a 7-day trial if you want to poke around. Would appreciate feedback on which strategies you'd want to try first, so we can make sure we have the infra to support them. Thank you! https://ift.tt/cpZbohW April 24, 2026 at 08:47PM
Thursday, April 23, 2026
Show HN: Agent Vault – A HTTP credential proxy and vault for AI agents https://ift.tt/zCrwKIH
Show HN: Agent Vault – A HTTP credential proxy and vault for AI agents https://ift.tt/mNbvEqk April 22, 2026 at 09:55PM
Show HN: AgentSearch – Self-hosted search and MCP for AI agents, no API keys https://ift.tt/PWSFz2m
Show HN: AgentSearch – Self-hosted search and MCP for AI agents, no API keys https://ift.tt/1RdrNAw April 23, 2026 at 11:55PM
Show HN: Turning a Gaussian Splat into a videogame https://ift.tt/k40HC2X
Show HN: Turning a Gaussian Splat into a videogame https://ift.tt/eNdAS4P April 23, 2026 at 07:48PM
Wednesday, April 22, 2026
Show HN: One ESLint rule to kill the "ChatGPT em dash" in your codebase https://ift.tt/e2ibl0f
Show HN: One ESLint rule to kill the "ChatGPT em dash" in your codebase https://ift.tt/ALx4DZX April 23, 2026 at 01:27AM
Show HN: Netlify for Agents https://ift.tt/vcZf6wF
Show HN: Netlify for Agents I launched Netlify with a Show HN more than 11 years today, for humans. Today we're launching our Agent first version of Netlify. Super early days for this, but I expect it to become as important as our original launch over time. It's as hard to perfect these flows as it was to perfect some of the initial human DX flows, since the agents are non-deterministic and keeps changing and evolving, and we'll have more to show soon on our eval tooling for this. Try it out with an agent, and we would love feedback on what works and what doesn't as we keep iterating on making Netlify better for our new agent friends. https://netlify.ai April 22, 2026 at 10:27PM
Tuesday, April 21, 2026
Show HN: Agent Brain Trust, customisable expert panels for AI agents https://ift.tt/UuIovi4
Show HN: Agent Brain Trust, customisable expert panels for AI agents Agent Brain Trust lets you summon a panel of real, named experts to critique your architecture, review your writing, pressure your product strategy, or debate your design patterns. 10 built-in trusts, an extensible roster, and a working turn-taking protocol that ensures nothing useful gets skipped. Guest experts are drafted via an MCP server that maps topics to real persona cards so the panel can reach into niche and novel territory without inventing expertise it does not have. Wrote up the full thinking here: https://tinyurl.com/agent-brain-trust https://ift.tt/ELmYfTX April 22, 2026 at 04:33AM
Show HN: Almanac MCP, turn Claude Code into a Deep Research agent https://ift.tt/g9lw6WQ
Show HN: Almanac MCP, turn Claude Code into a Deep Research agent I am Rohan, and I have grown really frustrated with CC's search and read tools. They use Haiku to summarise all the search results, so it is really slow and often ends up being very lossy. I built this MCP that you can install into your coding agents so they can actually access the web properly. Right now it can: - search the general web - search Reddit - read and scrape basically any webpage Install it: npx openalmanac setup The MCP is completely free to use. We have also built a central store where you can contribute things you learned while exploring. If you find something useful, you can contribute it to the encyclopedia we're building at Almanac using the same MCP. https://ift.tt/wNWkHb4 April 22, 2026 at 03:42AM
Show HN: Backlit Keyboard API for Python https://ift.tt/Cj5Y0wT
Show HN: Backlit Keyboard API for Python It currently supports Linux as of now. You can use this package to tinker with many things. Let's say, if you want to make a custom notification system, like if your website is down, you can make a blink notification with it. MacOS support is underway. I haven't tested Windows yet, I don't use it anymore btw. In future, if this package reaches nice growth, I'll be happy to make a similar Rust crate for it. https://ift.tt/WfKFrC7 April 19, 2026 at 12:22PM
Monday, April 20, 2026
Show HN: Simple CLI tool to convert PDFs to dark mode, with TOC preservation https://ift.tt/zUGjxSl
Show HN: Simple CLI tool to convert PDFs to dark mode, with TOC preservation Hi HN, I made a little something that could be useful to those like me that read pdfs at night. https://ift.tt/Rp05mIs April 21, 2026 at 01:52AM
Show HN: Git Push No-Mistakes https://ift.tt/TBa0DRN
Show HN: Git Push No-Mistakes no-mistakes is how I kill AI slop. It puts a local git proxy in front of my real remote. I push to no-mistakes instead of origin, and it spins up a disposable worktree, runs my coding agent as a validation pipeline, forwards upstream only after every check passes, opens a clean PR automatically, and babysits CI pipeline for me. https://ift.tt/0imugwr April 21, 2026 at 12:10AM
Show HN: AI Coding Agent Guardrails enforced at runtime https://ift.tt/OmHtoIU
Show HN: AI Coding Agent Guardrails enforced at runtime Hello, looking for some users interested using a devtool that allows developers to centrally manage AI Coding Agent tools that supports all AI Coding Agent tools like Claude Code, Codex, Antigravity, etc. Try it free! https://ift.tt/k8ZKD4F... https://sigmashake.com April 20, 2026 at 10:55PM
Sunday, April 19, 2026
Show HN: How context engineering works, a runnable reference https://ift.tt/Fw3ufSy
Show HN: How context engineering works, a runnable reference I've been presenting at local meetups about Context Engineering, RAG, Skills, etc.. I even have a vbrownbag coming up on LinkedIn about this topic so I figured I would make a basic example that uses bedrock so I can use it in my talks or vbrownbags. Hopefully it's useful. https://ift.tt/VhCs8ta April 17, 2026 at 11:50PM
Show HN: Newsmaps.io a map of how news topics are covered by different countries https://ift.tt/OXlukeE
Show HN: Newsmaps.io a map of how news topics are covered by different countries https://ift.tt/lFzq3rB April 20, 2026 at 02:32AM
Show HN: A privacy-first, local-LLM note app for iOS (Google Keep alternative) https://ift.tt/zyYlo6A
Show HN: A privacy-first, local-LLM note app for iOS (Google Keep alternative) https://ift.tt/1FLPVZk April 19, 2026 at 10:29PM
Show HN: Free PDF redactor that runs client-side https://ift.tt/YEPwLih
Show HN: Free PDF redactor that runs client-side I recently needed to verify past employment and to do so I was going to upload paystubs from a previous employer, however I didn't want to share my salary in that role. I did a quick search online and most sites required sign-up or weren't clear about document privacy. I conceded and signed up for a free trial of Adobe Acrobat so I could use their PDF redaction feature. I figured there should be a dead simple way of doing this that's private, so I decided to create it myself. What this does is rasterize each page to an image with your redactions burned in, then it rebuilds the PDF so the text layer is permanently destroyed and not just covered up and easily retrievable. I welcome any and all feedback as this is my first live tool, thanks! https://redactpdf.net April 20, 2026 at 12:09AM
Saturday, April 18, 2026
Show HN: AI Subroutines – Run automation scripts inside your browser tab https://ift.tt/1oVpPrw
Show HN: AI Subroutines – Run automation scripts inside your browser tab We built AI Subroutines in rtrvr.ai. Record a browser task once, save it as a callable tool, replay it at: zero token cost, zero LLM inference delay, and zero mistakes. The subroutine itself is a deterministic script composed of discovered network calls hitting the site's backend as well as page interactions like click/type/find. The key architectural decision: the script executes inside the webpage itself, not through a proxy, not in a headless worker, not out of process. The script dispatches requests from the tab's execution context, so auth, CSRF, TLS session, and signed headers get added to all requests and propagate for free. No certificate installation, no TLS fingerprint modification, no separate auth stack to maintain. During recording, the extension intercepts network requests (MAIN-world fetch/XHR patch + webRequest fallback). We score and trim ~300 requests down to ~5 based on method, timing relative to DOM events, and origin. Volatile GraphQL operation IDs are detected and force a DOM-only fallback before they break silently on the next run. The generated code combines network calls with DOM actions (click, type, find) in the same function via an rtrvr.* helper namespace. Point the agent at a spreadsheet of 500 rows and with just one LLM call parameters are assigned and 500 Subroutines kicked off. Key use cases: - record sending IG DM, then have reusable and callable routine to send DMs at zero token cost - create routine getting latest products in site catalog, call it to get thousands of products via direct graphql queries - setup routine to file EHR form based on parameters to the tool, AI infers parameters from current page context and calls tool - reuse routine daily to sync outbound messages on LinkedIn/Slack/Gmail to a CRM using a MCP server We see the fundamental reason that browser agents haven't taken off is that for repetitive tasks going through the inference loop is unnecessary. Better to just record once, and get the LLM to generate a script leveraging all the possible ways to interact with a site and the wider web like directly calling backed API's, interacting with the DOM, and calling 3P tools/APIs/MCP servers. https://ift.tt/XtmCZne April 18, 2026 at 02:33AM
Show HN: Praxis – Lab data to publication-ready figures in one Python package https://ift.tt/DB2RvWE
Show HN: Praxis – Lab data to publication-ready figures in one Python package https://ift.tt/GJqnglu April 18, 2026 at 11:45PM
Friday, April 17, 2026
Show HN: Pyra – a Python toolchain experiment inspired by uv and Bun https://ift.tt/DYZj17t
Show HN: Pyra – a Python toolchain experiment inspired by uv and Bun I’ve been working on Pyra for the past few months and wanted to start sharing it in public. Right now it’s focused on the core package/project management workflow: Python installs, init, add/remove, lockfiles, env sync, and running commands in the managed env. The bigger thing I’m exploring is whether Python could eventually support a more cohesive toolchain story overall, more in the direction of Bun: not just packaging, but maybe over time testing, tasks, notebooks, and other common workflow tools feeling like one system instead of a bunch of separate pieces. It’s still early, and I’m definitely not claiming it’s as mature as uv. I’m mostly sharing it now because I want honest feedback on whether the direction feels interesting or misguided. https://ift.tt/81YnRZ9 April 18, 2026 at 03:20AM
Show HN: I turned my MacBook notch into a live Claude Code dashboard https://ift.tt/RDJtFzP
Show HN: I turned my MacBook notch into a live Claude Code dashboard https://ift.tt/Yhg9saf April 17, 2026 at 09:13PM
Show HN: Waputer – The WebAssembly Computer https://ift.tt/Yy02Spv
Show HN: Waputer – The WebAssembly Computer Waputer is an operating system that runs entirely in the browser. When you visit the website at https://waputer.app , a kernel written in JavaScript sets up a filesystem and launches a WebAssembly program, which in turn talks to the kernel to handle the display and input. A purely terminal-based version is at https://waputer.dev . My original intention was to create programs that run in the browser that have a lot more in common with the desktop. The traditional "hello world" program is not really suited for the web. Waputer changes that. The GitHub repo at https://ift.tt/TQqLWZ5 gives a very brief overview of compiling a C program and running it on Waputer. There is a blog available from the main site that has a long-form explanation of Waputer and my motivations if you want some additional reading. https://waputer.app April 17, 2026 at 11:16PM
Thursday, April 16, 2026
Show HN: Spice simulation → oscilloscope → verification with Claude Code https://ift.tt/ZDuwzh0
Show HN: Spice simulation → oscilloscope → verification with Claude Code I built MCP servers for my oscilloscope and SPICE simulator so Claude Code can close the loop between simulation and real hardware. https://ift.tt/0Y6KFcO April 17, 2026 at 06:07AM
Show HN: Marky – A lightweight Markdown viewer for agentic coding https://ift.tt/2BL3dqv
Show HN: Marky – A lightweight Markdown viewer for agentic coding Hey HN, In this age of agentic coding I've found myself spending a lot of time reviewing markdown files. Whether it's plans or documentation that I've asked my agent to generate for me, it seems that I spend more time reading markdown than code. I've tried a few different solutions to make it easier to read such as Obsidian however I've found their Vault system to be quite limiting for this use case and I've found TUI solutions to not quite be as friendly to read as I've wanted so I made Marky. Marky is a lightweight desktop application that makes it incredibly easy to read and track your markdown files. It also has a helpful cli so you can just run marky FILENAME and have the app open to the md file that you pointed it at. I've been using the daily over the past week and I really enjoy it so I figured I'd share it. Here's a video if you want to check out a demo: https://www.youtube.com/watch?v=nGBxt8uOVjc . I have plans to add more features such as incorporating agentic tools such as claude code and codex into the UI as well as developing a local git diff reviewer to allow me to do local code review before pushing up to git. I'd love to hear your thoughts and any feature suggestions you may have :) https://ift.tt/RD8eOUC April 16, 2026 at 09:38PM
Show HN: Online Sound Decibel Meter https://ift.tt/qIUM8mD
Show HN: Online Sound Decibel Meter https://ift.tt/myRFVQT April 17, 2026 at 12:09AM
Wednesday, April 15, 2026
Show HN: I built a Wikipedia based AI deduction game https://ift.tt/VtJQm5L
Show HN: I built a Wikipedia based AI deduction game I haven't seen anything like this so I decided to build it in a weekend. How it works: You see a bunch of things pulled from Wikipedia displayed on cards. You ask yes or no questions to figure out which card is the secret article. The AI model has access to the image and wiki text and it's own knowledge to answer your question. Happy to have my credits burned for the day but I'll probably have to make this paid at some point so enjoy. I found it's not easy to get cheap+fast+good responses but the tech is getting there. Most of the prompts are running through Groq infra or hitting a cache keyed by a normalization of the prompt. https://ift.tt/qKZFuv5 April 16, 2026 at 05:43AM
Show HN: US keyboards don't have enough keys, so I switched to Japanese https://ift.tt/Buh9x20
Show HN: US keyboards don't have enough keys, so I switched to Japanese https://ift.tt/wTMn6Xt April 16, 2026 at 02:27AM
Show HN: Jeeves – TUI for browsing and resuming AI agent sessions https://ift.tt/90Tv64P
Show HN: Jeeves – TUI for browsing and resuming AI agent sessions I made Jeeves to search, preview, read through, and resume AI agent sessions in your terminal. It shows sessions across claude and codex in a single view, with more AI agent framework integrations to come. https://ift.tt/rGwFmVg April 16, 2026 at 01:01AM
Show HN: Monadic Networking Library for Go https://ift.tt/S6fPoW2
Show HN: Monadic Networking Library for Go A library built on top of ibm/fp-go for use in networking applications (servers, etc.) https://ift.tt/jaRmipd April 15, 2026 at 11:37PM
Tuesday, April 14, 2026
Show HN: Uninum – All elementary functions from a single operator, in Python https://ift.tt/ZxCkcL9
Show HN: Uninum – All elementary functions from a single operator, in Python https://ift.tt/NFOTLda April 15, 2026 at 03:16AM
Show HN: Run Python tools on rust agents https://ift.tt/tZWP0DH
Show HN: Run Python tools on rust agents Over at Tools-rs, we wanted to script tools faster with the help of large communities. The interest arose to build a way to bridge our Rust LLM runtimes together with more traditional scripting languages, so we decided to find a way to bring Python tools into our ecosystem. Hence, we're introducing our first FFI on Python (powered by PyO3)! Calling a Python tool is as easy as writing a decorator in the Python function and then passing the script's (or folder) path to the tool collection builder. They get serialized as JSON objects so they're fully observable by the AI, and you can call them directly from Rust. https://ift.tt/ezVSZqm April 15, 2026 at 02:01AM
Show HN: Sk.illmd.com, a forum for talking about and showing off agent skills https://ift.tt/5DRFYfu
Show HN: Sk.illmd.com, a forum for talking about and showing off agent skills https://ift.tt/CqwSV7U April 15, 2026 at 01:07AM
Monday, April 13, 2026
Show HN: Encrypted, nothing stored, nothing repeated face-gated asset sharing https://ift.tt/GoC4KYD
Show HN: Encrypted, nothing stored, nothing repeated face-gated asset sharing https://veylt.net/ April 13, 2026 at 11:40PM
Show HN: pg_grpc – Call gRPC services directly from PostgreSQL https://ift.tt/1VuxwPo
Show HN: pg_grpc – Call gRPC services directly from PostgreSQL https://ift.tt/30qKloU April 13, 2026 at 11:20PM
Sunday, April 12, 2026
Show HN: Stork – MCP server so Claude/Cursor can search 14k MCP servers AI tools https://ift.tt/oIzTPpM
Show HN: Stork – MCP server so Claude/Cursor can search 14k MCP servers AI tools https://www.stork.ai April 13, 2026 at 01:19AM
Show HN: A social feed with no strangers https://ift.tt/QWhyGVM
Show HN: A social feed with no strangers Grateful is a gratitude app with a simple social layer. You write a short entry, keep it private or share it to a circle. A circle is a small private group of your own making — family, close friends, whoever you'd actually want to hear from. It shows you the most recent post first. People in the circle can react or leave a comment. There's also a daily notification that sends you something you were grateful for in the past. Try it out on both iOS and Android. Go to grateful.so https://ift.tt/LOgZpkn April 13, 2026 at 04:11AM
Show HN: Rekal – Long-term memory for LLMs in a single SQLite file https://ift.tt/yW7kj5x
Show HN: Rekal – Long-term memory for LLMs in a single SQLite file I got tired of repeating myself to my LLM every session. rekal is an MCP server that stores memories in SQLite and retrieves them with hybrid search (BM25 + vectors + recency decay). One file, local embeddings, no API keys. https://ift.tt/slyefGr April 13, 2026 at 02:55AM
Saturday, April 11, 2026
Show HN: Bitcoin and Quantum Computing – a three-part research series https://ift.tt/6QyTJq7
Show HN: Bitcoin and Quantum Computing – a three-part research series https://bitcoinquantum.space April 12, 2026 at 12:47AM
Show HN: A living Vancouver. Connor is walking dogs at the SPCA this morning https://ift.tt/Wd4JvE2
Show HN: A living Vancouver. Connor is walking dogs at the SPCA this morning I've spent most of my career in marketing, which for the last few years has meant building consumer personas for campaigns. I wanted to see if I could make these real, living in real neighborhoods, had real weather, real budgets, real Saturday lunches. I always wanted to build a world, not a segment. This is that. 140 people so far, split across Vancouver (100), San Francisco (20), and Tokyo (20). Each one is about 1,000 lines of profile — family, finances, daily schedule, health, worldview, media diet, the channels you'd actually reach them through and the ones that will explicitly never work on them. Demographics are census-grounded income, age, ethnicity, household composition follow normal distributions against StatsCan, ACS, and Japanese e-Stat data, so the panel is roughly representative of the city instead of representative of whatever's overrepresented in an LLM's training corpus. The specific details come from real stories. They live in real local time on a live map. Right now it's Saturday 11:32 AM in Vancouver. Connor Hughes, a 31-year-old software developer at Clio in Gastown, is on his SPCA volunteer shift, he walks shelter dogs at the Boundary Road location every other Saturday morning. Hassan Khoury is in the morning lunch rush with Tony at his Lebanese café — it's his busiest day of the week. Ahmad Noori is pulling Saturday overtime on a construction site. Jordan Whitehorse is on mid-shift at East Cafe on Hastings. Every day is unique, no two days repeat. A 3 AM job fetches live data: weather from Open-Meteo, grocery CPI from StatsCan food vectors, Metro Vancouver transit delays from Google Routes API against specific corridors, Vancouver gas prices, sunrise and sunset. Each persona has a modifier file that reacts to all of it. When Vancouver gas hits $1.85/L, Jaspreet the long-haul trucker's Coquihalla run to Calgary stops feeling worth it, his margins are thin, his mood takes a hit. When food CPI spikes, Gurinder at the Amazon warehouse stops buying the $9 Subway and brings roti from home. A health flare rolls probabilistically each morning which maybe nothing, maybe Tanya's six month old had a rough night, maybe Frank's back is acting up. The days stack up and get remembered. Every persona has a journal, today's entry in a markdown file, a week of them compressed into a "dream" of ~30 lines that keeps the shape without the texture, a month compressed into ~15 lines. It's their journal. I'm not writing it; the simulation is. Click any persona to open their detail, or hit "Talk to [name]" to have a conversation and they run on Claude Haiku with their full profile and recent diary entries as context. Not a product, not a startup, just a thing I've been quietly working on. They feel, in a way I didn't expect, like my fully grown kids. Happy to answer questions. https://brasilia-phi.vercel.app April 12, 2026 at 12:12AM
Show HN: We scanned uscis.gov for third-party trackers. The results are jarring https://ift.tt/FhMTZIQ
Show HN: We scanned uscis.gov for third-party trackers. The results are jarring https://ift.tt/g4FqmCp April 11, 2026 at 07:13PM
Friday, April 10, 2026
Show HN: Eve – Managed OpenClaw for work https://ift.tt/D0ipyWP
Show HN: Eve – Managed OpenClaw for work Eve is an AI agent harness that runs in an isolated Linux sandbox (2 vCPUs, 4GB RAM, 10GB disk) with a real filesystem, headless Chromium, code execution, and connectors to 1000+ services. You give it a task and it works in the background until it's done. I built this because I wanted OpenClaw without the self-hosting, pointed at actual day-to-day work. I’m thinking less personal assistant and more helpful colleague. Here’s a short demo video: https://ift.tt/qKz2AJc The main interface is a web app where you can watch work happen in real time (agents spawning, files being written, use of the CLI). There's also an iMessage integration so you can fire a task asynchronously, put your phone down, and get a reply when it's finished. Under the hood, there's an orchestrator (Claude Opus 4.6) that routes to the right domain-specific model for each subtask: browsing, coding, research, and media generation. For complex tasks it spins up parallel sub-agents that coordinate through the shared filesystem. They have persistent memory across sessions so context compounds over time. I’ve packaged it with a bunch of pre-installed skills so it can execute in a variety of job roles (sales, marketing, finance) at runtime. Here are a few things Eve has helped me with in the last couple days: - Edit this demo video with a voice over of Garry: https://www.youtube.com/watch?v=S4oD7H3cAQ0 - Do my tax returns - To build HN as if it was the year 2030: https://ift.tt/94RiUF3 AMA on the architecture and lmk your thoughts :) P.S. I've given every new user $100 worth of credits to try it. https://eve.new/login April 10, 2026 at 11:01PM
Show HN: FluidCAD – Parametric CAD with JavaScript https://ift.tt/nk9w8vT
Show HN: FluidCAD – Parametric CAD with JavaScript Hello HN users, This is a CAD by code project I have been working on on my free time for more than year now. I built it with 3 goals in mind: - It should be familiar to CAD designers who have used other programs. Same workflow, same terminology. - Reduce the mental effort required to create models as much as possible. This is achieved by: - Provide live rendering and visual guidance as you type. - Allow the user to reference existing edges/faces on the scene instead of having to calculate everything. - Provide interactive mouse helpers for features that are hard to write by code: Only 3 interactive modes for now: Edge trimming, Sketch region extrude, Bezier curve drawing. - Implicit coding whenever possible: e.g: There are sensible defaults for most parameters. The program will automatically fuse intersecting objects together so you do not have to worry about what object needs to be fused with what. - It should be reasonably fast: The scene objects are cached and only the updated objects are re-computed. I think I have achieved these goals to a good extent. The program is still in early stages and there are many features I want to add, rewrite but I think it is already usable for simple models. https://fluidcad.io/ April 11, 2026 at 12:09AM
Thursday, April 9, 2026
Show HN: Last Year I wrote a (Sci)fictional story where the EFF was a player [pdf] https://ift.tt/cfpUb0n
Show HN: Last Year I wrote a (Sci)fictional story where the EFF was a player [pdf] https://ift.tt/qEvI9pg April 9, 2026 at 11:43PM
Show HN: Logoshi, a brand kit generator for solo founders https://ift.tt/mSB3aDv
Show HN: Logoshi, a brand kit generator for solo founders https://logoshi.com/ April 9, 2026 at 10:12PM
Show HN: I built Dirac, Hash Anchored AST native coding agent, costs -64.8 pct https://ift.tt/6Q4GeUW
Show HN: I built Dirac, Hash Anchored AST native coding agent, costs -64.8 pct Fully open source, a hard fork of cline. Full evals on the github page that compares 7 agents (Cline, Kilo, Ohmypi, Opencode, Pimono, Roo, Dirac) on 8 medium complexity tasks. Each task, each diff and correctness + cost info on the github Dirac is 64.8% cheaper than the average of the other 6. https://ift.tt/nzZ7pXa April 9, 2026 at 05:36PM
Show HN: Homebutler – I manage my homelab from chat. AI never gets raw shell https://ift.tt/JyI6DT4
Show HN: Homebutler – I manage my homelab from chat. AI never gets raw shell https://homebutler.dev April 9, 2026 at 05:39PM
Show HN: CSS Studio. Design by hand, code by agent https://ift.tt/LlzXoE7
Show HN: CSS Studio. Design by hand, code by agent Hi HN! I've just released CSS Studio, a design tool that lives on your site, runs on your browser, sends updates to your existing AI agent, which edits any codebase. You can actually play around with the latest version directly on the site. Technically, the way this works is you view your site in dev mode and start editing it. In your agent, you can run /studio which then polls (or uses Claude Channels) an MCP server. Changes are streamed as JSON via the MCP, along with some viewport and URL information, and the skill has some instructions on how best to implement them. It contains a lot of the tools you'd expect from a visual editing tool, like text editing, styles and an animation timeline editor. https://cssstudio.ai April 9, 2026 at 04:53PM
Show HN: Moon simulator game, ray-casting https://ift.tt/Zzgnm2a
Show HN: Moon simulator game, ray-casting Did this a few years ago. Seems apropos. Sources and more here: https://ift.tt/pJcOWBw https://ift.tt/Y0rCkLz April 6, 2026 at 10:39PM
Wednesday, April 8, 2026
Show HN: Skrun – Deploy any agent skill as an API https://ift.tt/y6HZWs3
Show HN: Skrun – Deploy any agent skill as an API https://ift.tt/s9AKyF8 April 8, 2026 at 06:04PM
Show HN: 500k+ events/sec transformations for ClickHouse ingestion https://ift.tt/1cDuIan
Show HN: 500k+ events/sec transformations for ClickHouse ingestion Hi HN! We are Ashish and Armend, founders of GlassFlow. Over the last year, we worked with teams running high-throughput pipelines into self-hosted ClickHouse. Mostly for observability and real-time analytics. A question that came repeatedly was: What happens when throughput grows? Usually, things work fine at 10k events/sec, but we started seeing backpressure and errors at >100k. When the throughput per pipeline stops scaling, then adding more CPU/memory doesn’t help because often parts of the pipeline are not parallelized or are bottlenecked by state handling. At this point, engineers usually scale by adding more pipeline instances. That works but comes with some trade-offs: - You have to split the workload (e.g., multiple pipelines reading from the same source) - Transformation logic gets duplicated across pipelines - Stateful logic becomes harder to manage and keep consistent - Debugging and changes get more difficult because the data flow is fragmented Another challenge arises when working with high-cardinality keys like user IDs, session IDs, or request IDs, and when you need to handle longer time windows (24h or more). The state grows quickly and many systems rely on in-memory state, which makes it expensive and harder to recover from failures. We wanted to solve this problem and rebuild our approach at GlassFlow. Instead of scaling by adding more pipelines, we scale within a single pipeline by using replicas. Each replica consumes, processes, and writes independently, and the workload is distributed across them. In the benchmarks we’re sharing, this scales to 500k+ events/sec while still running stateful transformations and writing into ClickHouse. A few things we think are interesting: - Scaling is close to linear as you add replicas - Works with stateful transformations (not just stateless ingestion) - State is backed by a file-based KV store instead of relying purely on memory - The ClickHouse sink is optimized for batching to avoid small inserts - The product is built with Go Full write-up + benchmarks: https://ift.tt/ol5djf9... Repo: https://ift.tt/BCG9pDw Happy to answer questions about the design or trade-offs. https://ift.tt/BCG9pDw April 8, 2026 at 10:56PM
Tuesday, April 7, 2026
Show HN: Unicode Steganography https://ift.tt/KN4i0s2
Show HN: Unicode Steganography I built a demo of two Unicode steganography techniques, zero-width characters and homoglyph substitution, in the context of AI misalignment. The first is about the use of two invisible zero-width characters (ZWS and ZWNJ) to binary encode text. The second is much cooler. Most characters in the Latin and Cyrillic alphabets look nearly identical, but have different unicode. If you have text to encode and convert it into binary representation (1s and 0s), you could take plain english "carrier" text and for each 1 in the binary representation you could substitute the Cyrillic letter equivalent. Decoding the message requires traversing the text and seeing where Cyrillic letters could have been substituted but weren't, and where they were, leading to 0s and 1s respectively, which can be built back into your original hidden text. In both cases, these are detectable, but the interesting problem for me is whether an LLM could eventually invent an encoding that goes unnoticed by both us, and automated detection. If LLMs were able to covertly include messages in plaintext, misaligned AI Agents could eventually communicate across MCP/A2A and individual chat session boundaries undetected. A deceptive LLM might seem helpful, but work against your goals. It could tell other agents it interacts with over MCP/A2A to help it discreetly fail, signal intent, and avoid tripping oversight/safety mechanisms. Furthermore, oversight mechanisms become more difficult to implement if we can't believe our own eyes. https://ift.tt/d0hXRoK April 7, 2026 at 06:27PM
Show HN: Marimo pair – Reactive Python notebooks as environments for agents https://ift.tt/F8LdsG3
Show HN: Marimo pair – Reactive Python notebooks as environments for agents Hi HN! We're excited to share marimo pair [1] [2], a toolkit that drops AI agents into a running marimo notebook [3] session. This lets agents use marimo as working memory and a reactive Python runtime, while also making it easy for humans and agents to collaborate on computational research and data work. GitHub repo: https://ift.tt/zL9cIKD Demo: https://www.youtube.com/watch?v=6uaqtchDnoc marimo pair is implemented as an agent skill. Connect your agent of choice to a running notebook with: /marimo-pair pair with me on my_notebook.py The agent can do anything a human can do with marimo and more. For example, it can obtain feedback by running code in an ephemeral scratchpad (inspect variables, run code against the program state, read outputs). If it wants to persist state, the agent can add cells, delete them, and install packages (marimo records these actions in the associated notebook, which is just a Python file). The agent can even manipulate marimo's user interface — for fun, try asking your agent to greet you from within a pair session. The agent effects all actions by running Python code in the marimo kernel. Under the hood, the marimo pair skill explains how to discover and create marimo sessions, and how to control them using a semi-private interface we call code mode. Code mode lets models treat marimo as a REPL that extends their context windows, similar to recursive language models (RLMs). But unlike traditional REPLs, the marimo "REPL" incrementally builds a reproducible Python program, because marimo notebooks are dataflow graphs with well-defined execution semantics. As it uses code mode, the agent is kept on track by marimo's guardrails, which include the elimination of hidden state: run a cell and dependent cells are run automatically, delete a cell and its variables are scrubbed from memory. By giving models full control over a stateful reactive programming environment, rather than a collection of ephemeral scripts, marimo pair makes agents active participants in research and data work. In our early experimentation [4], we've found that marimo pair accelerates data exploration, makes it easy to steer agents while testing research hypotheses, and can serve as a backend for RLMs, yielding a notebook as an executable trace of how the model answered a query. We even use marimo pair to find and fix bugs in itself and marimo [5]. In these examples the notebook is not only a computational substrate but also a canvas for collaboration between humans and agents, and an executable, literate artifact comprised of prose, code, and visuals. marimo pair is early and experimental. We would love your thoughts. [1] https://ift.tt/zL9cIKD [2] https://ift.tt/JXcG5tK [3] https://ift.tt/kyulF0b [4] https://www.youtube.com/watch?v=VKvjPJeNRPk [5] https://ift.tt/JjdLoQC... https://ift.tt/zL9cIKD April 7, 2026 at 11:17PM
Show HN: C64 Ultimate Toolbox for macOS https://ift.tt/GnPqvRf
Show HN: C64 Ultimate Toolbox for macOS My wife got me a Commodore 64 Ultimate ( https://ift.tt/dC6pPVj ) for my birthday, and it became an obvious hassle to have to keep an entire monitor connected to it just to tinker with it. When I found out the Ultimate FPGA board has built-in support for streaming the video and audio data over the network, as well as a REST API allowing for file and configuration management, I set to work on an app to remotely control my new device. - View and hear your Commodore 64 Ultimate or Ultimate 64 device over the network, with a fully configurable CRT shader so you can dial in just the right retro feel. - View and manage files on your device, including support for drag and drop folder/file upload, as well as the ability to run and mount disks, create new disk images, and more. - BASIC Scratchpad is a mini-IDE in the app where you can write BASIC apps and send them directly to any of your connected devices to run. - Keyboard forwarding allows you to interact with your device with your computer keyboard, includes a keyboard overlay for Commodore specific keys your keyboard definitely doesn't have. - Visual memory viewer and editor, along with a terminal-like memory viewer and editor for debugging and tinkering. - Built-in support for recording videos and taking screenshots cleanly. - Fully native macOS AppKit app. Here's a rough and ready demo video I recorded and sent to App Review for the 2.0 release which was approved yesterday: https://www.youtube.com/watch?v=_2wJO2wOGm8 Please note again this app only works with Commodore 64 Ultimate or Gideon's Ultimate 64 devices. Ultimate II does not have the data streams feature to power the display. https://ift.tt/HfAn8Y0 April 7, 2026 at 10:09PM
Monday, April 6, 2026
Show HN: Meta-agent: self-improving agent harnesses from live traces https://ift.tt/EAjXO24
Show HN: Meta-agent: self-improving agent harnesses from live traces We built meta-agent: an open-source library that automatically and continuously improves agent harnesses from production traces. Point it at an existing agent, a stream of unlabeled production traces, and a small labeled holdout set. An LLM judge scores unlabeled production traces as they stream. A proposer reads failed traces and writes one targeted harness update at a time, such as changes to prompts, hooks, tools, or subagents. The update is kept only if it improves holdout accuracy. On tau-bench v3 airline, meta-agent improved holdout accuracy from 67% to 87%. We open-sourced meta-agent. It currently supports Claude Agent SDK, with more frameworks coming soon. Try it here: https://ift.tt/v8D0M3n https://ift.tt/v8D0M3n April 7, 2026 at 12:52AM
Show HN: ComputeLock – Insurance to reduce unpredictable compute spend https://ift.tt/UuNIZPb
Show HN: ComputeLock – Insurance to reduce unpredictable compute spend Reserved instances save money... until utilization changes, and you’re still paying. With ComputeLock, the risk of on-demand price spikes doesn’t exist - we offer burst insurance. 1. Send us an estimate of on-demand spend you expect and from what provider. 2. We confirm the maximum we'll cover for you for a small fee, and you get it in writing. 3. If on-demand prices spike, we'll reimburse you. We plan to work with smaller developers to start. How we do this is by monitoring supply and demand for compute. Of course, we'll get it wrong sometimes. But it's like insurance, you'll only need it when you NEED it. Would love to hear your feedback: https://ift.tt/ohfT3wK https://ift.tt/ohfT3wK April 6, 2026 at 10:53PM
Sunday, April 5, 2026
Show HN: I built a tool to show how much ARR you lose to FX fees https://ift.tt/8zQPyqr
Show HN: I built a tool to show how much ARR you lose to FX fees Hey HN, I started my career as a finance manager, transitioned into product management, and now I’m building my own products. Back in my finance days, while managing a £6M budget, I uncovered a £15k leak hiding in plain sight: FX fees. Today, I see solo founders making the exact same mistake. I realised most founders are quietly losing 2-5% of their revenue to what I call the Lazy Tax: - Stripe's ~2% auto-conversion fee on inbound revenue, - plus their local bank's ~3% spread when paying for global SaaS tools (AWS, Claude, Ads). So I built FixMyFX to show founders their exact leak and how to fix it (using multi-currency accounts to achieve a zero FX leak setup). Initially, I had Claude build this in React. Realised a simple calculator shouldn't need a 150kb payload and a complex build process. Threw the React code away and rebuilt it as a single lightweight HTML file using Alpine.js and Tailwind. It's completely free and ungated. I hope it helps you keep a bit more of your hard-earned revenue. Would love your feedback. Tania https://fixmyfx.com April 5, 2026 at 11:41PM
Show HN: A Dad Joke Website https://ift.tt/g8lTbwB
Show HN: A Dad Joke Website A dad joke website where you can rate random dad jokes, 1-5 groans. Sourced from 4 different places, all cited, all categorized, and ranked by top voted. Help me create the worlds best dadabase! https://joshkurz.net/ April 5, 2026 at 11:24PM
Saturday, April 4, 2026
Show HN: Vibooks – Local-first bookkeeping software built for AI agents https://ift.tt/MjUZCh2
Show HN: Vibooks – Local-first bookkeeping software built for AI agents https://vibooks.ai/ April 5, 2026 at 06:09AM
Show HN: Contrapunk – Real-time counterpoint harmony from guitar input, in Rust https://ift.tt/O2quUFj
Show HN: Contrapunk – Real-time counterpoint harmony from guitar input, in Rust https://contrapunk.com/ April 5, 2026 at 06:10AM
Show HN: Dev Personality Test https://ift.tt/ON9cqFX
Show HN: Dev Personality Test Was curious how a personality test would look for developers. So created this using FastAPI, HTMX, and AlpineJS. https://ift.tt/rPDzBQh April 5, 2026 at 02:59AM
Show HN: M. C. Escher spiral in WebGL inspired by 3Blue1Brown https://ift.tt/ulCJPcZ
Show HN: M. C. Escher spiral in WebGL inspired by 3Blue1Brown The latest 3Blue1Brown video [1] about the M. C. Escher print gallery effect inspired me to re-implement the effect as WebGL fragment shader on my own. [1]: https://www.youtube.com/watch?v=ldxFjLJ3rVY https://ift.tt/q5sa4mN April 5, 2026 at 01:13AM
Friday, April 3, 2026
Show HN: Ismcpdead.com – Live dashboard tracking MCP adoption and sentiment https://ift.tt/uOnvNpI
Show HN: Ismcpdead.com – Live dashboard tracking MCP adoption and sentiment Built this to track the ongoing debate around Model Context Protocol - whether it's gaining real traction or just hype. Pulls live data from GitHub, HN, Reddit and a few other sources. Curious what the HN crowd thinks given how active the MCP discussion has been here. https://ismcpdead.com April 4, 2026 at 12:58AM
Show HN: Community Curated Lists https://ift.tt/FShw5eG
Show HN: Community Curated Lists https://ift.tt/mIvfQMP April 4, 2026 at 12:02AM
Thursday, April 2, 2026
Show HN: A P2P messenger with dual network modes (Fast and Tor) https://ift.tt/6XTlKwE
Show HN: A P2P messenger with dual network modes (Fast and Tor) Hello HN, I have been working on a desktop P2P messenger called Kiyeovo for the last ~8 months, and I just published its beta version. Quick backstory: It started out as a CLI application for my Graduate Thesis, where I tried to make the most secure and private messenger application possible. Then, I transformed it into a desktop application, gave it "clearnet" support and added a bunch of features. Short summary: The app runs in 2 completely isolated modes: - fast mode: relay/DCUtR -> lower latency, calls support - anonymous mode: Tor message routing -> slower, anonymous These modes use different protocol IDs, DHT namespaces, pubsub topics and storage scopes so there’s no data crossover between them. Messaging works peer-to-peer when both parties are online, but falls back to DHT "offline buckets" when one of them is not. To ensure robustness, messages are ACK-ed and deleted after being read. Group chats use GossipSub for realtime messaging. Group messages are also saved to offline buckets in order for offline users to be able to read them upon logging in. Kick/Join/Leave events are also propagated using the DHT. Group metadata and all offline data is of course encrypted. Other features: Chats are E2E, file sharing is supported, 1:1 audio/video calls are supported (only in fast mode though, using WebRTC) Tradeoffs: Tor has noticeable latency, offline delivery is not immediately guaranteed, but rather "eventually consistent"; beta version does not have group calls yet. I’d appreciate feedback, that's why I posted this as a beta version Repo: https://ift.tt/MxmPhlU https://ift.tt/wJuenEO April 2, 2026 at 09:02PM
Show HN: RiceVM – A Dis virtual machine and Limbo compiler in Rust https://ift.tt/MWHyNXD
Show HN: RiceVM – A Dis virtual machine and Limbo compiler in Rust Hi, I've made a Dis virtual machine and Limbo programming language compiler (called RiceVM) in Rust. It can run Dis bytecode (for example, Inferno OS applications), compile Limbo programs, and includes a fairly complete runtime with garbage collection, concurrency features, and many of the standard modules from Inferno OS's original implementation. The project is still in an early stage, but if you're interested in learning more about RiceVM or trying it out, you can check out the links below: Project's GitHub repo: https://ift.tt/QqTGWJt RiceVM documentation: https://habedi.github.io/ricevm/ April 3, 2026 at 01:19AM
Show HN: Most products have no idea what their AI agents did yesterday https://ift.tt/moRyNMc
Show HN: Most products have no idea what their AI agents did yesterday We build collaboration SDKs at Velt (YC W22). Comments, presence, real-time editing (CRDT), recording, notifications. A pattern we keep seeing: products add AI agents that write, edit, and approve things. Human actions get logged. Agent actions don't. Same workflow, different accountability. We shipped Activity Logs to fix this. Same record for humans and AI agents. Immutable by default. Auto-captures collaboration events, plus createActivity() for your own. Curious how others are handling this. https://ift.tt/5D7GUhM April 2, 2026 at 11:55PM
Wednesday, April 1, 2026
Show HN: Roadie – An open-source KVM that lets AI control your phone https://ift.tt/xnsRqKm
Show HN: Roadie – An open-source KVM that lets AI control your phone Roadie is an open-source hardware KVM controlled via HTTP. HDMI capture in, USB keyboard/mouse/touch out, all from a browser. Hardware KVMs with web UIs have existed for years (PiKVM, TinyPilot, JetKVM, etc.). Roadie adds two things they don't generally have: multi-touch support (so it works with phones and tablets) and a focus on agent-driven use: any browser automation tool can drive the /view page directly, or connect to the WebSocket endpoint for lower-level programmatic control. ~$86 in parts, including two CircuitPython boards, an HDMI-to-USB dongle, and a Go server running on the host. No software needed on the target. https://ift.tt/ED5i6AF April 2, 2026 at 01:16AM
Show HN: Canon PIXMA G3010 macOS driver, reverse-engineered with Claude https://ift.tt/trMiNec
Show HN: Canon PIXMA G3010 macOS driver, reverse-engineered with Claude Canon doesn't provide a working macOS driver for the PIXMA G3010. I was stuck using Canon's iPhone app for all printing and scanning. I pointed Claude Code at a packet capture from the iPhone app and it reverse-engineered Canon's proprietary CHMP protocol, wrote a pure Rust eSCL-to-CHMP bridge daemon, and built a .pkg installer. My role was the physical parts: capturing packets, testing on the printer, confirming Image Capture worked. The protocol docs in docs/ are probably the first public documentation of Canon's CHMP protocol. https://ift.tt/eIWVA2D April 1, 2026 at 11:58PM
Show HN: Modern AI assisted goals and performance management https://ift.tt/Eva62Uc
Show HN: Modern AI assisted goals and performance management Hey hey I'm launching this on product hunt and I did a show many months back but prfrm is way better now prfrm - by ArchitectFWD, is a performance management platform. It is a platform for Teams, Startups & Organizations and also Individuals, Solo Founders & Families to organise and track goals, set plans and review periods to stay on top of development plans and set out the path for success. Typically uses for Review periods, performance plans, goals Also just added Team OKR The Goals AI assistant can create meaningful goals linked to the OKR or to individual goals and plan outcomes I included a journal to track progress The AI assistant can go over the journal for next steps, talking points for the next meeting or check in An a Kanban style schedule tracking --- I built prfrm by ArchitectFWD because I was tired of traditional performance management Spreadsheet.. blank cell ..what is next.. No more. I can set myself up, set a period, set the plan and outcome and use the AI assistant to help generate meaningful goals. I can track how I’m going and plot my path to success. With the addition of team OKR (objectives and key results) the goals can be mapped to team objectives as well, strengthening goals to real business goals https://prfrmhq.com goes to https://ift.tt/i04wAbn There's a silent video on the landing of how it works in mobile view If you want to comment on product hunt that's welcome too at https://ift.tt/qOifck0... Lastly, want to see video's? They're on https://www.youtube.com/playlist?list=PLBYzijBKDTJVrBzOlYuU0... https://ift.tt/i04wAbn April 2, 2026 at 12:34AM
Tuesday, March 31, 2026
Show HN: How This Graybeard Built the Fastest and Freest Postgres BM25 Search https://ift.tt/HtEFZM8
Show HN: How This Graybeard Built the Fastest and Freest Postgres BM25 Search Last summer we faced a conundrum at my company, Tiger Data, a Postgres cloud vendor whose main business is in timeseries data. We were trying to grow our business towards emerging AI-centric workloads and wanted to provide a state-of-the-art hybrid search stack in Postgres. We'd already built pgvectorscale in house with the goal of scaling semantic search beyond pgvector's main memory limitations. We just needed a scalable ranked keyword search solution too. The problem: core Postgres doesn't provide this; the leading Postgres BM25 extension, ParadeDB, is guarded behind AGPL; developing our own extension appeared daunting. We'd need a small team of sharp engineers and 6-12 months, I figured. And we'd probably still fall short of the performance of a mature system like Parade/Tantivy. Or would we? I'd be experimenting long enough with AI-boosted development at that point to realize that with the latest tools (Claude Code + Opus) and an experienced hand (I've been working in database systems internals for 25 years now), the old time estimates pretty much go out the window. I told our CTO I thought I could solo the project in one quarter. This raised some eyebrows. It did take a little more time than that (two quarters), and we got some real help from the community (amazing!) after open-sourcing the pre-release. But I'm thrilled/exhausted today to share that pg_textsearch v1.0 is freely available via open source (Postgres license), on Tiger Data cloud, and hopefully soon, a hyperscalar near you: https://ift.tt/1b5TGhO In the blog post accompanying the release, I overview the architecture and present benchmark results using MS-MARCO. To my surprise, we were not only able to meet Parade/Tantivy's query performance, but exceed it substantially, measuring a 4.7x advantage on query throughput at scale: https://ift.tt/8wTo60m... It's exciting (and, to be honest, a little unnerving) to see a field I've spent so much time toiling in change so quickly in ways that enable us to be more ambitious in our technical objectives. Technical moats are moats no longer. The benchmark scripts and methodology are available in the github repo. Happy to answer any questions in the thread. Thanks, TJ (tj@tigerdata.com) https://ift.tt/1b5TGhO March 31, 2026 at 09:59PM
Show HN: PhAIL – Real-robot benchmark for AI models https://ift.tt/RiBwNOM
Show HN: PhAIL – Real-robot benchmark for AI models I built this because I couldn't find honest numbers on how well VLA models [1] actually work on commercial tasks. I come from search ranking at Google where you measure everything, and in robotics nobody seemed to know. PhAIL runs four models (OpenPI/pi0.5, GR00T, ACT, SmolVLA) on bin-to-bin order picking – one of the most common warehouse operations. Same robot (Franka FR3), same objects, hundreds of blind runs. The operator doesn't know which model is running. Best model: 64 UPH. Human teleoperating the same robot: 330. Human by hand: 1,300+. Everything is public – every run with synced video and telemetry, the fine-tuning dataset, training scripts. The leaderboard is open for submissions. Happy to answer questions about methodology, the models, or what we observed. [1] Vision-Language-Action: https://ift.tt/YjLrA6W https://phail.ai March 31, 2026 at 09:55PM
Monday, March 30, 2026
Show HN: Rusdantic https://ift.tt/isnh3m9
Show HN: Rusdantic A unified, high-performance data validation and serialization framework for Rust, inspired by Pydantic's ergonomics and powered by Serde. https://ift.tt/8zx7v3s March 31, 2026 at 03:27AM
Show HN: AI Spotlight for Your Computer (natural language search for files) https://ift.tt/QxvVaEe
Show HN: AI Spotlight for Your Computer (natural language search for files) Hi HN, I built SEARCH WIZARD — a tool that lets you search your computer using natural language. Traditional file search only works if you remember the filename. But most of the time we remember things like: "the screenshot where I was in a meeting" "the PDF about transformers" "notes about machine learning" Smart Search indexes your files and lets you search by meaning instead of filename. Currently supports: - Images - Videos - Audio - Documents Example query: "old photo where a man is looking at a monitor" The system retrieves the correct file instantly. Everything runs locally except embeddings. I'm looking for feedback on: - indexing approaches - privacy concerns - features you'd want in a tool like this GitHub: https://ift.tt/9NS08Wm Demo: https://deepanmpc.github.io/SMART-SEARCH/ March 30, 2026 at 08:43PM
Show HN: Memv – Memory for AI Agents https://ift.tt/qDpjuKz
Show HN: Memv – Memory for AI Agents memv is an open-source Python library that gives AI agents persistent memory. Feed it conversations; it extracts knowledge. The extraction mechanism is predict-calibrate (Nemori paper): given existing knowledge, it predicts what a new conversation should contain, then extracts only what the prediction missed. v0.1.2 adds the production path: - PostgreSQL backend (pgvector for vectors, tsvector for text search, asyncpg pooling). Single db_url parameter — file path for SQLite, connection string for Postgres. - Embedding adapters: OpenAI, Voyage, Cohere, fastembed (local ONNX). Other things it does: - Bi-temporal validity: event time (when was the fact true) + transaction time (when did we learn it), following Graphiti's model. - Hybrid retrieval: vector similarity + BM25 merged with Reciprocal Rank Fusion. - Episode segmentation: groups messages before extraction. - Contradiction handling: new facts invalidate old ones, with full audit trail. Procedural memory (agents learning from past runs) is next, deferred until there's usage data. https://ift.tt/edTYhpv March 30, 2026 at 10:39PM
Show HN: I made my fitness dashboard public and Apple Health needs an API https://ift.tt/sGATgCB
Show HN: I made my fitness dashboard public and Apple Health needs an API https://ift.tt/fHt09hc March 30, 2026 at 11:09PM
Subscribe to:
Posts (Atom)
Show HN: tltv – Federation protocol for 24/7 TV channels https://ift.tt/KMVr6Ng
Show HN: tltv – Federation protocol for 24/7 TV channels I spent six years trying to build a tv channel server. rewrote it eight times. flas...
-
Show HN: A directory of 800 free APIs, no auth required Explore reliable free APIs for developers — ideal for web and software development, ...
-
Show HN: I built Dirac, Hash Anchored AST native coding agent, costs -64.8 pct Fully open source, a hard fork of cline. Full evals on the gi...
-
Show HN: I built a FOSS tool to run your Steam games in the Cloud I wanted to play my Steam games but my aging PC couldn’t keep up, so I bui...