Sunday, September 14, 2025

Show HN: Worried about your pet? Health assessments with instant answers https://ift.tt/HlYxBaK

Show HN: Worried about your pet? Health assessments with instant answers https://petcheckai.com September 15, 2025 at 01:36AM

Show HN: DriftDB – An experimental append-only database with time-travel queries https://ift.tt/sYr3QEM

Show HN: DriftDB – An experimental append-only database with time-travel queries https://ift.tt/uRBQNV3 September 15, 2025 at 01:12AM

Show HN: Update: Open-source private home security camera(end-to-end encryption) https://ift.tt/4lQ95sW

Show HN: Update: Open-source private home security camera(end-to-end encryption) Several months ago, I posted in Show HN ( https://ift.tt/rcuH574 ) about this project (previously named Privastead, now changed to Secluso). It's a privacy-preserving home security camera that uses OpenMLS for end-to-end encryption between the camera and the mobile app. The post received a good amount of attention and there were many good comments. Since then, my project cofounder and I have made major improvements to the project. The project previously would act as a hub for an IP camera, which was otherwise closed source. But now, our camera software can also run directly on a Raspberry Pi (even one as weak as a Raspberry Pi Zero 2W), resulting in a security camera with a fully open source software stack. In addition, our Raspberry Pi-based camera can perform AI to detect people/pets/vehicles and send notifications. Moreover, our released camera binary can be verified using reproducible builds and our app now runs on iOS as well as Android. You can use this project to turn your Raspberry Pi into a fully functional and (more important) private security camera. Please check it out, use it, and provide us with feedback! In addition, we built a prototype of a standalone home security camera using this open source project and a Raspberry Pi. Please check it out here ( https://secluso.com/ ). It's not meant to replace the open source project, but to explore whether a plug-and-play camera could make it easier for people who are interested but don't have time to set up our project on a Raspberry Pi. We're curious if this kind of device would be useful to the community. If you'd like updates on our progress on that front, you can join our mailing list on the site. Finally, we'd love to hear your feedback and ideas on how we can improve the project. And we always welcome contributions to our open source project. Our site: https://secluso.com https://ift.tt/vCAunZk September 15, 2025 at 01:14AM

Show HN: From selling AI to QA teams to building a smooth test-management app https://ift.tt/KORXQ89

Show HN: From selling AI to QA teams to building a smooth test-management app Hey HN, Bootstrapped founder here. I've got a bit of a story for you. We started desplega.ai to build a sophisticated AI platform that could automate E2E testing. We spent the last few months talking to dozens of QA leaders, and trying to learn what are their actual challenges. We've got one consistent feedback from large teams: their daily reality is a living hell of slow, clunky tools. We're talking about teams at major companies still managing tests on spreadsheets. Or they're stuck in a Jira instance so customized and slow you can “make a full pot of coffee” while a page loads (and that’s why they acquired arc! t3.gg said it first). On top of that, they're paying 2k/mo+!? for these tools that feel like they were designed in ‘05. Soon, it became obvious that our AI tool was way too advanced for them, and why it was much easier for younger startups to start using us. But we didn't want to give up on them just yet so... Because I grew up when the internet was still free, and I actually miss that a lot, we decided to create a free test management tool. Our vision is still AI, but we learned AI is not the silver bullet large teams are wishing for. (We wrote something about it at https://ift.tt/vL2kyWa ). Our hypothesis right now is that we can be that team building the right tools for each QA team, leveraging AI. We would love to hear your thoughts on (a) Should we make the project open-source? Any key features? (b) Would you ever trust an entity to do your QA first pass? Cheers, https://ift.tt/9qREcUB September 15, 2025 at 12:12AM

Saturday, September 13, 2025

Show HN: council - ai groupchat of ctos (no more asking ai to roleplay) https://ift.tt/rtvTkP4

Show HN: council - ai groupchat of ctos (no more asking ai to roleplay) i made a council of advisors to help me code. as a self taught dev, ive been heavily reliant on ai for the past two years. found myself often prompting claude to take on different personalities, so i built a web app. it's great for the step before telling cursor what to do and reviewing prs once theyre ready. PLEASE DON"T NUKE MY APY KEY. ty https://ift.tt/QoGI47E September 14, 2025 at 02:08AM

Show HN: I built an open source drag and drop editor for Genkit AI flows https://ift.tt/dvhl8zO

Show HN: I built an open source drag and drop editor for Genkit AI flows Hi, I have been building small AI Agents for quite some time now using various frameworks and one thing that always bugged me was that iterating on small things like prompts, flows, tools etc always took a code change + deployment of the app. While the prompt part can be solved with Langfuse I haven't found a good way to keep the flow management remote (and open source). Lately I have been working with Genkit ( https://ift.tt/yTvqepN ) and love how modular it is. So I thought why not build a UI builder on top of that that can handle simple flows, prompts and basic tracing. And here we are with a first early version: https://flowshapr.ai Repo: https://ift.tt/6F7Gpxd This first release can - Manage and execute simple flows remotely - Works with GoogleAI, Anthropic or OpenAi - Integrate with remote MCP tools - API Endpoint to execute flows remotely - Flows and flow urls are compatible with the genkit client sdk Upcoming - Support for Ollama - Support for various vector stores - More complex multi agent flows - Session management Any feedback and suggestions are welcome! September 14, 2025 at 12:24AM

Show HN: A store that generates products from anything you type in search https://ift.tt/smKpYrh

Show HN: A store that generates products from anything you type in search https://anycrap.shop/ September 13, 2025 at 05:32PM

Friday, September 12, 2025

Show HN: I got tired of Base64, so I made a numeric-only alternative https://ift.tt/81v2SdQ

Show HN: I got tired of Base64, so I made a numeric-only alternative I created numbase is an alternative to Base64 that encodes data into a single large number instead of ASCII characters. It's useful if you want to store or transmit data in numeric form and easily apply compression algorithms like Huffman. GitHub: https://ift.tt/hHSdOr1 September 12, 2025 at 10:42PM

Show HN: 47jobs – A Fiverr/Upwork for AI Agents https://ift.tt/92XRkIt

Show HN: 47jobs – A Fiverr/Upwork for AI Agents Hi HN, I’ve been working on something I’d love to share: 47jobs ( https://47jobs.com ) – a marketplace where you can hire AI agents to do tasks instead of human freelancers. Why? I kept noticing that many tasks on Upwork/Fiverr—coding, content generation, data analysis, automation—can now be handled by AI in minutes, not hours. But there wasn’t a platform built around hiring AI directly. So I built 47jobs: 100% AI agents doing the work (no humans in the loop). Jobs get delivered 10x faster, at transparent prices. You can “hire” an agent for coding, automation, research, etc. I’d love your thoughts: Does a pure AI-agent marketplace make sense? What types of jobs would you want AI agents to handle first? Any UX or trust issues you’d expect with this model? This is an early version, and I’m here to learn from your feedback. Thanks! https://47jobs.xyz September 13, 2025 at 01:29AM

Show HN: Lumro – AI agents for customer support, sales, and more https://ift.tt/Rlsx2kn

Show HN: Lumro – AI agents for customer support, sales, and more Hey HN, We just launched Lumro, a platform that lets you create AI agents that actually do things, not just chat. With Lumro you can: Handle customer support instantly, 24/7 Capture leads and qualify them Book demos or route tickets automatically The idea is to take repetitive work off human teams so they can focus on strategy and relationships. We launched yesterday and so far: 200+ people checked it out 15 signed up Our agent booked 1 demo Our agent captured 2 leads It’s early days, but we’re excited about the traction. Would love your feedback especially on what you’d want to see in an AI agent for your business. https://www.lumro.co/ September 12, 2025 at 09:46PM

Thursday, September 11, 2025

Show HN: Kafkatop, top-like CLI for Kafka https://ift.tt/QPJbTVg

Show HN: Kafkatop, top-like CLI for Kafka Hey HN, for those of you tired of running kafka-consumer-groups.sh and similar tools, here's a small real-time monitoring CLI tool for Apache Kafka, that displays consumer lag and event rates in a clean, top-like interface. You can quickly assess which consumers are lagging and when they will catch up. I've made this to quickly assess the health of remote on-premises clusters which most of the time lack proper monitoring. The tool can be found here: https://ift.tt/cX356g4 I'd be very interested to hear your feedback or any features you think would add value to this tool! https://ift.tt/cX356g4 September 11, 2025 at 11:33PM

Show HN: Real-time texture compression in Three.js https://ift.tt/z3rIot5

Show HN: Real-time texture compression in Three.js With the latest three.js update (r180) the use of the Spark GPU codecs is now straightforward and integration into existing gltf loaders requires just one line of code. This blog post outlining the few steps involved, goes over some of the surprises I encountered, and takes a close look at performance. The spark.js GitHub repository now includes three.js examples that are trivial to run, just: ``` npm install npm run dev ``` https://ift.tt/R9YbhXM September 11, 2025 at 11:50PM

Wednesday, September 10, 2025

Show HN: Haystack – Review pull requests like you wrote them yourself https://ift.tt/4kgI8sb

Show HN: Haystack – Review pull requests like you wrote them yourself Hi HN! We’re Akshay and Jake. We put together a tool called Haystack to make pull requests straightforward to read. What Haystack does: -- Builds a clear narrative. Changes in Haystack aren’t just arranged as unordered diffs. Instead, they unfold in a logical order, each paired with an explanation in plain, precise language -- Focuses attention where it counts. Routine plumbing and refactors are put into skimmable sections so you can spend your time on design and correctness -- Provides full cross-file context. Every new or changed function/variable is traced across the codebase, showing how it’s used beyond the immediate diff Here’s a quick demo: https://youtu.be/w5Lq5wBUS-I If you’d like to give it a spin, head over to haystackeditor.com/review! We set up some demo PRs that you should be able to understand and review even if you’ve never seen the repos before! We used to work at big companies, where reviewing non-trivial pull requests felt like reading a book with its pages out of order. We would jump and scroll between files, trying to piece together the author’s intent before we could even start reviewing. And, as authors, we would spend time to restructure our own commits just to make them readable. AI has made this even trickier. Today it’s not uncommon for a pull request to contain code the author doesn’t fully understand themselves! So, we built Haystack to help reviewers spend less time untangling code and more time giving meaningful feedback. We would love to hear about whether it gets the job done for you! How we got here: Haystack began as (yet another) VS Code fork where we experimented with visualizing code changes on a canvas. At first, it was a neat way to show how pieces of code worked together. But customers started laying out their entire codebase just to make sense of it. That’s when we realized the deeper problem: understanding a codebase is hard, and engineers need better ways to quickly understand unfamiliar code. As we kept building, another insight emerged: with AI woven into workflows, engineers don’t always need to master every corner of a codebase to ship features. But in code review, deep and continuous context still matters, especially to separate what’s important to review from plumbing and follow-on changes. So we pivoted. We took what we’d learned and worked closely with engineers to refine the idea. We started with simple code analysis (using language servers, tree-sitter, etc.) to show how changes relate. Then we added AI to explain and organize those changes and to trace how data moves through a pull request. Finally, we fused the two by empowering AI agents to use static analyses. Step by step, that became the Haystack we’re showing today. We’d love to hear your thoughts, feedback, or suggestions! https://ift.tt/cRG1udz September 10, 2025 at 11:51PM

Show HN: WorldView – Compare how different countries report the same news https://ift.tt/2EGTkX1

Show HN: WorldView – Compare how different countries report the same news https://worldview.up.railway.app/ September 10, 2025 at 11:47PM

Tuesday, September 9, 2025

Show HN: Real time visual saliency detection https://ift.tt/3wzaM2k

Show HN: Real time visual saliency detection I've just made public a library to perform real time visual saliency detection on videos (but static images are also supported). This started a couple months ago when, while working on another project, I ended up side-tracking and overkilling as usual. I'm pretty happy with the result and I think it could prove to be a useful piece of software. It should work on both Linux and macOS, but I'm yet to test Linux cause I only have a mac at hand. Windows may be doable through WSL. GitHub: https://ift.tt/tdPl4IJ Showcase: https://big-nacho.github.io/dosage-docs/showcase.html https://ift.tt/tdPl4IJ September 10, 2025 at 01:36AM

Show HN: An Open Source XR(AR/VR) Operating System https://ift.tt/Fys1YIE

Show HN: An Open Source XR(AR/VR) Operating System We're two college students building an XR(AR/VR) native Operating System with a custom kernel. We're also Open Source so feel free to check our GitHub Repository- https://ift.tt/ueK9TZs . The journey hasn't exactly been easy, we've been criticized by a lot saying that whatever we're doing is impractical and that we're too ambitious. Regardless, we've been committed to reach our goal. Here to answer all questions and doubts. Answering one question beforehand because we know someone is going to ask it - Q: Why use your own kernel/ Why don't you use Linux/ Why are you trying to reinvent the wheel? A: Using our own kernel helps us get rid of the baggage of legacy codes, bring the most optimal performance on our target hardware (XR/AR/VR) and achieve more efficiency than what we would've achieved on an existing kernel. We're not trying to reinvent the wheel, but just building Formula One racing tyres for it. https://ift.tt/aXHlW0S September 7, 2025 at 04:39PM

Show HN: Superagents – connect spreadsheets to any database, API or MCP server https://ift.tt/BNoMYms

Show HN: Superagents – connect spreadsheets to any database, API or MCP server Hi HN, I’m Eoin, founder of Sourcetable ( https://sourcetable.com ). Today, we’re launching Superagents. You can now connect your spreadsheet to any database, API or MCP server on the Internet. All of that data is available inside your spreadsheet, and you can use AI to analyze it and build models, reports and visualizations. The reason I started the company is because I spent 10 years at startups across engineering and operations roles and realized that Excel and Sheets weren't architected for the modern information environment. This creates a tremendous amount of nuisance and busywork cobbling together SaaS tools, reporting suites, and the misery of endless coordination meetings to make it all happen. (Boo meetings!) Spreadsheets aren’t just a business application: they’re the original thinking tool. The quality of these tools has a downstream impact on analytical thinking and creativity writ large, so this is a problem worth solving. Fast forward to today, we’re a 6 person team taking on Excel, Sheets and ChatGPT, so we’re excited to hear what you think! Who are Superagents for? Analysts, operators, and anyone doing data-centric work in spreadsheets. We see a tonne of finance people, of course, but also students, researchers and mom & pop shops. Sourcetable's superagents democratize data access and analysis, which is nice because our company’s mission is to make data accessible to everyone. Why “Superagents”? Because they can plan and orchestrate other task-specific agents to complete your work for you. We have a lot of different AI tools and agents inside Sourcetable, but there’s a whole lot more on the Agentic Web. Superagents are like the conductor that coordinates them all and calls on them when needed. Also, it’s a fun feature name (thanks, Alyssa!) If you remember the linked-data dream of the semantic web movement, that future is now: all of your business data is available and connected in Sourcetable. How does it work? Sourcetable is running a python virtual machine under the hood. Everything is sandboxed, and there are hundreds of AI tools and libraries our AI can access. Superagents are also doing code-gen on the fly to solve problems. The closest system we have found is Replit’s sandboxed operating systems. Beyond that Mixtral, ChatGPT and Anthropic offer some limited data connectivity features, except these AI chat services lack the storage, compute, and code execution that Sourcetable and Replit provide. This is all very new. How is this different to your previous data connectors, etc? We started out using ETL services to sync data and provide a GUI-driven PowerBI like experience in your spreadsheet. This was useful for people who knew SQL and how to write joins to combine fragmented data, but for everyone else (read: practically everyone), this solution just didn’t provide the frictionless, self-serve experience that we wanted. Our choices were to switch the GTM motion or change the product, so we shelved that reporting suite and focused on our AI spreadsheet and waited for models to catch up with our ambitions. Now that they have, we’re re-launching Sourcetable with our original goal in mind: building a spreadsheet-based operating system for the Agent Web, with fully networked data access for everyone on your team. AI is the great UX enabler. Caveats: * We heavily use Postgres, Google Analytics, Stripe and Google Search Console with Superagents. * We haven’t tested every endpoint on the Internet. We find that mainstream, well documented applications work best. * Yes, you can write data back to 3rd party applications and databases. We generally advise against this unless you understand the risks involved in giving AI write-access to your data. Bonus round: * All data connectors added during this launch week are FREE. (Regular AI messaging limits still apply.) Product Feedback? eoin@sourcetable.com https://ift.tt/TGw9L0E September 10, 2025 at 12:25AM

Monday, September 8, 2025

Show HN: C++ Compiler Support Page https://ift.tt/54uoZBk

Show HN: C++ Compiler Support Page Hi HN, I have created a webpage that displays all C++ features since C++20 in a simple, searchable table. It is intended to serve as a quick reference for C++ developers, whether as support for cross-platform development or simply to track the current support status out of curiosity. I created it as a simpler, more structured, and more up-to-date alternative to the cppreference compiler support site. Please note that the page intentionally does not list LWG and CWG papers. This might change as I am continually updating the site and trying out new ideas. Questions, feedback and suggestions are appreciated, either here or in the form of GitHub issues. https://cppstat.dev September 8, 2025 at 12:42PM

Show HN: I made a simple ASCII-art analog clock in Emacs https://ift.tt/7wfKHg3

Show HN: I made a simple ASCII-art analog clock in Emacs Just a toy, showing how easy it is to leverage built-in Emacs features (most notably Artist mode, which provides a set of functions for creating ASCII-art vector graphics) and things like trigonometric functions and timers to create something nice. A short blog post mentioning some background (and showing a screenshot): https://ift.tt/LZfhN8W . https://ift.tt/flizqFv September 8, 2025 at 11:48PM

Sunday, September 7, 2025

Show HN: Vizzly – Visual testing platform with built-in review workflows https://ift.tt/PRWud59

Show HN: Vizzly – Visual testing platform with built-in review workflows Hey HN! I’m Robert. I worked on [Percy’s SDKs/support from 2018–2022.]( https://ift.tt/MFfw14Y ) If you ever reached out to support or opened an issue, I probably helped you. Hopefully it was positive :) After a few years away, visual testing still felt stuck, so I'm building Vizzly. The problem: Design handoffs kinda suck. Designers make beautiful specs, devs implement them, then everyone realizes the details got lost somewhere. Current visual testing tools catch broken CSS but miss the real issue - making sure what ships actually matches what was designed, functionally (like really in the browser/OS/etc). What it is - Visual testing + review where you send actual screenshots (not DOM re-renders). Can be _any_ image to review (PDFS!) - Collaboration built-in: reviewer assignment, approvals, @mentions, screenshot-level threads. - Baselines: automatic (Git-aware), manual (not Git-based), or hybrid. - Team-based pricing; generous free plan for OSS; on-prem available. What’s different - Capture-first: use the pixels your app produced (no “but it doesn’t look like that on my machine/CI”). - Local TDD + CI parity: run locally with instant feedback; same flow in CI. - Custom properties to filter/slice reviews (component, viewport, theme, etc). Try it quickly (Playwright example) ``` npm i -D @vizzly-testing/cli export VIZZLY_TOKEN=your-token # in your tests: import { vizzlyScreenshot } from '@vizzly-testing/cli'; let img = await page.screenshot({ fullPage: true }); await vizzlyScreenshot('homepage-layout', img); ``` I would love feedback on everything! Rough edges you hit using the product/sdk, baseline expectations across branches, what you need for design/dev review to feel “done”, etc. Features like root cause analysis, an MCP, and more collab features are coming. But it's just me building :p I'm a big fan of OSS, so the OSS plan is pretty generous (10 seats + 10 review seats (20 total), unlimited public projects, 75GB, 6 concurrent builds). If it's not generous enough for teams, I'm willing to up it! This is my first time launching anything like this, I'm super keen on getting feedback and working any support or suggestions folks have. If anyone knew me from my support at Percy, I _really_ enjoy those conversations and opportunities to ship a fix or feature at the end of a chat. If Vizzly isn't it for your team, I wanna know why and what I can do to help you. Backstory + screenshots from my intro blog post: https://ift.tt/yY2N0Kl https://vizzly.dev September 7, 2025 at 09:14PM

Show HN: Kstack – Skill pack for monitoring/troubleshooting K8s in Claude Code https://ift.tt/GQauRgE

Show HN: Kstack – Skill pack for monitoring/troubleshooting K8s in Claude Code Hi All, Recently I've been using Claude Code a lot for de...