This is a autopost bolg frinds we are trying to all latest sports,news,all new update provide for you
Wednesday, August 28, 2024
Show HN: Warehouse OpenAI requests to your own database https://ift.tt/Q7xMKyk
Show HN: Warehouse OpenAI requests to your own database Today we’re launching Velvet, an AI gateway for warehousing OpenAI and Anthropic requests to your PostgreSQL instance. We originally built an AI SQL editor, but realized that customers were using it to monitor their AI requests in production. We had already built an AI request warehousing tool internally to debug our SQL editor and gave some customers access. A few days into testing this idea, our pilot customer launched [1] and we began warehousing 1,500 requests per second. We worked closely with their engineering team in the following weeks, completely re-architecting Velvet for scale and additional features (such as Batch support). Along the way, other companies began seeking out Velvet to get visibility into their own LLM requests. We’re launching our AI gateway as a self-serve product today, but our pilot customers are already warehousing over 3 million requests per week - so the system is stable and performant. What makes Velvet unique is that you own the data in your own database. Also, we’re the first proxy that gives visibility into OpenAI batch calls - so you can observe and monitor async calls that save you money. Some technical notes: - Supports OpenAI and Anthropic endpoints - Data is formatted as JSON and logged to your own PostgreSQL instance (can add support for other databases for paying customers). - You can include queryable metadata in the header, such as user ID, org ID, model ID, and version ID. - Built on Cloudflare workers, which keeps latency minimal (using our caching feature will reduce latency overall) - Built for security + starting process of SOC II soon Why warehouse your requests? - Understand where money is spent. Use custom headers to calculate the cost per customer, model, or service. - Download real request/response data, so you can evaluate new models (e.g., re-running requests with a cheaper mini model) - Monitor time to completion of batch jobs. (e.g., OpenAI says 24 hours, but our customers average 3-4 hours) - Export a subset of example requests for fine-tuning It’s just a 2 line code change to get started. Try a sandbox demoing the logging proxy here: https://ift.tt/voPtWJD More details in our docs https://ift.tt/bOnFAZh [1] https://ift.tt/YqDTEnC https://ift.tt/gf49epa August 28, 2024 at 10:21PM
Subscribe to:
Post Comments (Atom)
Show HN: Tablr – Supabase with AI Features https://ift.tt/ltABMro
Show HN: Tablr – Supabase with AI Features https://www.tablr.dev/ June 30, 2025 at 04:35AM
-
Show HN: Locksmith – detect locks taken by Postgres migrations https://ift.tt/0cBueJt February 10, 2025 at 02:26AM
-
Show HN: I built a FOSS tool to run your Steam games in the Cloud I wanted to play my Steam games but my aging PC couldn’t keep up, so I bui...
-
Show HN: TNX API – Natural Language Interactions with Your Database Hey HN! I built TNX API to make working with databases as simple as aski...
No comments:
Post a Comment