Growing India News, world news, nation news, our news, people's news, grow news, entertainment, fashion, movies, tech, automobile and many more..
Thursday, August 29, 2024
Show HN: Warehouse OpenAI requests to your own database https://ift.tt/5IZzVGb
Show HN: Warehouse OpenAI requests to your own database Today we’re launching Velvet, an AI gateway for warehousing OpenAI and Anthropic requests to your PostgreSQL instance. We originally built an AI SQL editor, but realized that customers were using it to monitor their AI requests in production. We had already built an AI request warehousing tool internally to debug our SQL editor and gave some customers access. A few days into testing this idea, our pilot customer launched [1] and we began warehousing 1,500 requests per second. We worked closely with their engineering team in the following weeks, completely re-architecting Velvet for scale and additional features (such as Batch support). Along the way, other companies began seeking out Velvet to get visibility into their own LLM requests. We’re launching our AI gateway as a self-serve product today, but our pilot customers are already warehousing over 3 million requests per week - so the system is stable and performant. What makes Velvet unique is that you own the data in your own database. Also, we’re the first proxy that gives visibility into OpenAI batch calls - so you can observe and monitor async calls that save you money. Some technical notes: - Supports OpenAI and Anthropic endpoints - Data is formatted as JSON and logged to your own PostgreSQL instance (can add support for other databases for paying customers). - You can include queryable metadata in the header, such as user ID, org ID, model ID, and version ID. - Built on Cloudflare workers, which keeps latency minimal (using our caching feature will reduce latency overall) - Built for security + starting process of SOC II soon Why warehouse your requests? - Understand where money is spent. Use custom headers to calculate the cost per customer, model, or service. - Download real request/response data, so you can evaluate new models (e.g., re-running requests with a cheaper mini model) - Monitor time to completion of batch jobs. (e.g., OpenAI says 24 hours, but our customers average 3-4 hours) - Export a subset of example requests for fine-tuning It’s just a 2 line code change to get started. Try a sandbox demoing the logging proxy here: https://ift.tt/rS8qkoh More details in our docs https://ift.tt/LBmR1V3 [1] https://ift.tt/UI9TW5Z https://ift.tt/RXh4pIH August 28, 2024 at 10:21PM
Subscribe to:
Post Comments (Atom)
Show HN: Pocket2Linkding – Migrate from Mozilla Pocket to Linkding https://ift.tt/IwYJfju
Show HN: Pocket2Linkding – Migrate from Mozilla Pocket to Linkding With the Mozilla Pocket shutdown coming up in about two weeks, I thought ...
-
Show HN: An AI logo generator that can also generate SVG logos Hey everyone, I've spent the past 2 weeks building an AI logo generator, ...
-
Show HN: Snap Scope – Visualize Lens Focal Length Distribution from EXIF Data https://ift.tt/yrqHZtDShow HN: Snap Scope – Visualize Lens Focal Length Distribution from EXIF Data Hey HN, I built this tool because I wanted to understand which...
-
Show HN: Federated IndieAuth Server implemented as a notebook https://ift.tt/32IC633 April 27, 2021 at 04:37PM
No comments:
Post a Comment