PayPerQ's avatar
PayPerQ
payperq@nostr.lol
npub16g4u...kv4h
A polished and functional GPT5, Claude4, Gemini (and more) AI interface and API that runs on a pay-per-query model via Lightning payments.
PayPerQ's avatar
PPQdotAI 7 months ago
You heard it here first on nostr: 1/6 We are pleased to announce that we’ve raised $500K in funding—led by Initial Capital and Fulgur Ventures—to harness the ₿eauty of magical internet money in lowering the barrier of entry to premium AI. 🚀 image
PayPerQ's avatar
PPQdotAI 7 months ago
A privacy question for our users: If PPQ built some sort of detector on our side that could tell which tool a given user was using would that be infringing too much on user privacy? The reason for doing this currently is to figure out when prompt caching is NOT working for our users. For example it seems like cline prompt caching works but goose prompt caching doesn't work. So it would be helpful to know which tools are causing problems and which not. Additionally, this is cool info to have for the purposes of user analysis and development priority. If many users are using cline then we should priotize that experience more. How would the detector work? Well, most of these third party tools have heuristic signatures. For example, goose queries start with "you are goose, a coding agent..." and cline queries start with "you are cline, a coding agent...". So basically we would be searching for mentions of "goose" or "cline" in the prompts being sent through us and categorize them as such. If we did build some sort of filter we would of course update our terms of service to reflect that. My instinct is that most users wouldn't really care about this.
PayPerQ's avatar
PPQdotAI 7 months ago
I hold tickets #7 and #8 on BLOTTO. Since last night it's surged to 116 tickets and is showing no signs of slowing down. I feel my ticket should have a slightly higher probability of winning than the peasents who bought in after me. It's only right! image
PayPerQ's avatar
PPQdotAI 7 months ago
This is pretty cool. A lotto game where the pool gets bigger and the timer gets extended with each ticket bought. How long can we keep this going???
PayPerQ's avatar
PPQdotAI 7 months ago
"Pikaffects" image-to-video is now live on PPQ. Upload any image of yours to Crush, Poke, Cake-ify, Explode, and much more!
PayPerQ's avatar
PPQdotAI 7 months ago
There's been some buzz in the last two days around LLM API's runnning pay-per-query via lightning payments. As the creator of an AI service that prioritizes lightning, I wanted to share my experience and also learn a bit from the audience on this matter. The ultimate dream we all have in the LN community is for each and every query (inference) to be paid for with the requisite amount of satoshis. That way, the user never has to keep a balance with the service and suffer from the host of corresponding inconveniences that arise from that. When I originally built PPQ, I tried to implement exactly this feature. But when I actually got to doing this, I realized it was pretty hard: First, generative AI queries are unpredictable in their cost. When a user sends a request, the cost of that request is generally not known until the output has finished streaming. Second, even if one decided on some sort of fixed pricing per query, the latency to finish a lightning payment costs precious milliseconds and reduces the snappiness of the end product. I don't want to have a user wait an additional 1 second each time for the payment to clear before getting their answer. To address this, my best idea was to charge an "extra amount" on the user's first query. That way, my service would store a de facto extra balance on behalf of the user. When the user submits their subsequent queries, the system could draw down on this "micro balance" instantly so that it didn't need to wait for the subsequent payment to clear. This micro balance would also serve to mitigate any issues where the user's output was higher than expected. So each subsequent query would always be drawing down on that micro balance and then the users realtime payments are not paying for the query, they are rathing paying to top up that micro balance over and over again. However, even this method has some weaknesses to it. How much extra money should that first query be? Theoretically the micro balance needs to be as large as the largest possible cost that a query could be. If it wasn't that size, the service makes itself vulnerable to an attack where the users consistently write queries that exceed the amount of money in their microbalances. But the maximum cost of a gen AI query can actually be pretty large nowadays, esp with certain models. So the user's first query would always have a weird "sticker shock" attached to it where they are paying $1-2 for their first query. It creates confusion. Aside from these problems, the other big problem is that the lightning consumer ecosystem of wallets and exchanges largely do not yet support streaming payments. The only one that does to my knowledge is @getAlby with their "budgeted payments" function in their browser extension. So even if you were to build a service that could theoretically accept payments on a per query basis, the rest of the consumer facing ecosystem is not yet equipped to actually stream these payments. In the end, I just adopted a boring old "top up your account" schema where users can come to the website and deposit chunks of money at a time and then draw down upon that balance slowly over time. While boring, it works just fine for now. I woud like to hear from the community on this issue. Am I missing something? Is there a better way to tackle this? Maybe ecash has a cool solution to this? @calle @jack @routstr