You are charging per request ($0.05) what amounts to about 1M tokens on GPT4-mini ($0.075); i.e. this is 10-100x more expensive than the cost of running the LLM. I don't think it's worth it at that price point.
Edit (just found out this):
"The $0.05 per request fee covers requests up to 30,000 tokens.
For requests exceeding this limit, additional nominal charges may apply."
$0.05/request is a staggering amount. Their demo to order a burger is funny considering that a long chat session could easily end up costing more than a burger.
Good feedback, I agree that our pay-as-you-go pricing may not fit everyone's budget. We're working on reducing our costs and simplifying our pricing. Goal is to get this much, much lower in the coming months. There's some complexity to how we handle pricing, since the core of what we do is based off of ingesting the requests and responses, and performing some processes on top of that info. Some people have small prompts (<1000 tokens), others have massive prompts (80k+). We found that the vast majority are well under 30k, though.
You are charging per request ($0.05) what amounts to about 1M tokens on GPT4-mini ($0.075); i.e. this is 10-100x more expensive than the cost of running the LLM. I don't think it's worth it at that price point.
Edit (just found out this):
"The $0.05 per request fee covers requests up to 30,000 tokens.
For requests exceeding this limit, additional nominal charges may apply."
Hmm ... it's just too much.