Skip to main contentOne of the most unique features of Nuabase is Row-Level Caching. Unlike standard HTTP caching or LLM-response caching which typically works on the entire request, Nuabase understands that your requests are often composed of multiple independent items.
How it works
When you use createArrayFn to process a list of items (e.g., rows in a database, log entries, user comments), Nuabase treats each item as a separate unit of work.
- You send a batch of 100 items.
- Nuabase checks its cache for each item individually, based on the prompt + schema + item content.
- If 90 items have been seen before (identical content), Nuabase serves them from the cache.
- Only the 10 new items are sent to the LLM.
- The results are merged and returned to you.
Benefits
1. Massive Cost Savings
You don’t pay for LLM tokens for data you’ve already processed. If you have a process that runs every hour on “all active users”, but only a few users have changed, you only pay for the changes.
2. Lower Latency
Cached hits are returned instantly (milliseconds) rather than waiting for the LLM (seconds).
3. Idempotency
You can safely retry failed jobs or re-run pipelines without worrying about double-spending on LLM costs.
Example Scenario
Run 1:
You send 3 leads: A, B, C.
- Nuabase processes A, B, C.
- Cost: 3 items.
Run 2:
You send 3 leads: A, B, D (where D is new).
- Nuabase sees A and B are in cache.
- Nuabase processes only D.
- Cost: 1 item.