Are there any plans to offer optional token caching for LLM calls?
Thanks
Are there any plans to offer optional token caching for LLM calls?
Thanks
Hi Will ![]()
Currently we have implemented token caching for the same prompt w/ temp = 0. We’re actively having discussions about offering optional token caching on a broader scale, among other features designed to enhance the usage of LLMs at Palantir. Cannot give any timelines, but be sure to stay tuned into Foundry Announcements for the latest offerings!
This topic was automatically closed 91 days after the last reply. New replies are no longer allowed.