Token Optimisation with LLMs

I understand prompt caching is an effective way for optimising for input tokens. What are other options for optimising input tokens? My input tokens maximise in using an xml file.

I have avoided using TOONs [1] because it works only with uniform arrays, not nested objects nor non-uniform structures.

[1]https://github.com/toon-format/toon

1 points | by xoptions 4 hours ago

0 comments