Tag: small LLMs
18Apr
Compression-Aware Prompting: Getting the Best from Small LLMs
Learn how compression-aware prompting helps small LLMs perform like giants by distilling prompts, reducing token costs, and improving RAG efficiency.