Prompt Compression Techniques: Fitting More Context in Less Tokens

Introduction: Context windows are limited and tokens are expensive. Long prompts with extensive context, examples, or retrieved documents quickly hit limits and drive up costs. Prompt compression techniques reduce token count while preserving the information LLMs need to generate quality responses. This guide covers practical compression strategies: token pruning to remove low-information tokens, extractive summarization […]

Read more β†’

Azure Cosmos DB–Multi Master

During the Ignite 2018, Microsoft has announced the general availability of Multi-Master feature being introduced to Azure Cosmos DB to provide more control into data redundancy and elastic scalability for your data from different regions with multiple writes and read instances. What is Multi-Master essentially? Multi-master is a capability that provided as part of Cosmos […]

Read more β†’

Document Chunking Strategies: Optimizing RAG Retrieval Quality

Introduction: RAG systems live or die by their chunking strategy. Chunk too large and you waste context window space with irrelevant content. Chunk too small and you lose semantic coherence, making it hard for the LLM to understand context. The right chunking strategy depends on your document types, query patterns, and retrieval approach. This guide […]

Read more β†’

Getting Started with Azure CLI 2.0

Older days we used to manage azure resources through AzureRM PowerShell modules . This was very much flexible for any Azure Administrator or Developers to run Automated Deployments to Azure Resource Manager resources. Azure CLIΒ  is the next improved version with simplified cmdlets to make life easier and it is cross-platform. You can use Azure […]

Read more β†’