Function Calling Deep Dive: Building LLM-Powered Tools and Agents

Introduction: Function calling transforms LLMs from text generators into action-taking agents. Instead of just describing what to do, the model can actually do it—query databases, call APIs, execute code, and interact with external systems. OpenAI’s function calling (now called “tools”) and similar features from Anthropic and others let you define available functions, and the model […]

Read more →

LlamaIndex: The Data Framework for Building Production RAG Applications

Introduction: LlamaIndex (formerly GPT Index) is the leading data framework for building LLM applications over your private data. While LangChain focuses on chains and agents, LlamaIndex specializes in data ingestion, indexing, and retrieval—the core components of Retrieval Augmented Generation (RAG). With over 160 data connectors through LlamaHub, sophisticated indexing strategies, and production-ready query engines, LlamaIndex […]

Read more →

Quantization Methods for LLMs: GPTQ, AWQ, and BitsAndBytes

Last year, I needed to run a 13B parameter model on a 16GB GPU. Full precision required 52GB. After testing GPTQ, AWQ, and BitsAndBytes, I reduced memory to 7GB with minimal accuracy loss. After quantizing 30+ models, I’ve learned which method works best for each scenario. Here’s the complete guide to LLM quantization. Figure 1: […]

Read more →

HL7 v3: Understanding RIM and Why v3 Failed to Replace v2

Executive Summary HL7 v3 was designed in the 1990s as the successor to HL7 v2, promising a rigorous, model-driven approach based on the Reference Information Model (RIM). Despite 20+ years of development and standardization, v3 never achieved widespread adoption. Understanding why v3 failed—and where it still matters—is crucial for architects navigating healthcare interoperability standards. 🏥 […]

Read more →

Advanced RAG Patterns: From Naive Retrieval to Production-Grade Systems (Part 1 of 2)

Introduction: Retrieval-Augmented Generation (RAG) has become the go-to architecture for building LLM applications that need access to private or current information. By retrieving relevant documents and including them in the prompt, RAG grounds LLM responses in factual content, reducing hallucinations and enabling knowledge that wasn’t in the training data. But naive RAG implementations often disappoint—the […]

Read more →

AKS pod managed identity

Kubernetes has become one of the most popular container orchestration tools, and Azure Kubernetes Service (AKS) is a managed Kubernetes service provided by Microsoft Azure. With the increasing use of Kubernetes and AKS, there is a growing need to improve the security and management of access to cloud resources. AKS pod managed identity is a […]

Read more →