• Skip to primary navigation
  • Skip to main content
  • Skip to primary sidebar

Meet Ashwin

Helping Engineers To Become Tech Leaders

  • Blog
  • Newsletter
  • Courses
    • Generative AI 101
    • LLMOps 101
    • Data Platforms Made Easy
  • Products
    • Generative AI Ecosystem Mindmap
    • 8 Mistakes to avoid in Tech Leadership (e-book)
  • Resources
  • Contact

Ashwin

Agent Evals 101 : Building a Strong Evaluation Framework for AI Agents

February 18, 2026 by Ashwin Leave a Comment

AI Agents are taking enterprises by storm. Organizations are racing to build autonomous systems that reason, plan, and execute complex tasks—whether solving real problems or keeping pace with competitors.

The promise is compelling: productivity gains and capabilities that seemed impossible months ago. But beneath the excitement lies a critical challenge: How do you actually evaluate and test AI agents?

This isn’t an engineering afterthought. It’s the difference between a reliable system and an unpredictable black box that works brilliantly one moment and fails catastrophically the next. Unlike traditional software, AI agents are probabilistic and context-dependent—our old testing playbooks don’t apply.

Yet evaluation remains the unattractive, ignored piece of the puzzle. Teams prototype excitedly while pushing “How do we know this works?” to later sprints—or ignoring it entirely.

This is a mistake organizations can’t afford to make.

How are Agent Evals different from Traditional Software Testing?

Here are 5 key differences between AI Agent testing and traditional software testing:

  • Non-deterministic outputs: Traditional software produces the same output for the same input every time. AI agents can generate different responses to identical prompts, making reproducible test cases nearly impossible and requiring statistical evaluation across multiple runs instead of single pass/fail checks.
  • Emergent behavior and edge cases: While traditional software fails in predictable ways based on code paths, AI agents can exhibit unexpected behaviors in novel situations they weren’t explicitly trained for. The failure modes are often subtle, contextual, and impossible to enumerate in advance.
  • Multi-step reasoning and chain-of-thought: Traditional tests verify individual functions or API calls. AI agents execute complex, multi-step workflows where each decision influences the next, requiring evaluation of the entire reasoning chain—not just the final output—to understand where and why things went wrong.
  • Subjective quality metrics: Traditional software testing relies on objective criteria like “does this API return a 200 status code?” AI agent outputs often require human judgment to evaluate quality, relevance, tone, and appropriateness—metrics that are inherently subjective and context-dependent.
  • Tool use and external interactions: Traditional software testing typically mocks external dependencies for isolation. AI agents dynamically choose which tools to use, when to use them, and how to chain them together, making it critical to test not just if they can use tools correctly, but whether they make good decisions about when and how to use them in the first place.

Building Blocks of Agent Evals

An Agent Evaluation system typically consists of these core building blocks:

Agents – The system under test that works toward specific objectives using tools and LLMs. While production systems often involve multiple agents, we treat them as a single entity for evaluation purposes.

Evaluation Suite – A comprehensive set of tasks or test cases that assess agents across various parameters and metrics. This is your complete testing framework.

Tasks – Individual test cases within the evaluation suite, each targeting specific agent capabilities or goals. Common metrics include task success rate, accuracy, latency, and cost per task.

Outcomes – The measurable outputs produced by agents during task execution. These represent what the agent actually accomplished and form the basis for evaluation—analogous to system test results in traditional software.

Graders – Automated or human evaluators that score outcomes against predefined rubrics, ground truth data, or reference outputs. Graders determine whether the agent succeeded and how well it performed.

While your evaluation system may include additional components depending on your specific testing needs, these building blocks provide a solid foundation for assessing any agentic system.

Eval-Driven Development (EDD) for Agentic Systems

Inspired by Test-Driven Development (TDD), which has improved software engineering for decades, Eval-Driven Development (EDD) places systematic evaluations at the core of the AI development lifecycle—not as an afterthought, but as the foundation.

Here are 4 core principles for a robust EDD approach:

1. Define success upfront – Establish clear, quantifiable metrics aligned with business goals before writing a single line of code. Define what “good” looks like: accuracy thresholds, acceptable latency, hallucination rates, task completion criteria.

2. Build comprehensive eval suites – Create test datasets covering common scenarios, edge cases, and adversarial examples. Your evals should stress-test your agents the way real-world usage will.

3. Evaluate continuously – Run your evaluation suite with every prompt change, model upgrade, or system modification. Treat evals like CI/CD—continuous measurement is the only way to catch regressions early and validate improvements.

4. Use multiple evaluation methods – Combine human grading for nuanced judgment, code-based assertions for objective metrics, and LLM-as-judge for scalable quality assessment. No single evaluation method captures the full picture.


Guided by these core principles, here’s a practical workflow for designing an evaluation system for your agentic application:

  • Understand the business problem – Start with clarity on what success looks like for your stakeholders. What specific tasks must the agent accomplish? What quality bar must it meet? What failures are unacceptable?
  • Assemble representative examples – Gather real data related to the problem: actual user queries, historical cases, common workflows, and known failure modes. Your eval dataset should mirror production reality.
  • Build an end-to-end V0 system – Create a working prototype quickly, even if crude or incomplete. You need something functional to evaluate before you can improve it.
  • Label data and build initial evals – Create ground truth labels for your examples and write your first evaluation metrics. Start simple: basic success/failure criteria and a handful of quality measures.
  • Align evals to business metrics – This is critical. Map your technical metrics (accuracy, latency, tool usage) to actual business outcomes (customer satisfaction, cost savings, error reduction). If your evals don’t predict business value, they’re measuring the wrong things.
  • Iterate on both system and evals – Improve your agent based on eval results, but also refine your evals as you discover what matters. Your evaluation framework should evolve alongside your system.
  • Integrate into your development workflow – Embed evals into CI/CD, code reviews, and release processes. Make evaluation a continuous practice, not a pre-launch gate.

This creates a virtuous flywheel: better evals surface real problems, fixes improve performance, and improved performance validates your eval framework.

Agent Eval Frameworks in Practice

The ecosystem of agent evaluation tools is rapidly maturing. Here are four leading frameworks that can accelerate your evaluation efforts:

1. LangSmith (by LangChain)

LangSmith provides end-to-end observability and evaluation for LLM applications with excellent tracing of multi-step agent workflows. Its strength lies in visualizing complex agent reasoning chains and tight integration with the LangChain ecosystem, supporting human-in-the-loop evaluation, automated scoring, and A/B testing.

2. Braintrust

Braintrust is an enterprise-grade evaluation platform focused on production AI systems with strong CI/CD integration and powerful analytics. It excels at experiment tracking and comparing different agent architectures side-by-side, supporting code-based assertions, LLM-as-judge, and human review workflows.

3. Promptfoo

Promptfoo is an open-source, configuration-driven framework that emphasizes simplicity and flexibility for teams wanting full control. Its lightweight, local-first design and YAML-based test definitions make it ideal for rapid iteration and easy integration into CI/CD pipelines.

Key Takeaways

  • Evaluation is non-negotiable. As AI agents become embedded in enterprise systems, a robust evaluation framework isn’t optional—it’s the foundation for reliability, trust, and measurable business impact.
  • Adopt Eval-Driven Development. Just as TDD transformed software engineering, EDD must become your default approach for AI development. Build evaluations first, then build agents that pass them.
  • Connect evals to business outcomes. Technical metrics matter only if they predict business value. Continuously align your evaluation framework with what stakeholders actually care about, and refine it as you gather more production data.
  • Leverage the ecosystem. The evaluation tooling landscape is evolving rapidly. Use open-source and commercial frameworks to accelerate your evaluation maturity rather than building everything from scratch.
  • The organizations that master agent evaluation today will be the ones deploying reliable, high-value AI systems tomorrow. Start building your evaluation practice now.

Disclaimer: The ideas, opinions, and recommendations expressed in this blog post are solely my own and do not reflect the views, policies, or positions of my employer.

Filed Under: Uncategorized

Enterprise AI Agents: Why Governance is Your Competitive Advantage

December 13, 2025 by Ashwin Leave a Comment

Building AI agents in an enterprise extends far beyond prototypes and POCs. When designing agents for real users who drive revenue, governance becomes critical.

This post explores the essential governance pillars for integrating AI agents into enterprise production applications. 

In the world of Agent Management Platforms, four governance pillars strengthen the successful rollout of AI agents in your organization.

Short on time? Check the TL;DR

Pillars of AI Agents Governance

AI Agents Governance Pillars
  • Evaluation – How do you evaluate the correctness and completeness of agent outputs
  • Security – How do you protect agents against malicious attacks and breaches
  • Guardrails – How do you set boundaries and checkpoints, often tied to your business and compliance rules
  • Auditability – How do you understand everything that has gone on before an agent produced the outputs

For this article, I will pick examples of each pillar from the open-source CrewAI Agentic framework.  But you can essentially implement it with other similar frameworks like LangGraph, Mastra, etc.


AI Agents Evaluation

Evaluation in AI agent systems serves a similar purpose to testing in traditional software development—validating that the system performs as intended. 

However, AI agents built on large language models are non-deterministic systems. Unlike conventional software, where the same input consistently produces the same output, AI agents may generate different responses across multiple runs. This variability requires specialized evaluation approaches.

Understanding Non-Deterministic Behavior

The non-deterministic nature of LLM-based agents stems from temperature settings, sampling methods, and the probabilistic nature of language generation. 

While this enables creative problem-solving and natural interactions, it means traditional unit tests with exact output matching are insufficient. Instead, evaluation must assess whether outputs meet quality standards and functional requirements rather than matching predetermined values.

Two Complementary Evaluation Approaches

Effective AI agent evaluation employs two methodologies: subjective and objective evaluation.

Subjective Evaluation: LLM-as-Judge

Subjective evaluation leverages other LLMs to review, critique, and score agent outputs. This approach assesses qualities difficult to measure programmatically, such as relevance, coherence, helpfulness, and tone. 

An LLM judge evaluates whether an agent’s response appropriately addresses the user’s intent, maintains a consistent persona, or demonstrates sound reasoning.

The judge receives the original task, the agent’s output, and specific evaluation criteria. It then generates a critique and assigns scores, revealing subtle issues like logical inconsistencies or inappropriate confidence levels that traditional metrics might miss.

Objective Evaluation: Gold Standard Comparison

Objective evaluation relies on gold standard datasets—curated collections of inputs paired with known correct outputs. The process compares agent outputs against these benchmarks using quantifiable metrics such as accuracy, precision, recall, or task-specific success criteria.

For example, if an agent extracts structured information from documents, objective evaluation measures how accurately it identifies required fields compared to human-annotated examples. 

For multi-step workflows, evaluation verifies that the agent completes all necessary steps in the correct sequence.

Implementation in Practice

Modern agent frameworks like CrewAI provide built-in evaluation capabilities combining both approaches. 

A typical workflow includes running the agent multiple times with the same test cases, applying LLM judges for subjective quality assessment, comparing outputs against gold standards, and aggregating results to identify patterns and issues.

Evaluation results inform iterative improvement. Teams can adjust agent instructions, refine prompts, modify tool configurations, or retrain components based on performance gaps. This continuous evaluation and refinement process is essential for maintaining reliable AI agent systems in production.


AI Agents Security

AI agents face unique security challenges that differ from traditional software vulnerabilities. As agents interact with users, access tools, and process data autonomously, they become targets for malicious attacks designed to manipulate their behavior or extract sensitive information.

Common Security Threats

Prompt Injection occurs when attackers craft inputs that override an agent’s original instructions, causing it to ignore safety guidelines or perform unintended actions. These attacks exploit the agent’s natural language interface to introduce malicious directives disguised as legitimate requests.

Tool Misuse happens when agents are manipulated into using their capabilities inappropriately—such as accessing unauthorized data, executing harmful commands, or making API calls that violate business rules. Since agents have direct access to tools and systems, compromised decision-making can have immediate consequences.

Data Poisoning involves corrupting the training data or knowledge bases that agents rely on, causing them to learn incorrect patterns or biased behaviors. Attackers may inject false information into retrieval systems or databases that agents query.

Context Poisoning targets the agent’s working memory by inserting misleading information into conversation history or retrieved documents. This can cause the agent to make decisions based on false premises or manipulated context.

Identity Spoofing exploits weak authentication mechanisms, allowing attackers to impersonate legitimate users or systems. Agents may then grant unauthorized access or perform actions on behalf of fake identities.

Security Countermeasures

Protecting AI agents requires multiple defensive layers. Input validation and sanitization filters suspicious patterns before they reach the agent’s core reasoning. Output monitoring detects when responses deviate from expected behavior or contain sensitive information leaks.

Tool access controls implement permission boundaries, ensuring agents can only invoke authorized functions with appropriate parameters. Context isolation separates user inputs from system instructions, making it harder for malicious prompts to override core directives.

Authentication and authorization frameworks verify user identities and enforce role-based access controls before agents process requests. Audit logging tracks all agent interactions, tool usage, and decision points for forensic analysis and compliance.

Adversarial testing proactively simulates attacks to identify vulnerabilities before deployment. Security-focused evaluation frameworks can test agent resilience against known attack patterns and edge cases.

Modern agent platforms are increasingly incorporating these security primitives as built-in features, but organizations must still configure them appropriately and maintain vigilance as new attack vectors emerge.


AI Agents Guardrails

AI agents operating autonomously in production environments require guardrails—protective boundaries that ensure safe, compliant, and aligned behavior. Unlike traditional software with hardcoded logic paths, agents make dynamic decisions that need real-time constraints to prevent harmful or inappropriate actions.

Types of Guardrails

Input Guardrails filter and validate incoming requests before they reach the agent’s reasoning engine. These prevent processing of prohibited content, malicious prompts, or queries outside the agent’s intended scope. Input guardrails can reject requests containing personal identifiable information (PII), offensive language, or topics the agent shouldn’t address.

Output Guardrails scan agent responses before delivery to users, blocking content that violates policies or quality standards. These catch hallucinations, inappropriate recommendations, leaked sensitive data, or responses that contradict business rules. Output guardrails ensure the agent doesn’t make unauthorized commitments or provide advice beyond its mandate.

Behavioral Guardrails constrain how agents use tools and make decisions during execution. These include spending limits on API calls, restrictions on which databases can be accessed, approval requirements for certain actions, and constraints on autonomous decision-making scope. For example, an agent might require human approval before financial transactions exceed a threshold.

Contextual Guardrails adapt constraints based on user roles, conversation context, or operational conditions. An agent might have different permissions for internal employees versus external customers, or operate under stricter rules during high-risk scenarios.

Implementation Approaches

Guardrails can be implemented through rule-based systems that enforce explicit policies, LLM-based classifiers that evaluate content against nuanced criteria, or hybrid approaches combining both. Modern agent frameworks provide guardrail APIs that intercept agent workflows at key checkpoints—before tool execution, after response generation, and during multi-step reasoning chains.

Effective guardrails balance safety with functionality. Overly restrictive guardrails frustrate users and limit agent utility, while insufficient guardrails expose organizations to risk. Continuous monitoring and adjustment ensure guardrails remain calibrated to organizational needs.


AI Agents Auditability

Auditability establishes transparency and accountability for AI agent operations. As agents make autonomous decisions that impact business outcomes, organizations need comprehensive records of what agents did, why they did it, and what information influenced their decisions.

Essential Audit Components

ComponentWhyWhat to capture?

Decision Trails
Understand why an agent reached a particular conclusion or took a specific action.Capture the agent’s reasoning process, including which tools were invoked, what information was retrieved, how the agent weighted different factors, and what alternatives were considered.

Interaction Logs
To understand the complete context of each interaction, supporting compliance reviews, quality assurance, and issue resolutionAll user inputs, agent outputs, and conversation flows.Logs should include timestamps, user identifiers, session metadata, and any relevant business context.

Tool Usage Records
For security monitoring, cost management, and understanding the agent’s operational footprint across enterprise systems.Document every external system call, API request, database query, and file access performed by the agent.

Performance Metrics
Identify patterns that indicate degraded performance, emerging issues, or opportunities for improvement.Track success rates, response times, error frequencies, and user satisfaction scores.

Compliance and Governance

Auditability supports regulatory compliance by providing evidence of proper agent behavior. Industries with strict oversight—such as finance, healthcare, and legal services—require detailed records demonstrating that agents operated within approved parameters and didn’t make unauthorized decisions.

Audit data also enables retrospective analysis when issues arise. If an agent provides incorrect information or makes a problematic decision, audit trails allow teams to reconstruct the event, identify root causes, and implement corrective measures.

Implementation Considerations

Effective audit systems must balance comprehensiveness with storage and privacy concerns. Logs should capture sufficient detail for meaningful analysis while protecting sensitive information. Retention policies should align with regulatory requirements and business needs.

Modern agent platforms provide structured logging frameworks that automatically capture key events in standardized formats. Organizations should integrate these logs with existing observability and compliance systems, enabling centralized monitoring and analysis across all AI agent deployments.


Conclusion

AI agents represent a fundamental shift in how enterprises deliver value—moving from deterministic software to autonomous systems that reason, decide, and act on behalf of organizations. This power comes with responsibility. Without robust governance, agents can make costly mistakes, expose sensitive data, violate policies, or erode user trust.

The four governance pillars—evaluation, security, guardrails, and auditability—form an integrated framework for deploying AI agents safely and effectively in production environments.

Evaluation ensures agents perform reliably despite their non-deterministic nature, combining subjective LLM-based assessment with objective benchmarking to validate quality before and after deployment.

Security protects against emerging threats unique to AI systems, from prompt injection to identity spoofing, implementing defensive layers that safeguard both the agent and the systems it accesses.

Guardrails establish boundaries that keep agents aligned with organizational policies and ethical standards, constraining behavior without sacrificing the flexibility that makes agents valuable.

Auditability provides transparency into agent operations, creating decision trails that support compliance, enable root cause analysis, and build stakeholder confidence in autonomous systems.

Together, these pillars transform AI agents from experimental prototypes into trustworthy enterprise tools. Organizations that invest in governance upfront accelerate adoption, reduce risk, and unlock the full potential of AI agents to drive business value. As agents become more capable and autonomous, governance won’t be optional—it will be the foundation of successful AI implementation.

Filed Under: AI, AI Agents, Data & AI, Tech Tagged With: agent governance, agents, ai, ai agents, aiagents, data

How Do LLMs Work? A Simple Guide for Kids, Teens & Everyone!

November 30, 2025 by Ashwin Leave a Comment

How LLM Works

Have you ever wondered how ChatGPT or other AI chatbots can write stories, answer questions, and have conversations with you? Let me explain it in a way that’s easy to understand!

The Magic Black Box

Imagine a large language model (LLM) as a mysterious black box. You type something into it (like a question or a story prompt), and it gives you text back as an answer. Simple, right? But what’s happening inside?

Before we peek inside, here’s something important: this black box has been “trained” by reading millions and millions of books, websites, and articles. Think of it like a student who has read every book in the world’s biggest library! All that reading becomes the LLM’s vocabulary and reference material.

Now, let’s open up that black box and see what’s really going on inside.

Inside the Black Box: Three Important Parts

When we look inside, we actually find three smaller boxes working together:

  1. The Encoder – The Translator
  2. The Attention Mechanism – The Detective
  3. The Decoder – The Writer

Let’s explore each one!

Part 1: The Encoder (The Translator)

The Encoder’s job is to translate your words into a language that computers understand: numbers!

Step 1: Making Tokens – First, your sentence gets broken into pieces called “tokens.” These are like puzzle pieces made of words or parts of words. Each token gets assigned a number. For example:

  • “apple” might become token #5234
  • “car” might become token #891

Step 2: Creating a Meaning Map – But here’s where it gets cool! The Encoder doesn’t just turn words into random numbers. It places them on a special map called a “vector embedding.” This map shows how words relate to each other based on their meaning.

Imagine a huge playground where similar words stand close to each other:

  • The word “apple” would stand near “fruit,” “orange,” and “banana”
  • It would also stand somewhat near “computer” (because of Apple computers)
  • But it would be really far away from “car” or “rocket”

This map helps the LLM understand that words can have similar meanings or be used in similar ways.

Part 2: The Attention Mechanism (The Detective)

This is where the real magic happens! The Attention Mechanism is like a detective trying to figure out what you really mean.

Understanding Context Let’s say you type: “The bat flew out of the cave.”

The word “bat” could mean:

  • A flying animal, OR
  • A baseball bat

The Attention Mechanism’s job is to figure out which meaning you’re talking about by looking at the other words around it. When it sees “flew” and “cave,” it realizes you’re probably talking about the animal!

How Does It Do This?

The Attention Mechanism uses something called Multi-Head Attention. Instead of looking at one word at a time, it looks at groups of words together to understand the full picture.

Think of it like this: If you’re trying to understand a painting, you don’t just look at one tiny spot. You step back and look at different parts of it from different angles. That’s what multi-head attention does with your sentence!

The Scoring Game: Q-K-V

Here’s how the detective assigns importance scores to words:

  1. Query (Q): “What am I looking for?” – This is your input word asking a question
  2. Key (K): “What do I know?” – This is the relevant information from the LLM’s huge knowledge base
  3. Value (V): “How important is this?” – This is the score that tells the LLM which words matter most

For our bat example, the word “flew” would get a high score because it’s super important for understanding that we’re talking about the animal, not the baseball bat!

The Feed-Forward Network

After scoring all the words, something called a Feed-Forward Neural Network (FFN) steps in. Think of it as a teacher organizing messy notes into a clean outline. It takes all those scores and organizes them neatly.

This whole process—the scoring and organizing—repeats several times to make sure the LLM really, really understands what you’re asking. Each time through, the understanding gets sharper and clearer.

Part 3: The Decoder (The Writer)

Now that the LLM understands what you’re asking, it’s time to create an answer! That’s the Decoder’s job.

Finding the Best Word

The Decoder looks at all the attention scores and context, then asks: “What’s the best word to say next?”

It searches through its vocabulary and calculates probabilities. For example, if you asked “What color is the sky?” the Decoder might find:

  • “blue” has a 70% probability
  • “gray” has a 15% probability
  • “pizza” has a 0.001% probability (doesn’t make sense!)

The Decoder picks the word with the highest probability—in this case, “blue.”

Building Sentences Word by Word

Here’s something cool: the LLM doesn’t write the whole answer at once. It writes one word at a time, super fast!

After it writes “blue,” it asks again: “What should the next word be?” Maybe it adds “and” or “on” or “during.” Each word it picks becomes part of the context for choosing the next word.

This keeps going—pick a word, add it to the response, pick the next word—until the full answer is complete.

Back to Human Language

Remember how we turned your words into numbers at the beginning? Well, the Decoder does the opposite at the end! It takes all those number tokens and converts them back into words you can read.

And voila! You get your answer!

Putting It All Together

Let’s see the whole process with an example:

You type: “What do cats like to eat?”

  1. Encoder: Converts your question into tokens and places them on the meaning map. It knows “cats” are near “pets” and “animals,” and “eat” is near “food” and “hungry.”
  2. Attention Mechanism: The detective analyzes the question and realizes the important words are “cats” and “eat.” It assigns high scores to these words and understands you’re asking about cat food.
  3. Decoder: Looks at the context and starts writing: “Cats” (highest probability first word) → “like” (next best word) → “to” → “eat” → “fish,” → “chicken,” → “and” → “cat” → “food.”

Each word gets converted back from numbers to text, and you see the complete answer appear on your screen!

The Speed of Thought

All of this—the encoding, the attention detective work, the decoding—happens in just seconds or even split seconds! The LLM processes your input through these three stages so quickly that it feels like magic.

But now you know the secret: it’s not magic. It’s a clever system of translating, understanding context, and finding the most likely words to respond with, all powered by the massive amount of reading the LLM did during its training.

Remember the Key Ideas

  • LLMs are like super-readers who’ve read millions of books and can use that knowledge to chat with you
  • The Encoder turns your words into numbers and maps their meanings
  • The Attention Mechanism is a detective figuring out what you really mean
  • The Decoder picks the best words one by one to answer you
  • Everything happens lightning-fast, even though there are many steps!

Now you know how an LLM works! Pretty cool, right? Next time you chat with an AI, you’ll know exactly what’s happening behind the scenes.

Filed Under: AI, Generative AI, Tech Tagged With: 101, ai, data, genai, llm, llmfundamentals, tech

  • Page 1
  • Page 2
  • Page 3
  • Interim pages omitted …
  • Page 12
  • Go to Next Page »

Primary Sidebar

Connect with me on

  • GitHub
  • LinkedIn
  • Twitter

Recent Posts

  • Agent Evals 101 : Building a Strong Evaluation Framework for AI Agents
  • Enterprise AI Agents: Why Governance is Your Competitive Advantage
  • How Do LLMs Work? A Simple Guide for Kids, Teens & Everyone!
  • Scaling AI Impact: Growing Your CoE and Charting the Future
  • Operationalizing AI Excellence: Processes, Tools, and Talent Strategy for AI CoE
  • Topics

    • Data & AI
      • AI Agents
    • Life
      • Leadership
      • Negotiations
      • Personal Finance
      • Productivity
      • Reading
      • Self Improvement
    • Post Series
      • Intro to Blockchain
    • Tech
      • AI
      • Blockchain
      • Career
      • Certifications
      • Cloud
      • Data
      • Enterprise
      • Generative AI
      • Leadership
      • Presentations
      • Reporting
      • Software Design
      • Stakeholders
    • Uncategorized

Top Posts

  • A Framework to Acing Your Next Tech Presentation
  • Create your first NFT with Smart Contract
  • Why and How to start with Mindful Meditation?
  • Certifications are like licenses!
  • The Evolution of Data Platforms : Beyond Data Lakehouses

Copyright © 2026 · Ashwin Chandrasekaran · WordPress · Log in
All work on this website is licensed under a Creative Commons Attribution-ShareAlike 4.0 International License
The views and opinions expressed in this website are those of the author and do not necessarily reflect the views or positions of the organization he is employed with