NLP

Why Do Language Models Hallucinate? featured image

Why Do Language Models Hallucinate?

An analysis of why language models hallucinate — hallucinations arise from statistical pressures in training and evaluation procedures that reward guessing over acknowledging …

Why Entropy Matters in Machine Learning? featured image

Why Entropy Matters in Machine Learning?

Understanding entropy and why it's a core concept in decision trees, neural networks, and loss functions like cross-entropy.

Understanding FlashAttention: Inner vs Outer Loop Optimization featured image

Understanding FlashAttention: Inner vs Outer Loop Optimization

FlashAttention is a groundbreaking optimization technique for computing attention in Transformer models, drastically improving GPU memory efficiency through inner vs outer loop …

Adversarial Attacks on Large Language Models (LLMs) featured image

Adversarial Attacks on Large Language Models (LLMs)

An overview of adversarial attacks on large language models (LLMs) — how manipulated inputs can deceive models into generating harmful or incorrect outputs, covering key attack …

GLiNER: A Generalist Model for Named Entity Recognition using Bidirectional Transformers featured image

GLiNER: A Generalist Model for Named Entity Recognition using Bidirectional Transformers

A detailed summary of the GLiNER paper, introducing a lightweight, scalable, and highly effective model for open-type named entity recognition using bidirectional transformers with …

Comparing batch vs layer normalization featured image

Comparing batch vs layer normalization

The purpose of this post is just to understand the key difference between two types of well-known normalization techniques.

admin