All Blogs
Byte Pair Encoding in LLMs
In this blog, we will learn about BPE (Byte Pair Encoding) - the tokenization algorithm used by most modern Large Language Models (LLMs) to break text into smaller pieces before processing it.
Paged Attention in LLMs
In this blog, we will learn about Paged Attention, a technique that solves the memory waste problem of KV Cache, allowing LLMs to serve many more users at the same time.
KV Cache in LLMs
In this blog, we will learn about KV Cache - where K stands for Key and V stands for Value - and why it is used in Large Language Models (LLMs) to speed up text generation.
Causal Masking in Attention
In this blog, we will learn about the Causal Masking in Attention.
Linear Regression vs Logistic Regression
In this blog, we will learn about Linear Regression vs Logistic Regression in Machine Learning.
Supervised vs Unsupervised Learning
In this blog, we will learn about Supervised vs Unsupervised Learning in Machine Learning.
Android Push Notification Flow using FCM
In this blog, we will learn about the Android Push Notification Flow using FCM.
Android System Design Interviews
Today, we are going to discuss everything about the Android System Design Interviews.
Write-Ahead Logging (WAL)
In this blog, we will learn about Write-Ahead Logging (WAL) and why it is used internally in databases.