📌 Technical Origin of OpenML - A Quantum Physics Problem
The conceptual foundation of OpenML dates all the way back to 2013 when a college boy in Physics major was reading and implementing the Matrix Numerov Method for Solving Schrödinger Equation
2161 words
|
11 minutes
Cover Image of the Post
Transformer
2025-09-19
2026-02-03
I was originally studying LLM on Hugging Face (https://huggingface.co/learn/llm-course) and notice that transformer occupied significant portion of their learning materials so I decided to look into it deeply
14842 words
|
74 minutes
Cover Image of the Post
AI Agent
2025-11-12
2026-02-02
Building AI Agent from first principles with a durable understanding that frameworks alone cannot, This post structures a learning path that separates the core theory from the "from scratch" implementation, using resources that are independent of major agent frameworks, such as LlamaIndex.
5169 words
|
26 minutes
Cover Image of the Post
The Universal Approximation Theorem - The Mathematical Guarantee of Learnability for Neural Networks
2025-06-05
2026-01-30
This theorem serves as the fundamental existence proof for deep learning, demonstrating that a standard neural network with sufficient neurons has the theoretical capacity to approximate any continuous function. It assures engineers that the architecture itself is capable of modeling any complex real-world pattern, shifting the challenge from "is this representable?" to "how do we find the parameters to represent it?"
3869 words
|
19 minutes
Cover Image of the Post
Why is AI Deeply-Seated with Philosophy and Language?
2025-05-29
2026-01-27
AI is not just a consumer of linguistic data; it is a producer of new philosophical questions. It forces us to be more precise about what we mean by "meaning," "understanding," and "thought," turning centuries of abstract debate into a pressing, practical challenge.
6139 words
|
31 minutes
Cover Image of the Post
Gradient Checkpoint in Training Neural Networks
My first ASR model fine-tuning hit really hard at an error by gradient checkpoint which draw my attention to this topic. Gradient Checkpointing (also known as activation checkpointing) is a memory-optimization strategy used during the training of deep neural networks. It allows you to train significantly larger models (or use larger batch sizes) on limited GPU memory by trading computation time for memory space.
711 words
|
4 minutes
Cover Image of the Post
An Enquiry Concerning Human Understanding - The Ultimate Justification of Big Data, the Foundation of AI
2025-10-03
2026-01-08
Why would AI work? Certainly technological advancements and market conditions have their contributions. But it is the philosophical justification presented by one of the greatest philosophers, David Hume, that has made it inevitable which this post shows by examining and linking his most phenomenon work - "An Enquiry Concerning Human Understanding"
11788 words
|
59 minutes
Cover Image of the Post
Why Do People Study Ancient Languages?
2025-08-03
2026-01-06
A fascination with things from the past, like ancient languages, is a common and deeply rooted human interest. There are a number of reasons for this, ranging from the intellectual and practical to the emotional and psychological.
5894 words
|
29 minutes
Cover Image of the Post