About
📌 Technical Origin of OpenML - A Quantum Physics Problem
2026-01-26
The conceptual foundation of OpenML dates all the way back to 2013 when a college boy in Physics major was
reading and implementing the Matrix Numerov Method for Solving Schrödinger Equation
2161 words
|
11 minutes
Transformer
I was originally studying LLM on Hugging Face (https://huggingface.co/learn/llm-course) and notice that transformer
occupied significant portion of their learning materials so I decided to look into it deeply
14842 words
|
74 minutes
AI Agent
Building AI Agent from first principles with a durable understanding that frameworks alone cannot, This post
structures a learning path that separates the core theory from the "from scratch" implementation, using resources that
are independent of major agent frameworks, such as LlamaIndex.
5169 words
|
26 minutes
The Universal Approximation Theorem - The Mathematical Guarantee of Learnability for Neural Networks
This theorem serves as the fundamental existence proof for deep learning, demonstrating that a standard neural network
with sufficient neurons has the theoretical capacity to approximate any continuous function. It assures engineers that
the architecture itself is capable of modeling any complex real-world pattern, shifting the challenge from "is this
representable?" to "how do we find the parameters to represent it?"
3869 words
|
19 minutes
Why is AI Deeply-Seated with Philosophy and Language?
AI is not just a consumer of linguistic data; it is a producer of new philosophical questions. It forces us to be more
precise about what we mean by "meaning," "understanding," and "thought," turning centuries of abstract debate into a
pressing, practical challenge.
6139 words
|
31 minutes
Gradient Checkpoint in Training Neural Networks
2026-01-21
My first ASR model fine-tuning hit really hard at an error by gradient checkpoint which draw my attention to this
topic. Gradient Checkpointing (also known as activation checkpointing) is a memory-optimization strategy used during
the training of deep neural networks. It allows you to train significantly larger models (or use larger batch sizes)
on limited GPU memory by trading computation time for memory space.
711 words
|
4 minutes
An Enquiry Concerning Human Understanding - The Ultimate Justification of Big Data, the Foundation of AI
Why would AI work? Certainly technological advancements and market conditions have their contributions. But it is the
philosophical justification presented by one of the greatest philosophers, David Hume, that has made it inevitable
which this post shows by examining and linking his most phenomenon work - "An Enquiry Concerning Human Understanding"
11788 words
|
59 minutes
Why Do People Study Ancient Languages?
A fascination with things from the past, like ancient languages, is a common and deeply rooted human interest. There
are a number of reasons for this, ranging from the intellectual and practical to the emotional and psychological.
5894 words
|
29 minutes