AI Agent
2025-11-12
2026-02-09
Building AI Agent from first principles with a durable understanding that frameworks alone cannot, This post structures a learning path that separates the core theory from the "from scratch" implementation, using resources that are independent of major agent frameworks, such as LlamaIndex.
5151 words
|
26 minutes
Cover Image of the Post
Transformer
2025-09-19
2026-02-04
I was originally studying LLM on Hugging Face (https://huggingface.co/learn/llm-course) and notice that transformer occupied significant portion of their learning materials so I decided to look into it deeply
14909 words
|
75 minutes
Cover Image of the Post
The Universal Approximation Theorem - The Mathematical Guarantee of Learnability for Neural Networks
2025-06-05
2026-01-30
This theorem serves as the fundamental existence proof for deep learning, demonstrating that a standard neural network with sufficient neurons has the theoretical capacity to approximate any continuous function. It assures engineers that the architecture itself is capable of modeling any complex real-world pattern, shifting the challenge from "is this representable?" to "how do we find the parameters to represent it?"
3869 words
|
19 minutes
Cover Image of the Post
Why is AI Deeply-Seated with Philosophy and Language?
2025-05-29
2026-01-27
AI is not just a consumer of linguistic data; it is a producer of new philosophical questions. It forces us to be more precise about what we mean by "meaning," "understanding," and "thought," turning centuries of abstract debate into a pressing, practical challenge.
6139 words
|
31 minutes
Cover Image of the Post
Gradient Checkpoint in Training Neural Networks
My first ASR model fine-tuning hit really hard at an error by gradient checkpoint which draw my attention to this topic. Gradient Checkpointing (also known as activation checkpointing) is a memory-optimization strategy used during the training of deep neural networks. It allows you to train significantly larger models (or use larger batch sizes) on limited GPU memory by trading computation time for memory space.
711 words
|
4 minutes
Cover Image of the Post
An Enquiry Concerning Human Understanding - The Ultimate Justification of Big Data, the Foundation of AI
2025-10-03
2026-01-08
Why would AI work? Certainly technological advancements and market conditions have their contributions. But it is the philosophical justification presented by one of the greatest philosophers, David Hume, that has made it inevitable which this post shows by examining and linking his most phenomenon work - "An Enquiry Concerning Human Understanding"
11788 words
|
59 minutes
Cover Image of the Post
Where is the computational limits of AI? The Halting Problem
2025-10-06
2025-12-11
If the Universal Approximation Theorem expresses the unprecedented power of AI, the Halting Problem poses the opposite - its computational limit.
1714 words
|
9 minutes
Cover Image of the Post
Kant – Eine große Synthese zwischen Rationalismus und Empirismus
2024-12-27
2025-12-11
A Masterpiece of Idealism of Anti-Idealism - Transcendental Idealism
2902 words
|
15 minutes
Cover Image of the Post