Agentic AI Systems: Planning, Memory, and Tool Use in Autonomous Large Language Model Agents
Abstract The emergence of large language models (LLMs) capable of reasoning, tool invocation, and multi-step planning has precipitated a new…
Read MoreBlog and Portfolio
Abstract The emergence of large language models (LLMs) capable of reasoning, tool invocation, and multi-step planning has precipitated a new…
Read MoreAbstract Vision-language models (VLMs) have emerged as one of the most consequential developments in modern deep learning, enabling systems to…
Read MoreAbstract The transformer architecture has dominated sequence modeling for half a decade, but its $O(n^2)$ attention complexity imposes a hard…
Read MoreAbstract Position encoding is a foundational design choice in transformer architectures, enabling models to exploit token order without recurrence. Rotary…
Read MoreAbstract Full fine-tuning of large language models (LLMs) has become computationally prohibitive at scales exceeding tens of billions of parameters.…
Read MoreAbstract The standard self-attention mechanism in Transformer architectures exhibits quadratic time and memory complexity in sequence length, forming a fundamental…
Read MoreAbstract Model quantization — the process of representing neural network weights and activations with reduced numerical precision — has become…
Read MoreAbstract Neural Architecture Search (NAS) automates the discovery of high-performing neural network architectures, offering a principled alternative to manual design.…
Read MoreAbstract Federated learning (FL) offers a compelling paradigm for training natural language processing models across distributed clients without centralizing raw…
Read More