Nested Learning
This post introduces a new learning paradigm, Nested Learning (NL), which represents a model as a coherent set of nested, multi-level, and/or parallel optimization problems, each with its own context flow.
Common ML Concepts Explained Simply
This post introduces basic concepts of tokenization, decoding, prompting, tool-augmented agents, RAG, RLHF, VAEs, diffusion models, and LoRA—presented with standard objective functions and probabilistic notation.
Performance Hints in C++
This post tidies up the core ideas for how to think about, estimate, measure, and ship performance improvements—mainly for single-binary software (not distributed systems or ML hardware tuning).
Accurate Sampling-Based Cardinality Estimation for Complex Graph Queries
This post provides a novel, strongly consistent method for accurately and efficiently estimating the cardinality of complex graph queries.
Towards Better Evaluation for Dynamic Link Prediction
This post provides novel evaluation strategies for dynamic link prediction, introducing new datasets, robust negative sampling methods, and the EdgeBank baseline for better model assessment.
RAFT: Integrating RAG with Fine-Tuning
This post provides a comprehensive overview of RAG and Fine-Tuning methods and introduces Retrieval Augmented Fine-Tuning (RAFT) as a unified approach to enhance language model training.
Reinforcement learning on graphs: A survey
This post provides a comprehensive overview of RL and graph mining methods and generalize these methods to Graph Reinforcement Learning (GRL) as a unified formulation.
Primal Wasserstein Imitation Learning
This post introduces Primal Wasserstein Imitation Learning (PWIL) method, based on Wasserstein distance, derives a reward function offline and efficiently replicates expert behavior in MuJoCo continuous control tasks.
11 post articles, 2 pages.