Page Not Found
Page not found. Your pixels are in another canvas.
A list of all the posts and pages found on the site. For you robots out there is an XML version available for digesting as well.
Page not found. Your pixels are in another canvas.
About me
Awards
This is a page not in th emain menu
Services
Published:
Multi-Reference Preference Optimization (MRPO) for Large Language Models (AAAI 2025) Read more
Published:
Why and How Memory Matters for LLMs? Read more
Published:
Plug, Play, and Generalize: Length Extrapolation with Pointer-Augmented Neural Memory Read more
Published:
xLSTM: Extended Long Short-Term Memory Read more
Published:
How to Spot When Your Large Language Model is Misleading You Read more
Published:
Scalable MatMul-free Language Modeling Read more
Published:
Mamba: Linear-Time Sequence Modeling with Selective State Spaces Read more
Published:
Beyond Surprise: Direct and Causal Exploration in Deep Reinforcement Learning Read more
Published:
About recent LLM alignment finetuning techniques such as RLHF, DPO, KTO, IPO and SPIN Read more
Published:
Novelty or Surprise: How to Make Your Deep Reinforcement Learning Agents Curious? Read more
Published:
Legacy of Exploration in Reinforcement Learning Read more
Published:
Published:
Memory is just storage. Whenever computation needs to store interim results, it must ask for memory. This fundamental principle applies to any scenario where memory is required, yet a closer interpretation of memory’s role in each domain reveals a different understanding of its functionality and benefit. Read more
Published:
When a human programmer codes, he often uses core libraries to construct his programs. Most of the time, the program memory stores these static libraries, and let big programs be created dynamically during computation. The libraries are unitary components constructing bigger programs. Maintaining small and functionally independent sub-programs such as libraries encourage program utilisations since an immense program must refer to different libraries to complete its task. Indeed, it also eliminates redundancies as the stored programs-the core libraries are not overlapping each other. Read more
Published:
A neural network uses its weight to compute inputs and return outputs as computation results. Hence, the weight can be viewed as the neural network’s program. If we maintain a program memory of different weights responsible for various computation functions, we have a neural Universal Turing Machine. Obvious scenarios where a Program Memory may help: Read more
Published:
Memory-augmented neural networks (MANNs) store data in their external memory, resembling Turning Machines. Despite being theoretically Turing-Complete, MANNs cannot be trained flexibly to solve any task due to the lack of program memory. Without storing programs, it is hard to perform complicated tasks such as simulating recursive functional calls or implementing divide-and-conquer algorithms. As long as programs are not treated as data, the computing capability of neural networks is still limited. Read more
Published:
Imagine this, you only have short-term memory. You can only remember what happen during the day, and when you wake up, your mind refreshes. Without long-term memory, you cannot remember even your birthday, your last month’s payment or where had you been last week. To survive, you must note down every thing, and re-learn these facts every morning. That would be so inconvenient for dementia patients who suffer this kind of disease. In the same vein, if your mind only revolves around System 2 (slow and sophisticated), you will fail to think quickly and intuitively, and process life events effortfully. No matter how simple System 1, it is required and stands apart from System 2. It seems like a good model of memory should treat the memory as a bunch of modules, representing different functions and collaborating to deliver the desired outcome. Inspired by this observations, I wrote couple of papers based on multi-memory systems that analyze various aspects of memory functions such as item-relational storage, view/channel fusion, information encoder-decoder. Read more
Published:
Check our papers: Read more
Published:
Check our papers: Read more
Published:
Published:
Despite huge successes in breaking human records, current training of RL agents is prohibitively expensive in terms of time, GPUs, and samples. For example, it requires hundreds of millions or even billions of environment steps to reach human-level performance on Atari games-a common benchmark in modern RL. That is only doable with simulation, not real-world problems like robotics or industrial planning. The problem of sample-inefficiency is exacerbated in real environments, which can be stochastic, partially observable, noisy or long-term. Another issue is model complexity. RL algorithms are getting more complicated, coupled with numerous hyperparameters that need to be tuned carefully. That again accelerates the cost of training RL agents. Read more
Published:
Memory is the core of intelligence. Thanks to memory, human can effortlessly recognize objects, recall past events, plan the future, explain surrounding environments and reason from facts. From cognitive perspective, memory can take many forms and functionalities (see figure below). Read more
Published:
Mini ARC Analog Programme (MAAP)
Amount: $254,200 AUD
Duration: 2021-2023
Published:
Discovery Early Career Researcher Award (DECRA)
Amount: $428,331 AUD
Duration: 2025-2028
Published:
Cohere For AI Research Grant
Program Amount: $10,000 USD
Duration: 2025-2026
Published:
Just some of my incomplete ideas: Read more
Published:
Variational Inference (VI) first starts as a handy tool in Bayesian inference to approximate intractable posterior. Now, its usage goes beyond Bayesian inference, and we can see VI everywhere in classic learning, and deep learning-anywhere needs an approximation. After this lecture, we should: Read more
Published:
Authors: Asjad Khan*, Hung Le*, Kien Do, Truyen Tran, Aditya Ghose, Hoa Dam, Renuka Sindhgatta.
Link
Published:
Authors: Duc Nguyen, Nhan Tran, Hung Le.
Link
Published:
Authors: Majid Abdolshah, Hung Le, Thommen Karimpanal George, Sunil Gupta, Santu Rana, Svetha Venkatesh.
Link
Published/Accepted at PAKDD, 2018
Authors: Hung Le, Truyen Tran, Svetha Venkatesh
Code•PDF•Link
Published/Accepted at KDD, 2018
Authors: Hung Le, Truyen Tran, Svetha Venkatesh
Code•Poster•PDF•Link
Published/Accepted at NeurIPS, 2018
Authors: Hung Le, Truyen Tran, Thin Nguyen, Svetha Venkatesh
Code•Poster•PDF•Link
Published/Accepted at ICLR (Oral), 2019
Authors: Hung Le, Truyen Tran, Svetha Venkatesh
Code•Poster•Blog•PDF•Link
Published/Accepted at ICLR, 2020
Authors: Hung Le, Truyen Tran, Svetha Venkatesh
Code•Talk•Slides•PDF•Link
Published/Accepted at ICPR, 2020
Authors: Huu Tin Hoang, Chun-Jen Peng, Hung Tran, Hung Le, Huy Hoang Nguyen.
Link
Published/Accepted at ICML, 2020
Authors: Hung Le, Truyen Tran, Svetha Venkatesh
Code•Talk•Slides•PDF•Link
Published/Accepted at ICML (Spotlight), 2021
Authors: Majid Abdolshah, Hung Le, Thommen George Karimpanal, Sunil Gupta, Santu Rana, Svetha Venkatesh.
Link
Published/Accepted at ICSOC, 2021
Authors: Asjad Khan, Aditya Ghose, Hoa Dam, Hung Le, Truyen Tran, Kien Do.
Link
Published/Accepted at KDD (Tutorial), 2021
Authors: Truyen Tran, Vuong Le, Hung Le, Thao M Le
Link
Published/Accepted at ICONIP, 2021
Authors: Duy-Hung Nguyen, Bao-Sinh Nguyen, Nguyen Viet Dung Nghiem, Dung Tien Le, Mim Amina Khatun, Minh-Tien Nguyen, Hung Le.
Link
Published/Accepted at NeurIPS, 2021
Authors: Hung Le, Thommen Karimpanal George, Majid Abdolshah, Truyen Tran, Svetha Venkatesh
Code•Talk•Poster•Slides•PDF•Link
Published/Accepted at AAAI (Oral), 2022
Authors: Hung Le, Thommen Karimpanal George, Majid Abdolshah, Kien Do, Dung Nguyen, Svetha Venkatesh
Code•Talk•Slides•PDF•Link
Published/Accepted at AAMAS, 2022
Authors: Dung Nguyen, Phuoc Nguyen, Hung Le, Kien Do, Truyen Tran, Svetha Venkatesh
Link
Published/Accepted at ICLR, 2022
Authors: Kha Pham, Hung Le, Man Ngo, Truyen Tran, Bao Ho, Svetha Venkatesh
Link
Published/Accepted at NAACL-Findings, 2022
Authors: Duy-Hung Nguyen, Nguyen Viet Dung Nghiem, Bao-Sinh Nguyen, Tien Dung Le, Minh-Tien Nguyen, Shahab Sabahi, Hung Le
Link
Published/Accepted at ICML (Spotlight), 2022
Published/Accepted at ECCV, 2022
Authors: Kien Do, Haripriya Harikumar, Hung Le, Dung Nguyen, Truyen Tran, Santu Rana, Dang Nguyen, Willy Susilo, Svetha Venkatesh
Link
Published/Accepted at ICONIP (Oral), 2022
Authors: Bao-Sinh Nguyen, Quang-Bach Tran, Tuan-Anh Nguyen Dang, Duc Nguyen, Hung Le.
Link
Published/Accepted at ICONIP (Oral), 2022
Authors: Bao-Sinh Nguyen, Dung Tien Le, Hieu M. Vu, Tuan-Anh D. Nguyen, Minh-Tien Nguyen, Hung Le
Link
Published/Accepted at NeurIPS, 2022
Authors: Kien Do, Hung Le, Dung Nguyen, Dang Nguyen, HARIPRIYA HARIKUMAR, Truyen Tran, Santu Rana, Svetha Venkatesh
Link
Published/Accepted at NeurIPS, 2022
Authors: Kha Pham, Hung Le, Man Ngo, Truyen Tran
Link
Published/Accepted at NeurIPS (Spotlight), 2022
Authors: Hung Le, Thommen Karimpanal George, Majid Abdolshah, Dung Nguyen, Kien Do, Sunil Gupta, Svetha Venkatesh
PDF•Link
Published/Accepted at AAAI, 2023
Authors: Dung Nguyen, Phuoc Nguyen, Hung Le, Kien Do, Svetha Venkatesh, Truyen Tran
Link
Published/Accepted at ICLR, 2023
Authors: Kha Pham, Hung Le, Man Ngo, Truyen Tran
Link
Published/Accepted at IJCAI, 2023
Authors: Dung Nguyen, Hung Le, Kien Do, Svetha Venkatesh, Truyen Tran
Link
Published/Accepted at International Journal of Impact Engineering, 2023
Authors: Shannon Ryan, Neeraj Mohan Sushma, Hung Le, Arun Kumar A V, Santu Rana, Sevvandi Kandanaarachchi, Svetha Venkatesh
Link
Published/Accepted at ACML, 2023
Authors: Ragja Palakkadavath, Thanh Nguyen-Tang, Hung Le, Svetha Venkatesh, Sunil Gupta
Link
Published/Accepted at Artificial Intelligence, 2023
Authors: Thommen George Karimpanal, Hung Le, Majid Abdolshah, Santu Rana, Sunil Gupta, Truyen Tran, Svetha Venkatesh.
Link
Published/Accepted at Transactions on Machine Learning Research (TMLR) , 2023
Authors: Thanh Duc Hoang, Do Viet Tung, Duy-Hung Nguyen, Bao-Sinh Nguyen, Huy Hoang Nguyen, and Hung Le.
Link
Published/Accepted at AAMAS (Oral), 2024
Authors: Hung Le, Kien Do, Dung Nguyen and Svetha Venkatesh
Code•PDF•Link
Published/Accepted at IJCAI, 2024
Authors: Dung Nguyen, Hung Le, Kien Do, Sunil Gupta, Svetha Venkatesh and Truyen Tran
Link
Published/Accepted at ECML-PKDD, 2024
Authors: Minh Hoang Nguyen, Hung Le, Svetha Venkatesh
Code•Link
Published/Accepted at ISSTA, 2024
Authors: Thanh-Dat Nguyen, Tung Do-Viet, Hung Nguyen-Duy, Tuan-Hai Luu, Hung Le, Bach Le, Patanamon (Pick) Thongtanunam
Link
Published/Accepted at ECAI, 2024
Authors: Kien Do, Dung Nguyen, Hung Le, Thao Le, Dang Nguyen, Haripriya Harikumar, Truyen Tran, Santu Rana and Svetha Venkatesh
Link
Published/Accepted at ECAI (Oral), 2024
Authors: Van Dai Do, Quan Tran, Svetha Venkatesh and Hung Le
PDF•Link
Published/Accepted at Transactions on Machine Learning Research (TMLR) , 2024
Authors: Hung Le, Dung Nguyen, Kien Do, Svetha Venkatesh, Truyen Tran
Link•Code
Published/Accepted at WACV, 2024
Authors: Ragja Palakkadavath, Hung Le, Thanh Nguyen-Tang, Svetha Venkatesh, Sunil Gupta
Link
Published/Accepted at AAAI, 2025
Authors: Hung Le, Quan Hung Tran, Dung Nguyen, Kien Do, Saloni Mittal, Kelechi Ogueji, Svetha Venkatesh
Link•Code•Blog
Published:
Published:
Published:
Published:
Published:
Undergraduate course (finished), Deakin University, 2018
SIT-112, Data Science Concepts. Spring, 2018.
PhD course (finished), Deakin University, 2021
S913, PhD. Candidate Bao Duong Nguyen. Advanced Machine Learning for Causal Discovery. 2021-2024
PhD course, Deakin University, 2021
S913, PhD. Candidate Ragja Palakkadavath. Domain Generalization for Algorithmic Robustness and Fairness. 2022-2025
PhD course (finished), Deakin University, 2021
S913, PhD. Candidate Kha Pham. Memory for Fast Adaptation in Neural Networks. 2021-2024
PhD course, Deakin University, 2023
F975, PhD. Candidate Minh Hoang Nguyen. Causal Reinforcement Learning, The Synergistic Relationship between Causal Inference and Reinforcement Learning. 2023-2026
PhD course, Deakin University, 2023
F975, PhD. Candidate Van Dai Do. Efficient And Safe Large Language Models With Reinforcement Learning. 2023-2026
PhD course, Deakin University, 2024
F975, PhD. Candidate Truong Giang Do. Improving Foundation Models by Addressing the Binding Problem. 2024-2027
PhD course, Deakin University, 2024
F975, PhD. Candidate Manh Nguyen. Understand Large Language Models with Counterfactual Explaination. 2024-2027