-
When do Neural Networks Learn World Models?
Tianren Zhang, Guanyu Chen, Feng Chen
ICLR 2025 Workshop on World Models |
paper
TL;DR: We show that in a multi-task setting, prediction models with a low-degree bias can provably identify latent data-generating variables under mild assumptions.
-
A Method of Supervised Learning from Conflicting Data with Hidden Contexts
Tianren Zhang, Yizhou Jiang, Feng Chen
arXiv preprint |
paper
TL;DR: A formulation and a theoretically grounded method for the problem of open-ended training on data with hidden contexts.
-
Exploring the Hidden Reasoning Process of Large Language Models by Misleading Them
Guanyu Chen*, Peiyang Wang*, ..., Tianren Zhang†, Feng Chen†
In preparation
TL;DR: We show that LLMs can generalize false mathematical reasoning rules to real-world problems, implying the existence of an "abstract-then-reason" process in LLMs.
-
Feature Contamination: Neural Networks Learn Uncorrelated Features and Fail to Generalize
Tianren Zhang*, Chujie Zhao*, Yizhou Jiang, Feng Chen
International Conference on Machine Learning (ICML) 2024 |
paper |
poster |
code
TL;DR: We identify that neural networks can learn task-irrelevant features due to an implicit bias of SGD, resulting in a failure to generalize under distribution shifts.
-
Spatio-Temporal Approximation: A Training-Free SNN Conversion for Transformers
Yizhou Jiang*, Kunlin Hu*, Tianren Zhang, Haichuan Gao, Yuqian Liu,
Ying Fang†, Feng Chen†
International Conference on Learning Representations (ICLR) 2024 |
paper |
code
TL;DR: We propose the first training-free method for converting transformers to purely event-driven spiking neural networks.
-
M3PL: Identifying and Exploiting View Bias of Prompt Learning
Chujie Zhao*, Tianren Zhang*, Guanyu Chen, Yizhou Jiang, Feng Chen
Transactions on Machine Learning Research (TMLR) 2024 |
paper |
code
TL;DR: We identify a view bias in prompt learning of foundation models, i.e., it may
extract only a partial subset of useful features while ignoring others, and we provide an effective fix.
-
Fast Counterfactual Inference for History-Based Reinforcement Learning
Haichuan Gao, Tianren Zhang, Zhile Yang, Yuqing Guo,
Jinsheng Ren, Shangqi Guo†, Feng Chen†
AAAI Conference on Artificial Intelligence (AAAI) 2023 |
paper
TL;DR: We propose a tree-based counterfactual inference method for learning history representations in reinforcement learning.
-
Adjacency Constraint for Efficient Hierarchical Reinforcement Learning
Tianren Zhang*, Shangqi Guo*†, Tian Tan,
Xiaolin Hu, Feng Chen†
IEEE Transactions on Pattern Analysis and Machine Intelligence (TPAMI) 2022 |
paper
-
Subjective Learning for Conflicting Data
Tianren Zhang, Yizhou Jiang, Xin Su,
Shangqi Guo, Feng Chen
ICLR 2022 Workshop on Agent Learning in Open-Endedness |
paper
TL;DR: An initial attempt of formulating and addressing the problem of data conflicts in open-ended learning.
-
CRIL: Continual Robot Imitation Learning via Generative
and Prediction Model
Chongkai Gao, Haichuan Gao, Shangqi Guo, Tianren Zhang, Feng Chen
IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS) 2021 |
paper |
code
TL;DR: A continual imitation learning method for robot learning based on generation and prediction.
-
Generating Adjacency-Constrained Subgoals for Hierarchical Reinforcement Learning
Tianren Zhang*, Shangqi Guo*, Tian Tan, Xiaolin Hu†, Feng Chen†
Conference on Neural Information Processing Systems (NeurIPS) 2020 (Spotlight) |
paper |
code
TL;DR: We show that a state representation based on state adjacency can significantly improve the sample efficiency of hierarchical reinforcement learning.
-
Deep Meta Metric Learning
Guangyi Chen, Tianren Zhang, Jiwen Lu, Jie Zhou
International Conference on Computer Vision (ICCV) 2019 |
paper |
code