About Me

I am currently working at Zhipu AI, where I lead research on long context models. I completed both my PhD and undergraduate studies in the Department of Computer Science at Tsinghua University, under the supervision of Professor Prof. Juanzi Li. My primary research interests include long context, pre-trained models, and knowledge graphs. Through my work, I aim to advance the understanding and application of these areas, contributing to the development of more sophisticated AI systems.

Publications & Preprints

2022

The list of papers is no longer updated, More details can be found on my Google Scholar.

Do Pre-trained Models Benefit Knowledge Graph Completion? A Reliable Evaluation and a Reasonable Approach
ACL 2022 Findings
Xin Lv, Yankai Lin, Yixin Cao, Lei Hou, Juanzi Li, Zhiyuan Liu, Peng Li, Jie Zhou
[PDF] [Code]
We highlight a more accurate evaluation setting OWA and propose a novel PLM-based KGC model named PKGC. In our experiments, we verify that CWA cannot bring accurate evaluation results.
Program Transfer for Answering Complex Questions over Knowledge Bases
ACL 2022
Shulin Cao, Jiaxin Shi, Zijun Yao, Xin Lv, Jifan Yu, Lei Hou, Juanzi Li, Zhiyuan Liu, Jinghui Xiao
[PDF] [Code]
We propose the approach of program transfer, which aims to leverage the valuable program annotations on the rich-resourced KBs as external supervision signals to aid program induction for the low-resourced KBs that lack program annotations.
Triple-as-Node Knowledge Graph and Its Embeddings
DASFAA 2022
Xin Lv, Jiaxin Shi, Shulin Cao, Lei Hou, Juanzi Li
[PDF] [Code]
We contribute a benchmark WD16K with additional fact-relevant relations, and a framework FactE, which can represent facts, entities and relations in the same space via attention.
SQUIRE: A Sequence-to-sequence Framework for Multi-hop Knowledge Graph Reasoning
Arxiv 2022
Yushi Bai, Xin Lv, Juanzi Li, Lei Hou, Yincen Qu, Zelin Dai, Feiyu Xiong
[PDF] [Code]
SQUIRE treats the triple query and the evidential path as sequences and utilizes Transformer to learn and infer in an end-to-end fashion. We propose rule-enhanced learning and iterative training to further boost performance.
ICLEA: Interactive Contrastive Learning for Self-supervised Entity Alignment
Arxiv 2022
Kaisheng Zeng, Zhenhao Dong, Lei Hou, Yixin Cao, Minghao Hu, Jifan Yu, Xin Lv, Juanzi Li, Ling Feng
[PDF] [Code]
We propose an interactive contrastive learning model for self-supervised EA. The model encodes not only structures and semantics of entities (including entity name, entity description, and entity neighborhood), but also conducts cross-KG contrastive learning by building pseudo-aligned entity pairs.

2021

Is Multi-Hop Reasoning Really Explainable? Towards Benchmarking Reasoning Interpretability
EMNLP 2021
Xin Lv, Yixin Cao, Lei Hou, Juanzi Li, Zhiyuan Liu, Yichi Zhang and Zelin Dai
[PDF] [Code]
We propose a framework to quantitatively evaluate the interpretability of multi-hop reasoning models. Based on this framework, we annotate a dataset to form a benchmark that can accurately evaluate the interpretability.
Interpretable and Low-Resource Entity Matching via Decoupling Feature Learning from Decision Making
ACL 2021
Zijun Yao, Chengjiang Li, Tiansi Dong, Xin Lv, Jifan Yu, Lei Hou, Juanzi Li, Yichi Zhang and Zelin Dai
[PDF] [Code]
We propose to decouple the representation learning stage and the decision making stage to fully utilize unlabeled data for entity matching task.
Are Missing Links Predictable? An Inferential Benchmark for Knowledge Graph Completion
ACL 2021
Yixin Cao, Xiang Ji, Xin Lv, Juanzi Li, Yonggang Wen and Hanwang Zhang
[PDF] [Code]
We highlighted three principles for KGC datasets: inferential ability, assumptions, and patterns, and contribute a large-scale dataset InferWiki. We established a benchmark with three types of seven KGC models on two tasks of triple classification and link prediction.
KACC: A Multi-task Benchmark for Knowledge Abstraction, Concretization and Completion
ACL 2021 Findings
Jie Zhou, Shengding Hu, Xin Lv, Cheng Yang, Zhiyuan Liu, Wei Xu, Jie Jiang, Juanzi Li and Maosong Sun
[PDF] [Code]
We focus on the problems of knowledge abstraction, concretization, and completion. We propose a benchmark to test the abilities of models on KACC.

2020

Dynamic Anticipation and Completion for Multi-Hop Reasoning over Sparse Knowledge Graph
EMNLP 2020
Xin Lv, Xu Han, Lei Hou, Juanzi Li, Zhiyuan Liu, Wei Zhang, Yichi Zhang, Hao Kong, Suhui Wu
[PDF] [Code]
We propose a reinforcement learning model named DacKGR with two strategies (i.e., dynamic anticipation and dynamic completion) designed for sparse KGs. These strategies can ease the sparsity of KGs.

2019

Adapting Meta Knowledge Graph Information for Multi-Hop Reasoning over Few-Shot Relations
EMNLP 2019
Xin Lv, Yuxian Gu, Xu Han, Lei Hou, Juanzi Li, Zhiyuan Liu
[PDF] [Code]
We propose a meta-learning based model named Meta-KGR for multi-hop reasoning over few-shot relations of knowledge graphs.

2018

Differentiating Concepts and Instances for Knowledge Graph Embedding
EMNLP 2018
Xin Lv, Lei Hou, Juanzi Li, Zhiyuan Liu
[PDF] [Code]
We propose a new knowledge embedding model named TransC. TransC embeds instances, concepts, and relations in the same space to deal with the transitivity of isA relations.
OpenKE: An Open Toolkit for Knowledge Embedding
EMNLP 2018
Xu Han, Shulin Cao, Xin Lv, Yankai Lin, Zhiyuan Liu, Maosong Sun, Juanzi Li
[PDF] [Code]
We propose an efficient open toolkit OpenKE for knowledge embedding. OpenKE builds a unified underlying platform to organize data and memory. It also applies GPU learning and parallel learning to speed up training.