Welcome to contribute!!!
Contributor: Runxin Xu, Shuang Zeng
And thank Yuxuan Fan, Yifan Song for their suggestions!
Content
- Named Entity Recognition
- Coreference Resolution
- Relation Extraction
- Event Extraction
- Joint Information Extraction
- Sentence-level Named Entity Recognition
- Chinese Named Entity Recognition
- Few-shot Named Entity Recognition
- Document-level Named Entity Recognition
- Survey
- A Survey on Deep Learning for Named Entity Recognition (TKDE2020): 2020
- A Survey on Recent Advances in Named Entity Recognition from Deep Learning models (COLING2018): 2018
- A survey of named entity recognition and classification (NYU): 2017
Given a sentence, the task aims at recognizing the entities and their entity types in the sentence. For example, Alpha is an Protein type entity in the figure.
- How to handle the nested entities?
- How to handle the discontinuous entities?
- Sequence Labeling
- End-to-end Sequence Labeling via Bi-directional LSTM-CNNs-CRF (ACL2016)
- A Neural Layered Model for Nested Named Entity Recognition (NAACL2018): stack多层来从内到外识别实体
- Pyramid: A Layered Model for Nested Named Entity Recognition (ACL2020): stack多层变成Pyramid
- TENER: Adapting Transformer Encoder for Named Entity Recognition (arXiv): 改进position embedding
- Nested Named Entity Recognition with Partially-Observed TreeCRFs (AAAI2021): TreeCRF
- Span-level
- A Local Detection Approach for Named Entity Recognition and Mention Detection (ACL2017)
- A Boundary-aware Neural Model for Nested Named Entity Recognition (EMNLP2019): 先判断哪些token可以作为entity的开头结尾,然后遍历判断所有potential spans
- Multi-Grained Named Entity Recognition (ACL2019): word-level+sentence-level特征先找possible spans再分类
- Sequence-to-Nuggets: Nested Entity Mention Detection via Anchor-Region Networks (ACL2019): 先找anchor word再向两边扩展
- Named Entity Recognition as Dependency Parsing (ACL2020)
- Boundary Enhanced Neural Span Classification for Nested Named Entity Recognition (AAAI2020): 同时分类boundary label和span label
- SpanNER: Named Entity Re-/Recognition as Span Prediction (ACL2021)
- Hypergraph-based
- Joint Mention Extraction and Classification with Mention Hypergraphs (EMNLP2015)
- Nested Named Entity Recognition Revisited (NAACL2018): 将序列构建成一个hypergraph
- Neural Segmental Hypergraphs for Overlapping Mention Recognition (EMNLP2018): hypergraph-based
- MRC
- A Unified MRC Framework for Named Entity Recognition (ACL2020)
- Coarse-to-Fine Pre-training for Named Entity Recognition (EMNLP2020): 基于MRC-NER框架做pre-trainning
- Transition-based
- A Neural Transition-based Model for Nested Mention Recognition (EMNLP2018)
- An Effective Transition-based Model for Discontinuous NER (ACL2020): seq2seq
- Sequence to Sequence
- CoNLL-2002: Introduction to the CoNLL-2002 Shared Task: Language-Independent Named Entity Recognition
- CoNLL-2003: Introduction to the CoNLL-2003 Shared Task: Language-Independent Named Entity Recognition
- OntoNotes: Towards Robust Linguistic Analysis using OntoNotes
- ACE04: Website
- ACE05: Website
- WNUT-2016: Results of the WNUT16 Named Entity Recognition Shared Task
- WNUT-2017: Results of the WNUT2017 Shared Task on Novel and Emerging Entity Recognition
- Genia: GENIA Corpus - A Semantically Annotated Corpus For Bio-textmining
- CADEC: Cadec: A corpus of adverse drug event annotations
- ShARe13: Task 1: ShARe/CLEF eHealth Evaluation Lab 2013
- ShARe14: Task 2: ShARe/CLEF eHealth Evaluation Lab 2014
Different from English sentence, Chinese sentence features different sttructure of characters and words, and does not has explicity word boundary, which makes it more challenging.
- How to better utilize the character-word lattice structure to enhance Chinese NER?
- How to accelerate the model so that we can fully utilize the parallel computation of GPUs with a high inference speed?
- Model Architecture
- Chinese NER Using Lattice LSTM (ACL2018): Lattice LSTM
- CNN-Based Chinese NER with Lexicon Rethinking (IJCAI2019): CNN + n-gram金字塔
- Leverage Lexical Knowledge for Chinese Named Entity Recognition via Collaborative Graph Network (EMNLP2019): Graph
- A Lexicon-Based Graph Neural Network for Chinese NER (EMNLP2019): Graph
- Porous Lattice Transformer Encoder for Chinese NER (COLING2020): 改进attention
- FLAT: Chinese NER Using Flat-Lattice Transformer (ACL2020): flat structure
- Entity Enhanced BERT Pre-training for Chinese NER (EMNLP2020): 利用entity信息
- Lattice-BERT: Leveraging Multi-Granularity Representations in Chinese Pre-trained Language Models (NAACL2021): 多粒度的针对中文NER的BERT预训练
- Enhanced Embedding
- An Encoding Strategy Based Word-Character LSTM for Chinese NER (NAACL2019): 除了character embedding之外还引入对应的word embedding
- A Neural Multi-digraph Model for Chinese NER with Gazetteers (ACL2019): 引入外部entity词典信息
- Simplify the Usage of Lexicon in Chinese NER (ACL2020)
- Ontonotes 4.0: Website
- MSRA: The Third International Chinese Language Processing Bakeoff: Word Segmentation and Named Entity Recognition
- Resume: Chinese NER Using Lattice LSTM
- Weibo: Named Entity Recognition for Chinese Social Media with Jointly Trained Embeddings
- CLUENER-2020: CLUENER2020: Fine-grained Named Entity Recognition Dataset and Benchmark for Chinese
The few-shot NER problem is usually modeled as a N-way-K-shot task following traditional meta-learning paradigm.
- How to precisely capture the characteristics of different entity types given few training data?
- How to handle the imbalance of entity types (the domination of O class)?
- Meta-learning
- Prototype Network
- Few-shot Slot Tagging with Collapsed Dependency Transfer and Label-enhanced Task-adaptive Projection Network (ACL2020): TapNet + Collapsed Dependency Transfer + Label Information
- Simple and Effective Few-Shot Named Entity Recognition with Structured Nearest Neighbor Learning (EMNLP2020): Collapsed Dependency Transfer + Nearset Neighbor
- Few-Shot Named Entity Recognition A Comprehensive Study (arXiv): 针对entity的pre-training + self training
- Few-shot Intent Classification and Slot Filling with Retrieved Examples (NAACL2021): Span-based
- An Enhanced Span-based Decomposition Method for Few-Shot Sequence Labeling (arXiv): span-based + 各类attention + 对O类细分
- CONTaiNER: Few-Shot Named Entity Recognition via Contrastive Learning (arXiv): 更复杂的距离计算函数
- Learning from Language Description: Low-shot Named Entity Recognition via Decomposed Framework (EMNLP2021 findings): 对O类细分
- MAML
- Prototype Network
- Transfer Learning
- Augmented Natural Language for Generative Sequence Labeling (EMNLP2020): reformulate成Seq2Seq的方式
- Frustratingly simple few-shot slot tagging (ACL2021 findings): reformulate成MRC的方式
- ConVEx Data-Efficient and Few-Shot Slot Labeling (NAACL2021): 在额外reddit语料上做预训练
- Template-Based Named Entity Recognition Using BART (ACL2021 findings): 使用template的方式遍历所有可能span和可能的entity type
- Template-free Prompt Tuning for Few-shot NER (arXiv): 对前面template-based需要enumerate all spans的改进
- LightNER: A Lightweight Generative Framework with Prompt-guided Attention for Low-resource NER (arXiv): BART + prompt
- SNIPS: Snips Voice Platform: an embedded Spoken Language Understanding system for private-by-design voice interfaces
- Few-NERD: Few-NERD: A Few-shot Named Entity Recognition Dataset
Given a document that consists of multiple sentences, document-level NER aims at recognizing entities in the whole document.
- How to better utilize the document-level context?
- How to maintain the consistency of entities occurring multiple times across the document?
- Global Attention
- An attention-based BiLSTM-CRF approach to document-level chemical named entity recognition (Bioinformatics 2018): 每个句子各自encode,但是每个单词有一个attend整个document的word做一个aggregation的document-level特征
- Improving Clinical Named Entity Recognition with Global Neural Attention (APWeb2018)
- Global Attention for Name Tagging (CoNLL2018): 每个sentence各自encode,每个单词attend跟他一样的其他位置的词
- Leveraging Multi-Token Entities in Document-Level Named Entity Recognition (AAAI2020): 引入对multi-token的entity更多地关注
- Named Entity Recognition in Multi-level Contexts (AACL2020)
- Graph
- GraphIE: A Graph-Based Framework for Information Extraction (NAACL2019): 引入word-level graph + sentence-level graph,使用图神经网络
- Exploiting Global Contextual Information for Document-level Named Entity Recognition (arXiv): 每个节点是一个word,相同word会有相应的边,使用图神经网络
- Global-level Memory
- Pooled Contextualized Embeddings for Named Entity Recognition (NAACL2019): 用memory维护word的global的表示
- Hierarchical Contextualized Representation for Named Entity Recognition (AAAI2020): sentence-level representation + 基于memory的document-level representation
- Leveraging Document-Level Label Consistency for Named Entity Recognition (IJCAI2020): 先粗糙句子级别预测每个句子,然后利用整个document各个句子的初步结果使用一个uncertainty+memory的机制进行refine得到最终结果
- Concatenate (left and right) Context
- Fast and Accurate Entity Recognition with Iterated Dilated Convolutions (EMNLP2017): sentences concat起来迭代地stack dilated CNN
- FLERT: Document-Level Features for Named Entity Recognition (arXiv)
- Multilingual is not enough: BERT for Finnish (arXiv)
- Exploring Cross-sentence Contexts for Named Entity Recognition with BERT (CoNLL2020)
- A Frustratingly Easy Approach for Entity and Relation Extraction (NAACL2021)
- CoNLL2003: Introduction to the CoNLL-2003 Shared Task: Language-Independent Named Entity Recognition
- OntoNotes 5.0: Website
- CHEMDNER: CHEMDNER: The drugs and chemical names extraction challenge
- Survey
- Bridging Resolution: A Survey of the State of the Art (COLING): 2020
- A Brief Survey and Comparative Study of Recent Development of Pronoun Coreference Resolution (arxiv): 2020
- A Neural Entity Coreference Resolution Review (arxiv): 2019
- Anaphora and coreference resolution: A review (Information Fision): 2018
- Supervised noun phrase coreference research: The first fifteen years (ACL): 2020
Coreference resolution is the task of clustering mention spans in text that refer to the same underlying real world entities.
Example:
"I", "my", and "she" belong to the same cluster and "Obama" and "he" belong to the same cluster.
In this task, mention span may be named entity mention, pronoun, verb, etc.
- How to better represent mention span?
- How to mitigate the number of enormous invalid mention spans?
- How to model contextual representation?
- How to distinguish pronoun coreference and entity coreference?
- mention span pair-wise
- Understanding the value of features for coreference resolution (EMNLP2008): first neural mention pair model with average perceptron model, focusing on advanced feature selection
- mention word pair-wise
- wl-coref Word-Level Coreference Resolution: consider coreference links between individual words rather than word spans and then reconstruct the word spans
- mention-antecedant ranking
- Learning anaphoricity and antecedent ranking features for coreference resolution (ACL2015): an adaption of the piece-wise scoring function via a feed-forward function
- e2e-coref End-to-end Neural Coreference Resolution (EMNLP2017): span enumeration + mention pair-wise scoring (mention span score + antecedant mention span score + coreference score)
- Neural coreference resolution with deep biaffine attention by joint mention detection and mention clustering (ACL2018): (1) using a biaffine attention model to get antecedent scores for each possible mention, and (2) jointly optimizing the mention detection accuracy and the mention clustering log-likelihood given the mention cluster labels
- c2f-coref Higher-order Coreference Resolution with Coarse-to-fine Inference (NAACL2018): high-order span representation + coarse-to-fine inference
- c2f-coref-BERT BERT for Coreference Resolution: Baselines and Analysis (EMNLP2019): c2f-coref + BERT encoding
- coref-hoi-SpanBERT+cluster_mergeing Revealing the Myth of Higher-Order Inference in Coreference Resolution (EMNLP2020): analysis the impact of high-order inference (HOI) and four HOI variants
- c2f-coref-EE Coreference Resolution with Entity Equalization (EMNLP2020): use equalization approach represents each mention in a cluster via an approximation of the sum of all mentions in the cluster
- graph neural network
- Improving coreference resolution by leveraging entity-centric features with graph neural networks and second-order inference (ArXiv2020): use graphs to model entities + GNN
- question answering
- CorefQA CorefQA: Coreference Resolution as Query-based Span Prediction (ACL2020): mention detection + use mention to query antecedant mention on inpu text
- cluster based
- Learning global features for coreference resolution (NAACL2016): global scoring function between the cluster assigned to each possible antecedent and the mention + local scoring function of the mention and each possible antecedent
- Improving coreference resolution by learning entity-level distributed representations (ACL2016): define each mention as a single entity cluster and combine them during inference
- Latent-Structure models
- c2f-coref Higher-order Coreference Resolution with Coarse-to-fine Inference (NAACL2018): high-order span representation + coarse-to-fine inference
- others
- Improving generalization in coreference resolution via adversarial training (SEM2019): use adversarial gradient-based training to retrain e2e-coref
- Rewarding coreference resolvers for being consistent with world knowledge (EMNLP2019): makes use of Relation Extraction systems and the “distill” multi-task reinforcement learning technique to fuse world knowledge into CR
- Incorporating context and external knowledge for pronoun coreference resolution (NAACL2019): incorporate contextual information and external knowledge to represent pronominal expression
- Knowledge-aware pronoun coreference resolution (ACL2019): leverage different types of knowledge in the format of triplet to better resolve pronoun coreference with a neural model
- Resolving gendered ambiguous pronouns with BERT (ACL2019): BERT-based approach to solving the problem of gender-balanced pronoun resolution
- Look again at the syntax: Relational graph convolutional network for gendered ambiguous pronoun resolution (ACL2019): BERT + RGCN on dependency graph
- OntoNotes5.0 (also called CoNLL 2012): OntoNotes: A Large Training Corpus for Enhanced Processing: Manual Annotation + Three Languages (English, Chinese, and Arabic)
- Mind the GAP: Mind the GAP: A Balanced Corpus of Gendered Ambiguous Pronouns: Towards Gender-balanced Coreference, Kaggle’s Gender Pronoun Resolution competition
- DWIE: DWIE: An Entity-centric Dataset for Multi-task Document-level Information Extraction: joint IE including four task: Named Entity Recognition, Coreference Resolution, Relation Extraction, and Entity Linking
- ParCor: ParCor 1.0: A parallel pronoun-coreference corpus to support statistical MT (LREC2014):
- CIC: Character identification on multiparty conversation: Identifying mentions of characters in tv shows (ACL2016): CR on ChatBots
- WikiCoref: Wikicoref: An english coreference-annotated corpus of wikipedia articles (LREC2016): unrestricted Coreference Resolution corpus with more types of coreference
- LitBank: An annotated dataset of coreference in english literature (LREC2020): cross-domain and long-distance coreference performance benchmark
- Sentence-level Relation Extraction
- Distant Supervised Relation Extraction
- Few-shot Relation Extraction
- Document-level Relation Extraction
- Survey
- Deep Neural Approaches to Relation Triplets Extraction: A Comprehensive Survey (arXiv): 2021
- Deep Neural Network Based Relation Extraction: An Overview (arXiv): 2021
- Neural relation extraction: a survey (arXiv): 2020
- More Data, More Relations, More Context and More Openness: A Review and Outlook for Relation Extraction (arXiv): 2020
- Relation Extraction Using Distant Supervision: A Survey (arXiv): 2018
- Relation Extraction : A Survey (arXiv): 2017
- A Survey of Deep Learning Methods for Relation Extraction (arXiv): 2017
- A Review of Relation Extraction (Literature review for Language and Statistics II): 2007
- Kernel methods for relation extraction (JMLR): 2003
Given a sentence, the task aims at extracting the (head entity, relation, tail entity) triples out of the sentence.
As illustrated in the figure, we extract (United States, Country_president, Trump) and (Apple Inc, Company_CEO, Tim Cook) out of the sentence.
- How to better understand the semantics of the sentence?
- How to better take advantage of the interactions between entity recognition and relation extraction?
- How to handle Single-Entity-Overlap (SEO) problem, which means two different relation triples have an entity overlap?
- How to handle Entity-Pair-Overlap (EPO) problem, which menas two different relation triples have the same entity pairs?
- How to handle the dependency and interactions between different relations?
- Sequence Labelling
- Joint Extraction of Entities and Relations Based on a Novel Tagging Scheme (ACL2017): 序列标注,tag有三部分,第一部分表示实体开头/中间/结尾,第二部分表示所属关系,第三部分表示是头实体还是尾实体
- Adversarial training for multi-context joint entity and relation extraction (EMNLP2018): 引入对抗训练
- A Novel Cascade Binary Tagging Framework for Relational Triple Extraction (ACL2020): 先序列标注出头实体,再根据头实体以及某个特定关系序列标注出尾实体
- PRGC: Potential Relation and Global Correspondence Based Joint Relational Triple Extraction (ACL2021): 同样先标头实体再标relation-specific的尾实体,改进在于先判断可能的relation,有可能出现的才去标对应的尾实体
- Sequence to Sequence
- Extracting Relational Facts by an End-to-End Neural Model with Copy Mechanism (ACL2018): 输入句子,输出提取结果序列,结果序列格式是 => r1, h1, t1, r2, h2, t2, ...
- CopyMTL: Copy Mechanism for Joint Extraction of Entities and Relations with Multi-Task Learning (AAAI2020): 同上,做了改进
- Learning the Extraction Order of Multiple Relational Facts in a Sentence with Reinforcement Learning (EMNLP2019): 提取结构本来无序但是序列生来有序,用强化学习解决这个问题
- Joint Entity and Relation Extraction with Set Prediction Networks (arXiv2020): 同样针对无序的问题,用non-autoregressive decoder变成集合预测问题
- Minimize Exposure Bias of Seq2Seq Models in Joint Entity and Relation Extraction (EMNLP2020 findings): Seq2Seq的方法time step过长导致exposure bias,所以尝试把sequential的decoder变成tree
- Effective Modeling of Encoder-Decoder Architecture for Joint Entity and Relation Extraction (AAAI2020)
- Question Answering
- Entity-Relation Extraction as Multi-turn Question Answering (ACL2019): 建模成多次问答
- Table
- Table Filling Multi-Task Recurrent Neural Network for Joint Entity and Relation Extraction (COLING2016): 使用RNN模型按预定义顺序遍历表格中cell捕捉cell之间依赖
- End-to-End Neural Relation Extraction with Global Optimization (EMNLP2017): 类似上文,但是引入syntactic信息 + 基于beam search的global optimization
- TPLinker: Single-stage Joint Extraction of Entities and Relations Through Token Pair Linking (COLING2020): 表格中的cell不是直接表示两个词之间的关系,而变成更细粒度的比如两个词是否属于同一个entity的第一个词与最后一个词、两个词是否分别是某个关系中的头实体跟尾实体的第一个词等等
- Two are Better than One: Joint Entity and Relation Extraction with Table-Sequence Encoders (EMNLP2020): 提出了table encoder以及序列encoder,table encoder内部cell会交互,table跟sequence的encoder也会交互
- UniRE: A Unified Label Space for Entity Relation Extraction (ACL2021): 实体类型跟关系类型放到同一个label space
- Graph
- A Walk-based Model on Entity Graphs for Relation Extraction (ACL2018): 每个实体是一个点构成全连接图,两个实体之间的边表示关系,该边表示同样based on其它路径表示
- Leveraging Dependency Forest for Neural Medical Relation Extraction (EMNLP2019): 利用多棵independent的dependency parsing tree构图
- Graph Neural Networks with Generated Parameters for Relation Extraction (ACL2019): 实体做节点,生成边的表示
- GraphRel: Modeling Text as Relational Graphs for Joint Entity and Relation Extraction (ACL2019): 两阶段。第一阶段dependency tree构图,第二阶段用第一阶段预测结果构图做refinement
- AGGCN Attention Guided Graph Convolutional Networks for Relation Extraction (ACL2019): 多个含有图的层,第一层用dependency tree构图,后面基于attention结果构图,最后把所有层结果利用起来
- Joint Type Inference on Entities and Relations via Graph Convolutional Networks (ACL2019): 二分图,实体在一边,关系在另一边
- Relation of the Relations: A New Paradigm of the Relation Extraction Problem (arXiv): 把实体两两之间关系也作为节点,连同实体节点一起构图
- Span-level
- Span-Level Model for Relation Extraction (ACL2019): 用span其实是为了解决nested NER
- Span-based Joint Entity and Relation Extraction with Transformer Pre-training (ECAI2020)
- Pre-training
- Matching the Blanks: Distributional Similarity for Relation Learning ( ACL2019): 假设两个句子如果含有相同实体对那么表达相同关系,句子表示应该尽量接近
- Learning from Context or Names? An Empirical Study on Neural Relation Extraction (EMNLP2020): 假设两个句子如果含有远程标注的相同关系,那么句子表示应该尽量接近
- Coreferential Reasoning Learning for Language Representation (EMNLP2020): mask掉实体,模型能够预测出句子中与该实体是共指关系的其他实体
- Pre-training Entity Relation Encoder with Intra-span and Inter-span Information (EMNLP2020): 针对intra-span和inter-span提出三种不同的预训练目标
- Others
- Relation Classification via Convolutional Deep Neural Network (COLING2014): 早期工作,CNN-based,给定实体对做关系分类
- Relation Classification via Recurrent Neural Network (arXiv): 早起工作,RNN-based,给定实体对做关系分类
- A Frustratingly Easy Approach for Entity and Relation Extraction (NAACL2021): pipeline,先找实体,再在实体左右加特殊符号做RE,加context会带来帮助
- Extracting Entities and Relations with Joint Minimum Risk Training (EMNLP2018): 使用joint minimum risk training更显式地考虑实体识别与关系识别两个任务的交互
- Recurrent Interaction Network for Jointly Extracting Entities and Classifying Relations (EMNLP2020): 实体识别跟关系识别模块像RNN在多个时刻不断地交互,增强他们之间的显式交互而不是简单共享encoder做multi-task
- Let’s Stop Incorrect Comparisons in End-to-end Relation Extraction! (EMNLP2020): 反思我们的evaluation问题
- SemEval-2010 Task8: SemEval-2010 Task 8: Multi-Way Classification of Semantic Relations Between Pairs of Nominals: manual-annotated
- ACE03/04/05: Website: manual-annotated, 包括英语/中文/阿拉伯语三种语言
- CoNLL04: A Linear Programming Formulation for Global Inference in Natural Language Tasks
- TACRED: Position-aware Attention and Supervised Data Improve Slot Filling (EMNLP2017): manual-annotated
- WebNLG: The WebNLG Challenge: Generating Text from RDF Data (INLG2017): 由Extracting Relational Facts by an End-to-End Neural Model with Copy Mechanism (ACL2018) adapt过来
- DuIE: DuIE: A Large-scale Chinese Dataset for Information Extraction: 百度发布的中文数据集
- ADE: Development of a benchmark corpus to support the automatic extraction of drug-related adverse effects from medical case reports: biomedical领域,只有一种关系
Distant supervised RE aims at annotate the unlabeled sentences with the large-scale knowledge base. If the sentence has two entities that also occur in the knowledge graph and have a specific relation, we assume that this sentence exactly expresses such relation. Obviously, it inevitably brings noise (false positive examples).
People usually formulate the problem as a bag-level RE task. A bag means multiple sentences with the same entity pair. Given a bag, we have to correctly predict the relations between such entity pair, while some sentences in the bag may be the false positive examples.
- How to better filter those false positive examples (noise) and decrease the impact of them?
- How to better make the most of the information of different sentences with the same entity pairs (or in the same bag)?
- Classical
- Distant Supervision for Relation Extraction via Piecewise Convolutional Neural Networks (EMNLP2015): PCNN + bag里边选max probability的instance作为预测结果
- Better Selection or Aggregation Methods
- Relation Extraction with Multi-instance Multi-label Convolutional Neural Networks (COLING2016): 提出跨句子max池化从而更好利用bag里边cross-sentence的信息来共同推断存在的关系
- Neural Relation Extraction with Selective Attention over Instances (ACL2016): bag里边的多个句子用attention聚合
- Large Scaled Relation Extraction with Reinforcement Learning (AAAI2018): 使用强化学习来select sentences
- DSGAN: Generative Adversarial Training for Distant Supervision Relation Extraction (ACL2018): 使用GAN来挑选positive sentence,Generator打分,Discriminator判断
- Self-Attention Enhanced CNNs and Collaborative Curriculum Learning for Distantly Supervised Relation Extraction (EMNLP2019): 使用两个网络互相约束和挑sentence
- External Knowledge
- Analysis
- Looking Beyond Label Noise: Shifted Label Distribution Matters in Distantly Supervised Relation Extraction (EMNLP2019): 在远程标注数据上训练的模型跟在人工标注数据上的区别探究
- NYT: Modeling Relations and Their Mentions without Labeled Text (ECML-PKDD2010): 远程标注得到,New York Times Corpus + Freebase
- GDS: Improving Distantly Supervised Relation Extraction using Word and Entity Based Attention: Google Distant Supervised Dataset,远程标注得到
- NYT-H: Towards Accurate and Consistent Evaluation: A Dataset for Distantly-Supervised Relation Extraction (COLING2020)
The few-shot RE problem is usually modeled as a N-way-K-shot task following traditional meta-learning paradigm. Given N relation types with K instances, the task aims at predicting which relation type the query (test) instance belongs to.
- How to precisely capture the characteristics of different relation types given few training data?
- How to better consider the interactions among instances within support set?
- How to better consider the interactions among instances in support set and the query?
- How to make our model more robust to deal with the noise?
- Hybrid Attention-Based Prototypical Networks for Noisy Few-Shot Relation Classification (AAAI2019): 聚合prototype vector的时候考虑跟query的关系进行softmax聚合 + 不同维度计算距离的时候有不同重要性
- Multi-Level Matching and Aggregation Network for Few-Shot Relation Classification (ACL2019): encode的时候考虑support set跟query的交互 + 聚合prototype vector的时候考虑query attention做softmax聚合 + 引入同类间距离要小的auxiliary loss约束
- Meta-Information Guided Meta-Learning for Few-Shot Relation Classification (COLING2020): 基于MAML进行改进提出MIML,区别主要在于MAML是为了得到最佳快速adapt到某个任务的初始参数,MIML是得到最佳的meta-parameter,这个meta-parameter能够根据某个任务生成一个能快速adapt到这个任务的初始化参数
- Few-shot Relation Extraction via Bayesian Meta-learning on Relation Graphs (ICML2020): 引入贝叶斯学习+图网络
- Label Verbalization and Entailment for Effective Zero and Few-Shot Relation Extraction (EMNLP2021): formulate成NLI任务来做,这样可以先用大量NLI数据做pre-train,将知识迁移过来
- Matching the Blanks: Distributional Similarity for Relation Learning ( ACL2019): 假设两个句子如果含有相同实体对那么表达相同关系,句子表示应该尽量接近,做预训练之后来做这个任务
- MapRE: An Effective Semantic Mapping Approach for Low-resource Relation Extraction (EMNLP2021): 考虑label的semantic information来做pre-training
- ZS-BERT: Towards Zero-Shot Relation Extraction with Attribute Representation Learning (NAACL2021): 学习两个映射,一个映射句子,一个映射关系的文本描述,这样只需要关系的文本描述就可以分类实现zero-shot
- FewRel: A Large-Scale Supervised Few-Shot Relation Classification Dataset with State-of-the-Art Evaluation (EMNLP2018)
- FewRel 2.0: Towards More Challenging Few-Shot Relation Classification (EMNLP2019)
Given a document that consists of multiple sentences, the task aims at extracting relation triples out of the document.
- How to better capture the semantics of the whole long document?
- An entity usually has many mentions across the document.
- How to handle the inter-sentence relations, which means the head entity and the tail entity do not locate in the same sentence and may be far away from each other?
- How to handle the reasoning among relations?
- Graph
- Word-level Graph (图的节点是每个token)
- Cross-Sentence N-ary Relation Extraction with Graph LSTMs (EMNLP2017): dependency tree的边、相邻词的边、相邻句子root的边 + Graph LSTM
- Inter-sentence Relation Extraction with Document-level Graph Convolutional Neural Network (ACL2019): 每个句子基于dependency tree建图,之后还有相邻句子边、相邻词边、共指边和自环边
- Coarse-to-Fine Entity Representations for Document-level Relation Extraction (arXiv): 基于dependency tree建图,同样有相邻句子、相邻词、共指、自环边 + 每个实体对在图上找各个mention之间路径再聚集起来预测
- MrGCN: Mirror Graph Convolution Network for Relation Extraction with Long-Term Dependencies (ArXiv2021): 依存图 + Graph U-Net + Entity/Mention-centric classification
- Non-word-level Graph (图的节点不是token而是mention/entity/sentence等)
- Connecting the Dots: Document-level Neural Relation Extraction with Edge-oriented Graphs (EMNLP2019): 图的节点包括mention、entity和sentence三种,启发式连边,之后与《A Walk-based Model on Entity Graphs for Relation Extraction》类似基于其他边进行边信息聚合
- Reasoning with Latent Structure Refinement for Document-Level Relation Extraction (ACL2020): 先基于dependency tree抽出关键节点以及mention node、entity node来构图,之后不断refine这个graph
- Double Graph Based Reasoning for Document-level Relation Extraction (EMNLP2020): Mention-level Graph解决跨句子依赖,Entity-level Graph解决推理
- Global-to-Local Neural Networks for Document-Level Relation Extraction (EMNLP2020): 同样图上有mention、entity和sentence,用图来捕捉跨句子交互 + 对于实体对更优的通过mention表示聚合成entity表示的方式 + 预测实体对relation的时候考虑其他实体对的关系表示
- The Dots Have Their Values: Exploiting the Node-Edge Connections in Graph-based Neural Models for Document-level Relation Extraction (EMNLP2020 findings): 基于前面《Connceting the Dots》的改进,原来只考虑边,现在把节点表示也考虑上
- Document-Level Relation Extraction with Reconstruction (AAAI2021): 图上有entity、mention、sentence节点 + 核心思路是对于存在某个关系的实体对在图上有一条meta path
- Both
- Document-level Relation Extraction with Dual-tier Heterogeneous Graph (COLING2020): 首先一个基于dependency tree的syntactic graph,之后接一个以mention、entity为节点的semantic graph
- Word-level Graph (图的节点是每个token)
- Non-graph
- Simultaneously Self-Attending to All Mentions for Full-Abstract Biological Relation Extraction (NAACL2018): Transformer+Convolution+Biaffine
- Document-Level N-ary Relation Extraction with Multiscale Representation Learning (NAACL2019): mention-level representation + entity-level representation
- Entity and Evidence Guided Document-Level Relation Extraction (repl4nlp@ACL2021): 引入evidence prediction作为auxiliary task + 编码的时候输入concat上头实体
- Document-Level Relation Extraction with Adaptive Thresholding and Localized Context Pooling (AAAI2021):引入一个辅助类来做adaptive threshold处理multi-label问题 + 每个实体对更有针对性地提取context用于关系预测
- Entity Structure Within and Throughout: Modeling Mention Dependencies for Document-Level Relation Extraction (AAAI2021): 魔改Transformer的self-attention模块,根据token跟token之间是否属于同一个句子、是否属于同一个实体、是否一个为实体一个为非实体之类的分为总共六类,每一类有不同的attention计算方法
- Multi-view Inference for Relation Extraction with Uncertain Knowledge (AAAI2021): 引入外部knowledge base协助编码
- Document-level Relation Extraction as Semantic Segmentation (IJCAI2021): 借鉴了CV中的U-Net
- Learning Logic Rules for Document-level Relation Extraction (EMNLP2021): 将关系的一些推理规则当成隐变量
- ERICA: Improving Entity and Relation Understanding for Pre-trained Language Models via Contrastive Learning (ACL2021): 基于远程标注做预训练
- Document-level Entity-based Extraction as Template Generation (EMNLP2021): Seq2Seq
- May Be We Do NOT Need All Sentences
- Three Sentences Are All You Need: Local Path Enhanced Document Relation Extraction (ACL2021): 对于每个实体对只需要最多三个句子就可以提取出来关系了
- SIRE: Separate Intra- and Inter-sentential Reasoning for Document-level Relation Extraction (ACL2021 findings): 如果实体对共现在同一个句子那么只需要用intra-sentence的表示即可,否则采用inter-sentence的表示 + 考虑多个关系的逻辑推理
- Discriminative Reasoning for Document-level Relation Extraction (ACL2021 findings): 定义三种推理路径,对于每个实体对抽这些路径来获取hidden state进行预测
- Eider: Evidence-enhanced Document-level Relation Extraction (arXiv): 使用整个文档预测的同时,引入另一个branch预测evidence sentences,然后只用这部分sentences组成伪文档进行预测,将两部分结果综合
- SAIS: Supervising and Augmenting Intermediate Steps for Document-Level Relation Extraction (arXiv): Eider的基础上,前面引入共指等更多的中间auxiliary tasks
- CDR: BioCreative V CDR task corpus: a resource for chemical disease relation extraction
- GDA: RENET: A Deep Learning Approach for Extracting Gene-Disease Associations from Literature
- SciERC: Multi-Task Identification of Entities, Relations, and Coreference for Scientific Knowledge Graph Construction (EMNLP2018)
- SciREX: SciREX: A Challenge Dataset for Document-Level Information Extraction (ACL2020), 含有四元关系
- DocRED: DocRED: A Large-Scale Document-Level Relation Extraction Dataset (ACL2019)
- HacRED: HacRED: A Large-Scale Relation Extraction Dataset Toward Hard Cases in Practical Applications (ACL2021 findings)
- CodRED: CodRED: A Cross-Document Relation Extraction Dataset for Acquiring Knowledge in the Wild (EMNLP2021)
- Dialogue-based RE: Dialogue-Based Relation Extraction (ACL2020)
- DDRel: DDRel: A New Dataset for Interpersonal Relation Classification in Dyadic Dialogues (AAAI2021)
- Sentence-level Event Extraction
- Distant-supervised Event Extraction
- Few-shot Event Extraction
- Document-level Event Extraction
- Relations Among Events
- Survey
Given a sentence, the task aims at handling four sub-tasks:
- Event Detection
- Trigger Identification: Identify the trigger that triggers a specific event (e.g., held and come in the figure).
- Event Classification: Predict what event type the trigger has triggerred (e.g., Meet and Transport in the figure).
- Event Argument Extraction
- Argument Identification: Identify potential arguments for a specific event (e.g., Thailand, 2024, and etc. in the figure).
- Argument Role Prediction: Predict what role the argument plays in the event (e.g., place, time, and etc. in the figure).
- How to better understand the semantics of the sentence?
- How to deal with the error propagation problems?
- How to capture the dependency and interaction between different events?
- How to capture the dependency and interaction between different arguments of the same event?
- Classic
- Event Extraction via Dynamic Multi-Pooling Convolutional Neural Network (ACL2015): pipeline-based + PCNN
- Joint Event Extraction via Recurrent Neural Networks (NAACL2016): joint + RNN + memory记录已经预测过的事件从而捕捉依赖关系
- Sequence Labelling
- Exploring Pre-trained Language Models for Event Extraction and Generation (ACL2019): joint + BERT 来做sequence labelling + 类别平衡reweight loss + 数据增强
- One for All: Neural Joint Modeling of Entities and Events (AAAI2019): joint
- Biomedical Event Extraction as Sequence Labeling (EMNLP2020)
- CasEE: A Joint Learning Framework with Cascade Decoding for Overlapping Event Extraction (ACL2021 findings): sequence labelling, 先标trigger,然后given trigger and a specific role去标出可能的argument
- Graph
- Jointly multiple events extraction via attention-based graph information aggregation (EMNLP2018): joint + dependency tree构图GNN
- Jointly Extracting Event Triggers and Arguments by Dependency-Bridge RNN and Tensor-Based Argument Interaction (AAAI2018): 引入dependency tree structure但依然是RNN架构
- Biomedical Event Extraction Based on Knowledge-driven Tree-LSTM (NAACL2019): 基于dependency的Tree-LSTM + 引入external knowledge
- GATE: Graph Attention Transformer Encoder for Cross-lingual Relation and Event Extraction (AAAI2021): Graph
- Question Answering
- Event Extraction as Machine Reading Comprehension (EMNLP2020)
- Event Extraction by Answering (Almost) Natural Questions (EMNLP2020)
- Event Extraction as Multi-turn Question Answering (EMNLP2020 findings): reformulate成QA
- Sequence to Sequence
- Text2Event: Controllable Sequence-to-Structure Generation for End-to-end Event Extraction (ACL2021): reformulate成Seq2Seq
- Others
- Joint Event Extraction with Hierarchical Policy Network (COLING2020): 使用强化学习引入层次化policy network
- Extracting entities and events as a single task using a transition-based neural model (IJCAI2019): 使用动作序列解决
- CLEVE: Contrastive Pre-training for Event Extraction (ACL2021): 基于AMR parsing结果做对比学习与训练
- Focus on Event Detection
- Document-level Context
- Collective Event Detection via a Hierarchical and Bias Tagging Networks with Gated Multi-level Attention Mechanisms (EMNLP2018): 利用document信息 + 利用不同事件信息
- Document Embedding Enhanced Event Detection with Hierarchical and Supervised Attention (ACL2018): 引入document-level信息
- Graph
- Event Detection with Multi-Order Graph Convolution and Aggregated Attention (EMNLP2019): dependency tree做一阶图,加上更高阶的图,做GNN
- Event Detection with Relation-Aware Graph Convolutional Networks (EMNLP2020 findings): dependency tree构图并考虑边的类型
- External Knowledge
- Improving Event Detection via Open-domain Trigger Knowledge (ACL2020): 引入trigger相关的外部知识 + unlabled data
- Exploiting the Ground-Truth: An Adversarial Imitation Based Knowledge Distillation Approach for Event Detection (AAAI2019): 引入外部知识
- Distilling Discrimination and Generalization Knowledge for Event Detection via Delta-Representation Learning (ACL2019)
- Others
- Event Detection without Triggers (NAACL2019): Trigger对于事件不是必要的
- Incremental Event Detection via Knowledge Consolidation Networks (EMNLP2020): 不断有新事件类型进来
- How Does Context Matter? On the Robustness of Event Detection with Context-Selective Mask Generalization (EMNLP2020 findings): 避免对trigger的过度依赖
- Document-level Context
- Focus on Event Argument Extraction
- HMEAE: Hierarchical Modular Event Argument Extraction (EMNLP2019): 利用argument role之间所属大类别的内在关系
- Resource-Enhanced Neural Model for Event Argument Extraction (EMNLP2020 findings): self-training + dependency tree捕捉长距离依赖
- Capturing Event Argument Interaction via A Bi-Directional Entity-Level Recurrent Decoder (ACL2021): 预测两次从而更好捕捉argument role之间关系
- What the Role is vs. What Plays the Role: Semi-Supervised Event Argument Extraction via Dual Question Answering (AAAI2021): MRC + 模型标数据
- ACE05: Website: manual-annotated, 包括英语/中文/阿拉伯语三种语言
- TAC KBP 2015: Website
- BioNLP Genia 2011: Overview of Genia Event Task in BioNLP Shared Task 2011
- MAVEN: MAVEN: A Massive General Domain Event Detection Dataset
Distant-supervised EE annotate the unlabeled sentence with the help of the large-scale knowledge base. As illustrated in the figure, we annotate the sentence from Wikipedia with the knowledge in Freebase.
- How to deal with the noise brought by distant supervision?
- Event Extraction Using Distant Supervision (LREC2014)
- Automatically Labeled Data Generation for Large Scale Event Extraction (ACL2017)
- Scale Up Event Extraction Learning via Automatic Training Data Generation (AAAI2018)
- KnowDis: Knowledge Enhanced Data Augmentation for Event Causality Detection via Distant Supervision (COLING2020)
- Automatic Labeling of Tweets for Crisis Response Using Distant Supervision (WWW2020)
Few-shot Event Extraction is usually formulated as a N-way-K-shot meta-learning problem. Given N event types with K instances, the task aims at predicting which event type the query instance belongs to.
- How to precisely capture the characteristics of different relation types given few training data?
- How to better consider the interactions among instances within support set?
- How to better consider the interactions among instances in support set and the query?
- How to address the trigger bias and avoid overfitting?
- Extensively Matching for Few-shot Learning Event Detection (nuse@ACL2020)
- Exploiting the Matching Information in the Support Set for Few Shot Event Classification (PAKDD2020)
- Meta-Learning with Dynamic-Memory-Based Prototypical Network for Few-Shot Event Detection (WSDM2020)
- Few-Shot Event Detection with Prototypical Amortized Conditional Random Field (ACL2021 findings)
- Graph Learning Regularization and Transfer Learning for Few-Shot Event Detection (SIGIR2021)
- Taxonomy-aware Learning for Few-Shot Event Detection (WWW2021)
- Analysis
- ACE05: Website
- FewEvent: Meta-Learning with Dynamic-Memory-Based Prototypical Network for Few-Shot Event Detection
Given a document that consists of multiple sentences, the task aims at extracting events out of the whole document.
- How to better capture the semantics of the whole long document?
- How to handle the cross-sentence events, which means the event arguments scattering across different sentences?
- How to capture the interdependency among different events and different argument roles?
- DCFEE: A Document-level Chinese Financial Event Extraction System based on Automatically Labeled Training Data (ACL2018 demo)
- Doc2EDAG: An End-to-End Document-level Framework for Chinese Financial Event Extraction (EMNLP2019)
- Document-level Event Extraction via Heterogeneous Graph-based Interaction Model with a Tracker (ACL2021)
- Document-level Event Extraction via Parallel Prediction Networks (ACL2021)
- Revisiting the Evaluation of End-to-end Event Extraction (ACL2021 findings)
- Exploring Sentence Community for Document-Level Event Extraction (EMNLP2021 findings)
- Template Filling with Generative Transformers (NAACL2021)
- Focus on Event Argument Extraction
- Document-Level Event Role Filler Extraction using Multi-Granularity Contextualized Encoding (ACL2020)
- GRIT: Generative Role-filler Transformers for Document-level Event Entity Extraction (EACL2021)
- [- Document-level Entity-based Extraction as Template Generation (EMNLP2021): Seq2Seq
- Multi-Sentence Argument Linking (ACL2020)
- A Two-Step Approach for Implicit Event Argument Detection (ACL2020)
- Inducing Rich Interaction Structures Between Words for Document-Level Event Argument Extraction (PAKDD2021)
- Document-Level Event Argument Extraction by Conditional Generation (NAACL2019)
- Trigger is Not Sufficient: Exploiting Frame-aware Knowledge for Implicit Event Argument Extraction (ACL2021)
- Machine Reading Comprehension as Data Augmentation: A Case Study on Implicit Event Argument Extraction (EMNLP2021)
- Doc2EDAG: Doc2EDAG: An End-to-End Document-level Framework for Chinese Financial Event Extraction
- MUC-4: Fourth Message Uunderstanding Conference (MUC-4): Proceedings of a Conference Held in McLean, Virginia, June 16-18, 1992
- RAMS: Multi-Sentence Argument Linking (ACL2020)
- WikiEvents: Document-Level Event Argument Extraction by Conditional Generation
Given a text, the tasks aims at predicting the relation among different events that are expressed by the text. The relation types usually include coreference, causality, temporal, sub-event, and etc.
- How to better understand the relations between different events?
- How to better consider the global constraints between different event relations?
- Coreference
- Graph-based Event Coreference Resolution
- Joint Inference for Event Coreference Resolution (COLING2016)
- Joint Learning for Event Coreference Resolution (ACL2017)
- Event Coreference Resolution by Iteratively Unfolding Inter-dependencies among Events (EMNLP2017)
- Improving Event Coreference Resolution by Modeling Correlations between Event Coreference Chains and Document Topic Structures (ACL2018)
- Revisiting Joint Modeling of Cross-document Entity and Event Coreference Resolution (ACL2019)
- Improving Event Coreference Resolution by Learning Argument Compatibility from Unlabeled Data (NAACL2019)
- Causality
- Modeling Document-level Causal Structures for Event Causal Relation Identification (NAACL2018)
- Event Causality Recognition Exploiting Multiple Annotators’ Judgments and Background Knowledge (EMNLP2019)
- Towards Causal Explanation Detection with Pyramid Salient-Aware Network (CCL2020)
- Knowledge Enhanced Event Causality Identification with Mention Masking Generalizations (IJCAI2020)
- KnowDis: Knowledge Enhanced Data Augmentation for Event Causality Detection via Distant Supervision (COLING2020)
- Temporal
- Neural Temporal Relation Extraction (EACL2017): 神经网络早期工作
- Neural Architecture for Temporal Relation Extraction: A Bi-LSTM Approach for Detecting Narrative Containers (ACL2017)
- Classifying Temporal Relations by Bidirectional LSTM over Dependency Paths (ACL2017): 引入dependency tree
- A Structured Learning Approach to Temporal Relation Extraction (EMNLP2017): 整数规划
- Context-Aware Neural Model for Temporal Information Extraction (ACL2018): memory
- Joint Reasoning for Temporal and Causal Relations (ACL2018): 借助事件因果关系
- Improving Temporal Relation Extraction with a Globally Acquired Statistical Resource (NAACL2018)
- Temporal Information Extraction by Predicting Relative Time-lines (EMNLP2018): 从pair-wise角度转变成map到time-line
- Domain Knowledge Empowered Structured Neural Net for End-to-End Event Temporal Relation Extraction (EMNLP2020)
- Joint Constrained Learning for Event-Event Relation Extraction (EMNLP2020): 借助事件间子事件关系
- TimeBank 1.2: website
- TimeBank-Dense: Dense Event Ordering with a Multi-Pass Architecture
- MATRES: A Multi-Axis Annotation Scheme for Event Temporal Relations
- TORGUE: TORQUE: A Reading Comprehension Dataset of Temporal Ordering Questions (EMNLP2020): 用MRC方式
- ESTER: ESTER: A Machine Reading Comprehension Dataset for Reasoning about Event Semantic Relations (EMNLP2021): 多种事件间关系 + MRC方式
- FinReason: Probing into the Root: A Dataset for Reason Extraction of Structural Events from Financial Documents
- EventCorefBank: Joint Entity and Event Coreference Resolution across Documents
- []
- surveys
- todos
Joint Information Extraction aims at handling multiple extraction tasks simultaneously, including named entity recognition, relation extraction, event extraction, and etc. As illustrated in the figure, the entities are recognized, with the relation triples (man, ART, taxicab), (man, PHYS, checkpoint), (soldiers, PHYS, checkpoint) and the Transport events with Artifact and Destination arguments.
- How to achieve satisfactory performance for each sub-tasks?
- How to make the most of the interactions among different sub-tasks?
- How to derive a globally optimal extraction results, with consideration of some global constraints?
-
Graph
- A General Framework for Information Extraction using Dynamic Span Graphs (NAACL2019): NER + coreference + RE
- Entity, Relation, and Event Extraction with Contextualized Span Representations (EMNLP2019): NER + coreference + RE + EE
- A Joint Neural Model for Information Extraction with Global Features (ACL2020): 引入global的约束
- Abstract Meaning Representation Guided Graph Encoding and Decoding for Joint Information Extraction (NAACL2021): 借助AMR parsing的信息
- Cross-Task Instance Representation Interactions and Label Dependencies for Joint Information Extraction with Graph Convolutional Networks (NAACL2021)
- Cross-lingual Structure Transfer for Relation and Event Extraction (EMNLP2019): universal dependency parsing. + 跨语言
-
Sequence to Sequence
- ACE05: Website: manual-annotated, 包括英语/中文/阿拉伯语三种语言
- Light ERE: A Comparison of the Events and Relations Across ACE, ERE, TAC-KBP, and FrameNet Annotation Standards
- Rich ERE: From Light to Rich ERE: Annotation of Entities, Relations, and Events
- DWIE: DWIE: An Entity-centric Dataset for Multi-task Document-level Information Extraction
- Open-domain Information Extraction
- Type/Schema Induction
- ...