Skip to main content
Log in

End-to-End Entity Detection with Proposer and Regressor

  • Published:
Neural Processing Letters Aims and scope Submit manuscript

Abstract

Named entity recognition is a traditional task in natural language processing. In particular, nested entity recognition receives extensive attention for the widespread existence of the nesting scenario. The latest research migrates the well-established paradigm of set prediction in object detection to cope with entity nesting. However, the manual creation of query vectors, which fail to adapt to the rich semantic information in the context, limits these approaches. An end-to-end entity detection approach with proposer and regressor is presented in this paper to tackle the issues. First, the proposer utilizes the feature pyramid network to generate high-quality entity proposals. Then, the regressor refines the proposals for generating the final prediction. The model adopts encoder-only architecture and thus obtains the advantages of the richness of query semantics, high precision of entity localization, and easiness of model training. Moreover, we introduce the novel spatially modulated attention and progressive refinement for further improvement. Extensive experiments demonstrate that our model achieves advanced performance in flat and nested NER, achieving a new state-of-the-art F1 score of 80.74 on the GENIA dataset and 72.38 on the WeiboNER dataset.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Subscribe and save

Springer+ Basic
$34.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6

Similar content being viewed by others

Explore related subjects

Discover the latest articles, news and stories from top researchers in related subjects.

Code Availability

The code is available at https://github.com/Rosenberg37/EntityDetection.

References

  1. Wang Q, Zhang Q, Zuo M, He S, Zhang B (2022) A entity relation extraction model with enhanced position attention in food domain. Neural Process Lett 54(2):1449–1464. https://doi.org/10.1007/s11063-021-10690-9

    Article  Google Scholar 

  2. Zhang L, Zhang S, Balog K (2019) Table2vec: Neural word and entity embeddings for table population and retrieval. In: Proceedings of the 42nd international ACM SIGIR conference on research and development in information retrieval. SIGIR’19. Association for Computing Machinery, New York, NY, pp 1029–1032 . https://doi.org/10.1145/3331184.3331333

  3. Chen H, Wei B, Liu Y, Li Y, Yu J, Zhu W (2018) Bilinear joint learning of word and entity embeddings for entity linking. Neurocomputing 294:12–18. https://doi.org/10.1016/j.neucom.2017.11.064

    Article  Google Scholar 

  4. Liu H, Song J, Peng W, Sun J, Xin X (2022) TFM: a triple fusion module for integrating lexicon information in Chinese named entity recognition. Neural Process Lett. https://doi.org/10.1007/s11063-022-10768-y

    Article  Google Scholar 

  5. Yan R, Jiang X, Dang D (2021) Named entity recognition by using XLNet-BiLSTM-CRF. Neural Process Lett 53(5):3339–3356. https://doi.org/10.1007/s11063-021-10547-1

    Article  Google Scholar 

  6. Shen Y, Ma X, Tan Z, Zhang S, Wang W, Lu W (2021) Locate and label: A two-stage identifier for nested named entity recognition. In: Proceedings of the 59th annual meeting of the association for computational linguistics and the 11th international joint conference on natural language processing, vol 1, (Long Papers). Association for Computational Linguistics, Online, pp 2782–2794. https://doi.org/10.18653/v1/2021.acl-long.216

  7. Ohta T, Tateisi Y, Kim JD (2002) The genia corpus: An annotated research abstract corpus in molecular biology domain. In: Proceedings of the second international conference on human language technology research. HLT ’02. Morgan Kaufmann Publishers Inc., San Francisco, CA, pp 82–86

  8. Lin T-Y, Maire M, Belongie S, Hays J, Perona P, Ramanan D, Dollár P, Zitnick CL (2014) Microsoft coco: common objects in context. In: Fleet D, Pajdla T, Schiele B, Tuytelaars T (eds) Computer vision - ECCV 2014. Springer, Cham, pp 740–755

    Chapter  Google Scholar 

  9. Finkel JR, Manning CD (2009) Nested named entity recognition. In: Proceedings of the 2009 conference on empirical methods in natural language processing. Association for Computational Linguistics, Singapore, pp 141–150. https://aclanthology.org/D09-1015

  10. Huang H, Lei M, Feng C (2021) Hypergraph network model for nested entity mention recognition. Neurocomputing 423:200–206. https://doi.org/10.1016/j.neucom.2020.09.077

    Article  Google Scholar 

  11. Alex B, Haddow B, Grover C (2007) Recognising nested named entities in biomedical text. In: Biological, translational, and clinical language processing. Association for Computational Linguistics, Prague, pp 65–72. https://aclanthology.org/W07-1009

  12. Fisher J, Vlachos A (2019) Merge and label: a novel neural network architecture for nested NER. In: Proceedings of the 57th annual meeting of the association for computational linguistics. Association for Computational Linguistics, Florence, pp 5840–5850. https://doi.org/10.18653/v1/P19-1585. https://aclanthology.org/P19-1585

  13. Ju M, Miwa M, Ananiadou S (2018) A neural layered model for nested named entity recognition. In: Proceedings of the 2018 conference of the North American chapter of the association for computational linguistics: human language technologies, vol 1 (Long Papers). Association for Computational Linguistics, New Orleans, Louisiana, pp 1446–1459. https://doi.org/10.18653/v1/N18-1131. https://aclanthology.org/N18-1131

  14. Xu M, Jiang H, Watcharawittayakul S (2017) A local detection approach for named entity recognition and mention detection. In: Proceedings of the 55th annual meeting of the association for computational linguistics, vol 1 (Long Papers). Association for Computational Linguistics, Vancouver, pp 1237–1247. https://doi.org/10.18653/v1/P17-1114. https://aclanthology.org/P17-1114

  15. Sohrab MG, Miwa M (2018) Deep exhaustive model for nested named entity recognition. In: Proceedings of the 2018 conference on empirical methods in natural language processing. Association for Computational Linguistics, Brussels, pp 2843–2849. https://doi.org/10.18653/v1/D18-1309. https://aclanthology.org/D18-1309

  16. Tan C, Qiu W, Chen M, Wang R, Huang F (2020) Boundary enhanced neural span classification for nested named entity recognition. In: AAAI . https://doi.org/10.1609/aaai.v34i05.6434

  17. Shen Y, Wang X, Tan Z, Xu G, Xie P, Huang F, Lu W, Zhuang Y (2022) Parallel instance query network for named entity recognition. In: Proceedings of the 60th annual meeting of the association for computational linguistics. Association for Computational Linguistics. arxiv:2203.10545

  18. Sui D, Chen Y, Liu K, Zhao J, Zeng X, Liu S (2020) Joint entity and relation extraction with set prediction networks. Computation and Language. arXiv:2011.01675

  19. Liu M, Tu Z, Zhang T, Su T, Xu X, Wang Z (2022) LTP: a new active learning strategy for CRF-based named entity recognition. Neural Process Lett. https://doi.org/10.1007/s11063-021-10737-x

    Article  Google Scholar 

  20. Wang W, Bao F, Gao G (2019) Learning morpheme representation for Mongolian named entity recognition. Neural Process Lett 50(3):2647–2664. https://doi.org/10.1007/s11063-019-10044-6

    Article  Google Scholar 

  21. Metke-Jimenez A, Karimi S (2016) Concept identification and normalisation for adverse drug event discovery in medical forums. In: Proceedings of the first international workshop on biomedical data integration and discovery (BMDID 2016). http://ceur-ws.org/Vol-1709/BMDID_2016_paper_5.pdf

  22. Wang J, Shou L, Chen K, Chen G (2020) Pyramid: A layered model for nested named entity recognition. In: Proceedings of the 58th annual meeting of the association for computational linguistics. Association for Computational Linguistics, Online, pp 5918–5928. https://doi.org/10.18653/v1/2020.acl-main.525. https://aclanthology.org/2020.acl-main.525

  23. Yan H, Gui T, Dai J, Guo Q, Zhang Z, Qiu X (2021) A unified generative framework for various NER subtasks. In: Proceedings of the 59th annual meeting of the association for computational linguistics and the 11th international joint conference on natural language processing, vol 1, (Long Papers). Association for Computational Linguistics, Online, pp 5808–5822. https://doi.org/10.18653/v1/2021.acl-long.451. https://aclanthology.org/2021.acl-long.451

  24. Tan C, Qiu W, Chen M, Wang R, Huang F (2020) Boundary enhanced neural span classification for nested named entity recognition. In: National conference on artificial intelligence . https://doi.org/10.1609/aaai.v34i05.6434

  25. Li F, Wang Z, Hui SC, Liao L, Zhu X, Huang H (2021) A segment enhanced span-based model for nested named entity recognition. Neurocomputing 465:26–37. https://doi.org/10.1016/j.neucom.2021.08.094

    Article  Google Scholar 

  26. Manning C (2016) Connecting images and natural language a dissertation submitted to the department of computer science and the committee on graduate studies of Stanford university in partial fulfillment of the requirements for the degree of doctor of philosophy

  27. Tan Z, Shen Y, Zhang S, Lu W, Zhuang Y (2021) A sequence-to-set network for nested named entity recognition. In: Zhou, Z.-H. (ed.) Proceedings of the thirtieth international joint conference on artificial intelligence, IJCAI-21. International Joint Conferences on Artificial Intelligence Organization, pp 3936–3942. https://doi.org/10.24963/ijcai.2021/542. Main Track

  28. Carion N, Massa F, Synnaeve G, Usunier N, Kirillov A, Zagoruyko S (2020) End-to-end object detection with transformers. In: Computer vision-ECCV 2020: 16th European conference, Glasgow, UK, August 23-28, 2020, proceedings, Part I. Springer, Berlin, Heidelberg, pp 213–229. https://doi.org/10.1007/978-3-030-58452-8_13

  29. Sun Z, Cao S, Yang Y, Kitani KM (2020) Rethinking transformer-based set prediction for object detection. arXiv: Computer Vision and Pattern Recognition . https://doi.org/10.48550/arXiv.2011.10881

  30. Lin TY, Dollár P, Girshick R, He K, Hariharan B, Belongie S (2016) Feature pyramid networks for object detection. arXiv: Computer Vision and Pattern Recognition . https://doi.org/10.48550/arXiv.1612.03144

  31. Gao P, Zheng M, Wang X, Dai J, Li H (2021) Fast convergence of detr with spatially modulated co-attention. In: 2021 IEEE/CVF international conference on computer vision (ICCV), pp 3601–3610. https://doi.org/10.1109/ICCV48922.2021.00360

  32. Zhu X, Su W, Lu L, Li B, Wang X, Dai J (2021) Deformable detr: Deformable transformers for end-to-end object detection. In: International conference on learning representations . https://doi.org/10.48550/arXiv.2010.04159

  33. Cho K, Van Merriënboer B, Gulcehre C, Bahdanau D, Bougares F, Schwenk H, Bengio Y (2014) Learning phrase representations using rnn encoder-decoder for statistical machine translation. arXiv preprint arXiv:1406.1078

  34. Kim Y (2014) Convolutional neural networks for sentence classification. In: Proceedings of the 2014 conference on empirical methods in natural language processing (EMNLP). Association for Computational Linguistics, Doha, Qatar, pp 1746–1751. https://doi.org/10.3115/v1/D14-1181. https://aclanthology.org/D14-1181

  35. Devlin J, Chang MW, Lee K, Toutanova K (2019) BERT: Pre-training of deep bidirectional transformers for language understanding. In: Proceedings of the 2019 conference of the North American chapter of the association for computational linguistics: human language technologies, vol 1, (Long and Short Papers). Association for Computational Linguistics, Minneapolis, Minnesota, pp 4171–4186. https://doi.org/10.18653/v1/N19-1423. https://aclanthology.org/N19-1423

  36. Wang Y, Li Y, Tong H, Zhu Z (2020) HIT: Nested named entity recognition via head-tail pair and token interaction. In: Proceedings of the 2020 conference on empirical methods in natural language processing (EMNLP). Association for Computational Linguistics, Online, pp 6027–6036. https://doi.org/10.18653/v1/2020.emnlp-main.486. https://aclanthology.org/2020.emnlp-main.486

  37. Wu Y, Schuster M, Chen Z, Le QV, Norouzi M, Macherey W, Krikun M, Cao Y, Gao Q, Macherey K, et al. (2016) Google’s neural machine translation system: bridging the gap between human and machine translation. arXiv preprint arXiv:1609.08144 . https://doi.org/10.48550/arXiv.1609.08144

  38. Pennington J, Socher R, Manning C (2014) GloVe: Global vectors for word representation. In: Proceedings of the 2014 conference on empirical methods in natural language processing (EMNLP). Association for Computational Linguistics, Doha, Qatar, pp 1532–1543. https://doi.org/10.3115/v1/D14-1162. https://aclanthology.org/D14-1162

  39. Ba J, Kiros JR, Hinton GE (2016) Layer normalization. arXiv: Machine Learning . https://doi.org/10.48550/arXiv.1607.06450

  40. Hendrycks D, Gimpel K (2016) Gaussian error linear units (gelus). arXiv preprint arXiv:1606.08415 . https://doi.org/10.48550/arXiv.1606.08415

  41. Jiang Z, Yu W, Zhou D, Chen Y, Feng J, Yan S (2020) Convbert: improving bert with span-based dynamic convolution. In: Neural information processing systems . https://doi.org/10.48550/arXiv.2008.02496

  42. Al-Rfou R, Choe D, Constant N, Guo M, Jones L (2018) Character-level language modeling with deeper self-attention. arXiv: Computation and Language . https://doi.org/10.48550/arXiv.1808.04444

  43. Vaswani A, Shazeer N, Parmar N, Uszkoreit J, Jones L, Gomez AN, Kaiser L, Polosukhin I (2017) Attention is all you need. arXiv . https://doi.org/10.48550/arXiv.1706.03762

  44. Cho K, van Merriënboer B, Gulcehre C, Bahdanau D, Bougares F, Schwenk H, Bengio Y (2014) Learning phrase representations using RNN encoder–decoder for statistical machine translation. In: Proceedings of the 2014 conference on empirical methods in natural language processing (EMNLP). Association for Computational Linguistics, Doha, Qatar, pp 1724–1734. https://doi.org/10.3115/v1/D14-1179. https://aclanthology.org/D14-1179

  45. Zhu E, Li J (2022) Boundary smoothing for named entity recognition. https://doi.org/10.48550/arXiv.2204.12031

  46. Yu J, Bohnet B, Poesio M (2020) Named entity recognition as dependency parsing. In: Proceedings of the 58th annual meeting of the association for computational linguistics. Association for Computational Linguistics, Online, pp 6470–6476. https://doi.org/10.18653/v1/2020.acl-main.577. https://aclanthology.org/2020.acl-main.577

  47. Kuhn HW (1955) The Hungarian method for the assignment problem. Naval Res Logist Q 2:83–97. https://doi.org/10.1002/nav.3800020109

    Article  MathSciNet  MATH  Google Scholar 

  48. Tjong Kim Sang EF, De Meulder F (2003) Introduction to the CoNLL-2003 shared task: language-independent named entity recognition. In: Proceedings of the seventh conference on natural language learning at HLT-NAACL 2003, pp. 142–147. https://aclanthology.org/W03-0419

  49. Peng N, Dredze M (2015) Named entity recognition for Chinese social media with jointly trained embeddings. In: Proceedings of the 2015 conference on empirical methods in natural language processing. Association for Computational Linguistics, Lisbon, Portugal, pp 548–554. https://doi.org/10.18653/v1/D15-1064. https://aclanthology.org/D15-1064

  50. Li X, Yan H, Qiu X, Huang X (2020)FLAT: Chinese NER using flat-lattice transformer. In: Proceedings of the 58th annual meeting of the association for computational linguistics. Association for Computational Linguistics, Online, pp 6836–6842. https://doi.org/10.18653/v1/2020.acl-main.611. https://aclanthology.org/2020.acl-main.611

  51. Wolf T, Debut L, Sanh V, Chaumond J, Delangue C, Moi A, Cistac P, Rault T, Louf R, Funtowicz M, Brew J (2019) Huggingface’s transformers: state-of-the-art natural language processing. arXiv: Computation and Language . https://doi.org/10.48550/arXiv.1910.03771

  52. Lee J, Yoon W, Kim S, Kim D, Kim S, So CH, Kang J (2019) BioBERT: a pre-trained biomedical language representation model for biomedical text mining. Bioinformatics 36(4):1234–1240. https://doi.org/10.1093/bioinformatics/btz682

    Article  Google Scholar 

  53. Chiu B, Crichton G, Korhonen A, Pyysalo S (2016)How to train good word embeddings for biomedical NLP. In: Proceedings of the 15th workshop on biomedical natural language processing. Association for Computational Linguistics, Berlin, Germany, pp 166–174. https://doi.org/10.18653/v1/W16-2922. https://aclanthology.org/W16-2922

  54. Li S, Zhao Z, Hu R, Li W, Liu T, Du X (2018)Analogical reasoning on Chinese morphological and semantic relations. In: Proceedings of the 56th annual meeting of the association for computational linguistics, vol 2 (Short Papers). Association for Computational Linguistics, Melbourne, Australia, pp 138–143. https://doi.org/10.18653/v1/P18-2023. https://aclanthology.org/P18-2023

  55. Yamada I, Asai A, Shindo H, Takeda H, Matsumoto Y (2020) LUKE: Deep contextualized entity representations with entity-aware self-attention. In: Proceedings of the 2020 conference on empirical methods in natural language processing (EMNLP). Association for Computational Linguistics, Online, pp 6442–6454. https://doi.org/10.18653/v1/2020.emnlp-main.523. https://aclanthology.org/2020.emnlp-main.523

  56. Li X, Feng J, Meng Y, Han Q, Wu F, Li J (2020) A unified MRC framework for named entity recognition. In: Proceedings of the 58th annual meeting of the association for computational linguistics. Association for Computational Linguistics, Online, pp 5849–5859. https://doi.org/10.18653/v1/2020.acl-main.519. https://aclanthology.org/2020.acl-main.519

  57. Yuan H, Yuan Z, Gan R, Zhang J, Xie Y, Yu S (2022) BioBART: pretraining and evaluation of a biomedical generative language model. arXiv . https://doi.org/10.48550/ARXIV.2204.03905. https://arxiv.org/abs/2204.03905

  58. Wang Y, Shindo H, Matsumoto Y, Watanabe T (2021) Nested named entity recognition via explicitly excluding the influence of the best path. In: Proceedings of the 59th annual meeting of the association for computational linguistics and the 11th international joint conference on natural language processing, vol 1 (Long Papers). Association for Computational Linguistics, Online, pp 3547–3557. https://doi.org/10.18653/v1/2021.acl-long.275. https://aclanthology.org/2021.acl-long.275

  59. Wang S, Li X, Meng Y, Zhang T, Ouyang R, Li J, Wang G (2022) \(k\)NN-NER: named entity recognition with nearest neighbor search . https://doi.org/10.48550/arXiv.2203.17103

  60. Hu D, Wei L (2020) Slk-ner: Exploiting second-order lexicon knowledge for Chinese ner. In: Software engineering and knowledge engineering . https://doi.org/10.48550/arXiv.2007.08416

  61. Chen C, Kong F (2021) Enhancing entity boundary detection for better Chinese named entity recognition. In: Proceedings of the 59th annual meeting of the association for computational linguistics and the 11th international joint conference on natural language processing, vol 2, (Short Papers). Association for Computational Linguistics, Online, pp 20–25. https://doi.org/10.18653/v1/2021.acl-short.4. https://aclanthology.org/2021.acl-short.4

  62. Sun Z, Li X, Sun X, Meng Y, Ao X, He Q, Wu F, Li J (2021) ChineseBERT: Chinese pretraining enhanced by glyph and Pinyin information. In: Proceedings of the 59th annual meeting of the association for computational linguistics and the 11th international joint conference on natural language processing, vol 1, (Long Papers). Association for Computational Linguistics, Online, pp 2065–2075. https://doi.org/10.18653/v1/2021.acl-long.161. https://aclanthology.org/2021.acl-long.161

  63. Nie Y, Tian Y, Song Y, Ao X, Wan X (2020) Improving named entity recognition with attentive ensemble of syntactic information. In: Empirical methods in natural language processing . https://doi.org/10.48550/arXiv.2010.15466

  64. Teed Z, Deng J (2022) Raft: recurrent all-pairs field transforms for optical flow. https://doi.org/10.48550/arXiv.2003.12039

Download references

Acknowledgements

This work was supported by the National Natural Science Foundation of China under Grant Nos. 62072211, 51939003 and U20A20285.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Yu Jiang.

Ethics declarations

Conflict of interest

The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Wen, X., Zhou, C., Tang, H. et al. End-to-End Entity Detection with Proposer and Regressor. Neural Process Lett 55, 9269–9294 (2023). https://doi.org/10.1007/s11063-023-11201-8

Download citation

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s11063-023-11201-8

Keywords

Navigation