SSAM: a span spatial attention model for recognizing named entities
SSAM: a span spatial attention model for recognizing named entities
Blog Article
Abstract Mapping a sentence into a two-dimensional (2D) representation can flatten nested semantic structures and build multi-granular span dependencies in named entity recognition.Existing approaches to recognizing named entities often classify each entity span independently, which ignores the spatial structures between neighboring spans.To address this issue, we propose a Span Spatial Attention Model (SSAM) that consists of a token encoder, a span generation module, and Cocktail Shakers a 2D spatial attention network.
The SSAM employs a two-channel span generation strategy to capture multi-granular features.Unlike traditional attention implemented on a sequential sentence representation, spatial attention is applied to a 2D Kitchen Cart sentence representation, enabling the model to learn the spatial structures of the sentence.This allows the SSAM to adaptively encode important features and suppress non-essential information in the 2D sentence representation.
Experimental results on the GENIA, ACE2005, and ACE2004 datasets demonstrate that our proposed model achieves state-of-the-art performance, with F1-scores of 81.82%, 89.04%, and 89.
24%, respectively.The code is available at https://github.com/Gzuwkj/SpatialAttentionForNer.