“…Deep learning techniques, particularly recurrent neural networks (RNNs) and recently, transformer-based architectures like BERT (Devlin et al, 2019) and GPT (Radford et al, 2019), revolutionized NER. These models leverage contextual embeddings to capture intricate relationships and dependencies, achieving state-of-the-art results in various languages and domains for both flat (Xia et al, 2019;Zheng et al, 2019;Arkhipov et al, 2019;Lothritz et al, 2020;Yu et al, 2020;Yang et al, 2021) and nested (Sohrab and Miwa, 2018;Katiyar and Cardie, 2018;Dadas and Protasiewicz, 2020;Wang et al, 2020) entities.…”