Relation classification is an important semantic processing task in the field of natural language processing (NLP). Data sources generally adopt remote monitoring strategies to automatically generate large-scale training data, which inevitably causes label noise problems. At the same time, another challenge is that important information can appear at any place in the sentence. This paper presents a sentence-level joint relation classification model. The model has two modules: a reinforcement learning (RL) agent and a joint network model. In particular, we combine bidirectional long short-term memory (Bi-LSTM) and attention mechanism as a joint model to process the text features of sentences and classify the relation between two entities. At the same time, we introduce an attention mechanism to discover hidden information in sentences. The joint training of the two modules solves the noise problem in relation extraction, sentence-level information extraction, and relation classification. Experimental results demonstrate that the model can effectively deal with data noise and achieve better relation classification performance at the sentence level.
Background:
K-means algorithm is implemented through two steps: initialization and
subsequent iterations. Initialization is to select the initial cluster center, while subsequent iterations
are to continuously change the cluster center until it won't change any more or the number of iterations
reaches its maximum. K-means algorithm is so sensitive to the cluster center selected during
initialization that the selection of a different initial cluster center will influence the algorithm performance.
Therefore, improving the initialization process has become an important means of
K-means performance improvement.
Methods:
This paper uses a new strategy to select the initial cluster center. It first calculates the minimum
and maximum values of the data in a certain index (For lower-dimensional data, such as twodimensional
data, features with larger variance, or the distance to the origin can be selected; for
higher-dimensional data, PCA can be used to select the principal component with the largest variance),
and then divides the range into equally-sized sub-ranges. Next adjust the sub-ranges based on
the data distribution so that each sub-range contains as much data as possible. Finally, the mean value
of the data in each sub-range is calculated and used as the initial clustering center.
Results:
The theoretical analysis shows that although the time complexity of the initialization process
is linear, the algorithm has the characteristics of the superlinear initialization method. This algorithm
is applied to two-dimensional GPS data analysis and high-dimensional network attack detection.
Experimental results show that this algorithm achieves high clustering performance and clustering
speed.
Conclusion:
This paper reduces the subsequent iterations of K-means algorithm without compromising
the clustering performance, which makes it suitable for large-scale data clustering. This algorithm
can not only be applied to low-dimensional data clustering, but also suitable for highdimensional
data.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.