Quantifying natural behavior from video recordings is a key component in ethological studies. Markerless pose estimation methods have provided an important step toward that goal by automatically inferring kinematic body keypoints. The next step in behavior quantification is utilization of these features toward organizing and interpreting behavioral segments into states. In this work, we introduce a novel deep learning toolset to address this aim. In particular, we introduce OpenLabCluster which clusters segments into groups according to the similarity of kinematic body keypoints and then employs active learning approach which refines the clusters and classifies them into behavioral states. The active learning approach is an iterative semi-supervised deep learning methodology selecting representative examples of segments to be annotated such that the annotation informs clustering and classification of all segments. With these methodologies, OpenLabCluster contributes to faster and more accurate organization of behavioral segments with only a sparse number of them being annotated. We demonstrate OpenLabCluster performance on four different datasets, which include different animal species exhibiting natural behaviors, and show that it boosts clustering and classification compared to existing methods, even when all segments have been annotated. OpenLabCluster has been developed as an open-source interactive graphic interface which includes all necessary functions to perform clustering and classification, informs the scientist of the outcomes in each step, and incorporates the choices made by the scientist in further steps.