DNA compaction and accessibility in eukaryotes are governed by nucleosomes and orchestrated through interactions between DNA and DNA-binding proteins. Using QuantAFM, a method for automated image analysis of atomic force microscopy (AFM) data, we performed a detailed statistical analysis of structural properties of mono-nucleosomes. QuantAFM allows fast analysis of AFM images, including image preprocessing, object segmentation, and quantification of different structural parameters to assess DNA accessibility of nucleosomes. A comparison of nucleosomes reconstituted with and without linker histone H1 quantified H1’s already described ability of compacting the nucleosome. We further employed nucleosomes bearing two charge-modifying mutations at position R81 and R88 in histone H2A (H2A R81E/R88E) to characterize DNA accessibility under destabilizing conditions. Upon H2A mutation, even in presence of H1, the DNA opening angle at the entry/exit site was increased and the DNA wrapping length around the histone core was reduced. Interestingly, a distinct opening of the less bendable DNA side was observed upon H2A mutation, indicating an enhancement of the intrinsic asymmetry of the Widom-601 nucleosomes. This study validates AFM as a technique to investigate structural parameters of nucleosomes and highlights how the DNA sequence, together with nucleosome modifications, can influence the DNA accessibility.
Temporal expressions in text play a significant role in language understanding and correctly identifying them is fundamental to various retrieval and natural language processing systems. Previous works have slowly shifted from rule-based to neural architectures, capable of tagging expressions with higher accuracy. However, neural models can not yet distinguish between different expression types at the same level as their rule-based counterparts. In this work, we aim to identify the most suitable transformer architecture for joint temporal tagging and type classification, as well as, investigating the effect of semi-supervised training on the performance of these systems. Based on our study of token classification variants and encoder-decoder architectures, we present a transformer encoder-decoder model using the RoBERTa language model as our best performing system. By supplementing training resources with weakly labeled data from rule-based systems, our model surpasses previous works in temporal tagging and type classification, especially on rare classes. Our code and pre-trained experiments are available at: https://github.com/satya77/ Transformer_Temporal_Tagger
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.