The coronavirus disease 2019 (COVID-19) pandemic is the most rapidly evolving global emergency since March 2020 and one of the most exercised topics in all aspects of the world. So far there are numerous articles that have been published related to COVID-19 in various disciplines of science and social context. Since from the very beginning, researchers have been trying to address some fundamental questions like how long it will sustain when it will reach the peak point of spreading, what will be the population of infections, cure, or death in the future. To address such issues researchers have been used several mathematical models from the very beginning around the world. The goal of such predictions is to take strategic control of the disease. In most of the cases, the predictions have deviated from the real data. In this paper, a mathematical model has been used which is not explored earlier in the COVID-19 predictions. The contribution of the work is to present a variant of the linear regression model is the piece-wise linear regression, which performs relatively better compared to the other existing models. In our study, the COVID-19 data set of several states of India has been used.
This paper 1 describes two systems for Named Entity Recognition (NER) and performance of two systems has been compared. The first system is a rule-based one whereas the second one is statistical (based on CRF) in nature. The systems vary in some other aspects too, for example, the first system works on untagged data (not even POS tag is done) to identify NER whereas the second system makes use of a POS tagger and a chunker. The rules used by the first system are mined from the training data. The CRF-based classification does not require any explicit linguistic rules but it uses a gazetteer built from Wiki and other sources.
With over 1.4 million Bodo speakers, there is a need for Automated Language Processing systems such as Machine translation, Part Of Speech tagging, Speech recognition, Named Entity Recognition, and so on. In order to develop such a system it requires a sufficient amount of dataset. In this paper we present a detailed description of the primary resources available for Bodo language that can be used as datasets to study Natural Language Processing and its applications. We have listed out different resources available for Bodo language: 8,005 Lexicon dataset collected from agriculture and health, Raw corpus dataset of 2,915,544 words, Tagged corpus consisting of 30,000 sentences, Parallel corpus of 28,359 sentences from tourism, agriculture and health and Tagged and Parallel corpus dataset of 37,768 sentences. We further discuss the challenges and opportunities present in Bodo language.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.