Web was developed to share information among the users through internet as some hyperlinked documents. If someone wants to collect some data from the web he has to search and crawl through the documents to fulfil his needs. Concept of Linked Data creates a breakthrough at this stage by enabling the links within data. So, besides the web of connected documents a new web developed both for humans and machines, i.e., the web of connected data, simply known as Linked Data Web. Since it is a very new domain, still a very few works has been done, specially the publication of legacy data within a University domain as Linked Data.In this paper our approach concentrates on publishing the data from our University's databases as linked Data on the web and also ensuring the trustworthiness of the data by addition of provenance data along with. For this purpose we have introduced a framework to make all the steps towards the Linked Data publication easy for legacy databases within a university domain.
Purpose
– This article aims to discuss how the emergence of advanced semantic web technology has transformed the conventional web into machine processable and understandable form.
Design/methodology/approach
– In this paper the authors survey the current research works, tools and applications on publishing legacy data as linked data with the aspiration of conferring healthier understanding of the working domain of the linked data world.
Findings
– Today, a vast amount of data are stored in various file formats other than RDF, which are called legacy data. In order to publish them as linked data they need to be extracted and converted into RDF or linked data without altering the original data schema or loss of information.
Originality/value
– Most of the key issues have to be addressed. A more sophisticated approach to this technology is the linked data, which constructs the transformation of web of documents into the web of connected data possible.
Nowadays with the rapid use of Internet makes it easy to collect information as a matter of mouse click over the Web. With this enormous opportunity it becomes easy to make copy and use of someone else's works or ideas without giving the proper credit to the original owner. Normally this kind of practice is known as plagiarism. Plagiarism has already become a serious issue for the different publishing house of journals or the conference organizers to prevent the authors to plagiarize their work. Considering this issue several tools have been developed to check for plagiarism mostly based on syntactical or word based matching. In this paper a new conceptual model is proposed to generate a decent tool to identify the plagiarized content based on syntactical and semantic matching.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.