Many image processing, computer graphics, and computer vision problems can be treated as image-to-image translation tasks. Such translation entails learning to map one visual representation of a given input to another representation. Image-to-image translation with generative adversarial networks (GANs) has been intensively studied and applied to various tasks, such as multimodal image-to-image translation, super-resolution translation, object transfiguration-related translation, etc. However, image-to-image translation techniques suffer from some problems, such as mode collapse, instability, and a lack of diversity. This article provides a comprehensive overview of image-to-image translation based on GAN algorithms and its variants. It also discusses and analyzes current state-of-the-art image-to-image translation techniques that are based on multimodal and multidomain representations. Finally, open issues and future research directions utilizing reinforcement learning and three-dimensional (3D) modal translation are summarized and discussed.
Precisely assessing the severity of persons with Covid-19 at an early stage is an effective way to increase the survival rate of patients. Based on the initial screening, to identify and triage the people at highest risk of complications that can result in mortality risk in patients is a challenging problem, especially in developing nations around the world. This problem is further aggravated due to the shortage of specialists. Using machine learning (ML) techniques to predict the severity of persons with Covid-19 in the initial screening process can be an effective method which would enable patients to be sorted and treated and accordingly receive appropriate clinical management with optimum use of medical facilities. In this study, we applied and evaluated the effectiveness of three types of Artificial Neural Network (ANN), Support Vector Machine and Random forest regression using a variety of learning methods, for early prediction of severity using patient history and laboratory findings. The performance of different machine learning techniques to predict severity with clinical features shows that it can be successfully applied to precisely and quickly assess the severity of the patient and the risk of death by using patient history and laboratory findings that can be an effective method for patients to be triaged and treated accordingly.
Drones/unmanned aerial vehicles (UAVs) have recently grown in popularity due to their inexpensive cost and widespread commercial use. The increased use of drones raises the possibility that they may be employed in illicit activities such as drug smuggling and terrorism. Thus, drone monitoring and automated detection are critical for protecting restricted areas or special zones from illicit drone operations. One of the most challenging difficulties in drone detection in surveillance videos is the apparent likeness of drones against varied backdrops. This paper introduces an automated image-based drone-detection system that uses an enhanced deep-learning-based object-detection algorithm known as you only look once (YOLOv5) to defend restricted territories or special zones from unauthorized drone incursions. The transfer learning to pretrain the model is employed for improving performance due to an insufficient number of samples in our dataset. Furthermore, the model can recognize the detected object in the images and mark the object’s bounding box by joining the results across the region. The experiments show outstanding results for the loss value, drone location detection, precision and recall.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.