Emotion processing has been a very intense domain of investigation in data analysis and NLP during the previous few years. Currently, the algorithms of the deep neural networks have been applied for opinion mining tasks with good results. Among various neuronal models applied for opinion mining a deep belief network (DBN) model has gained more attention. In this proposal, we have developed a combined classifier based on fuzzy Vader lexicon and a parallel deep belief network for emotion analysis. We have implemented multiple pretreatment techniques to improve the quality and soundness of the data and eliminate disturbing data. Afterward, we have performed a semi-automatic dataset labeling using a combination of two different methods: Mamdani's fuzzy system and Vader lexicon. As well, we have applied four feature extractors, which are: GloVe, TFIDF (Trigram), TFIDF (Bigram), TFIDF (Unigram) with the aim of transforming each incoming tweet into a digital value vector. In addition, we have integrated three feature selectors, namely: The ANOVA method, the chi-square approach and the mutual information technique with the objective of selecting the most relevant features. Further, we have implemented the DBN as classifier for classifying each inputted tweet into three categories: neutral, positive or negative. At the end, we have deployed our proposed approach in parallel way employing both Hadoop and Spark framework with the purpose of overcoming the problem of long runtime of massive data. Furthermore, we have carried out a comparison between our newly suggested hybrid approach and alternative hybrid models available in the literature. From the experimental findings, it was found that our suggested vague parallel approach is more powerful than the baseline patterns in terms of false negative rate (1.33%), recall (99.75%), runtime (32.95s), convergence, stability, F1 score (99.53%), accuracy (98.96%), error rate (1.04%), kappa-Static (99.1%), complexity, false positive rate (0.25%), precision rate (97.59%) and specificity rate (98.67%). As a conclusion, our vague parallel approach outperforms baseline and deep learning models, as well as certain other approaches chosen from the literature.