With the unprecedented growth of signal processing and machine learning application domains, there has been a tremendous expansion of interest in distributed optimization methods to cope with the underlying large-scale problems. Nonetheless, inevitable system-specific challenges such as limited computational power, limited communication, latency requirements, measurement errors, and noises in wireless channels impose restrictions on the exactness of the underlying algorithms. Such restrictions have appealed to the exploration of algorithms' convergence behaviors under inexact settings. Despite the extensive research conducted in the area, it seems that the analysis of convergences of dual decomposition methods concerning primal optimality violations, together with dual optimality violations is less investigated. Here, we provide a systematic exposition of the convergence of feasible points in dual decomposition methods under inexact settings, for an important class of global consensus optimization problems. Convergences and the rate of convergences of the algorithms are mathematically substantiated, not only from a dual-domain standpoint but also from a primal-domain standpoint. Analytical results show that the algorithms converge to a neighborhood of optimality, the size of which depends on the level of underlying distortions. Further elaboration of a generalized problem formulation is also furnished, together with the convergence properties of the underlying algorithms. Finally, theoretical derivations are verified by numerical experiments.