Fact-based Visual Question Answering (FVQA), a challenging variant of VQA, requires a QA-system to include facts from a diverse knowledge graph (KG) in its reasoning process to produce an answer. Large KGs, especially commonsense KGs, are known to be incomplete, i.e. not all non-existent facts are always incorrect. Therefore, being able to reason over incomplete KGs for QA is a critical requirement in realworld applications that has not been addressed extensively in the literature. We develop a novel QA architecture that allows us to reason over incomplete KGs, something current FVQA state-of-the-art (SOTA) approaches lack. We use KG Embeddings, a technique widely used for KG completion, for the downstream task of FVQA. We also employ a new image representation technique we call 'Image-as-Knowledge' to enable this capability, alongside a simple one-step co-Attention mechanism to attend to text and image during QA. Our FVQA architecture is faster during inference time, being O(m), as opposed to existing FVQA SOTA methods which are O(N logN ), where m = number of vertices, N = number of edges = O(m 2 ). We observe that our architecture performs comparably in the standard answer-retrieval baseline with existing methods; while for missing-edge reasoning, our KG representation outperforms the SOTA representation by 25%, and image representation outperforms the SOTA representation by 2.6%. The primary contribution of this paper is a method that permits FVQA to reason about common-sense facts that are absent from the knowledge graph (missing edges). KG embeddings permits us to offer two additional contributions to the SOTA in FVQA: an 'image-as-knowledge' representation of visual information, and a coAttention method for combining visual and textual inputs. 'Image-as-knowledge' represents the image as the span of the KG embedding vectors for the entities found in it. It is more effective than a 'bag-of-words' image representation, because the 'imageas-knowledge' vectors encode information about the graph structure. CoAttention uses the words of the query to compute weighted combinations of the entity vectors, in effect a vector within the span of the image entities. Two such iden-