This chapter introduces several state of the art techniques that could help to make deep underwater archaeological photogrammetric surveys easier, faster, more accurate, and to provide more visually appealing representations in 2D and 3D for both experts and public. We detail how the 3D captured data is analysed and then represented using ontologies, and how this facilitates interdisciplinary interpretation and cooperation. Towards more automation, we present a new method that adopts a deep learning approach for the detection and the recognition of objects of interest, amphorae for example. In order to provide more readable, direct and clearer illustrations, we describe several techniques that generate different styles of sketches out of orthophotos developed using neural networks. In the same direction, we present the Non-Photorealistic Rendering (NPR) technique, which converts a 3D model into a more readable 2D representation that is more useful to communicate and simplifies the identification of objects of interest. Regarding public dissemination, we demonstrate how recent advances in virtual reality to provide an accurate, high resolution, amusing and appropriate visualization tool that offers the public the possibility to 'visit' an unreachable archaeological site. Finally, we conclude by introducing the plenoptic approach, a new promising technology that can change the future of the photogrammetry by making it easier and less time consuming and that allows a user to create a 3D model using only one camera shot. Here, we introduce the concepts, the developing process, and some results, which we obtained with underwater imaging.