This study presents the initial study for a new approach to visualize an acoustic sound aimed at mimicking the traveling wave propagation of the basilar membrane within the human cochlea. Typically, a fast Fourier transform (FFT) is required to extract the frequency information from acoustic sound (i.e., voice) for speech recognition. Although this algorithm ensures real-time frequency extraction due to the inherent fast recursive structure, it is necessary to develop a new frequency selectivity technique for advanced speech recognition. We explore the potential of the cochlea-inspired sound visualization to deliver new frequency selectivity by using an image sensor. The experimental prototyping model is fabricated, and we capture images of frequency dependent wave propagation motion using a camera and reproduce 2D images through motion magnification. This approach offers a promising application for speech recognition systems because no FFT is required to extract the frequency information, although there are outstanding technical problems that need to be further examined.