This work explores our developments over a framework and over a platform for human-robot-interaction with emotions. The focus of our framework is on visuo-auditory perception and response. In other words, perception and response can be called analysis and synthesis; the analysis is responsible for the classification of human emotion and the synthesis is responsible for the synthetic expression that the robot must show. This paper is focused on the synthesis and also on how the synthesis can affect human engagement during an interactive conversation with the robot.