This paper presents a safe robust policy iteration (SR-PI) algorithm to design controllers with satisficing (good enough) performance and safety guarantee. This is in contrast to standard PI-based control design methods with no safety certification. It also moves away from existing safe control design approaches that perform pointwise optimization and are thus myopic. Safety assurance requires satisfying a control barrier function (CBF), which might be in conflict with the performancedriven Lyapunov solution to the Bellman equation arising in each iteration of the PI. Therefore, a new development is required to robustly certify the safety of an improved policy at each iteration of the PI. The proposed SR-PI algorithm unifies performance guarantee (provided by a Bellman inequality) with safety guarantee (provided by a robust CBF) at each iteration. The Bellman inequality resembles the satisficing decision making framework and parameterizes the sacrifice on the performance with an aspiration level when there is a conflict with safety. This aspiration level is optimized at each iteration to minimize the sacrifice on the performance. It is shown that the presented satisficing control policies obtained at each iteration of the SR-PI guarantees robust safety and performance. Robust stability is also guaranteed when there is no conflict with safety. Sum of squares (SOS) program is employed to implement the proposed SR-PI algorithm iteratively. Finally, numerical simulations are carried out to illustrate the proposed satisficing control framework.