Machine learning (ML) based classifiers are vulnerable to evasion attacks, as shown by recent attacks. However, there is a lack of systematic study of evasion attacks on ML‐based anti‐phishing detection. In this study, we show that evasion attacks are not only effective on practical ML‐based classifiers, but can also be efficiently launched without destructing the functionalities and appearance. For this purpose, we propose three mutation‐based attacks, differing in the knowledge of the target classifier, addressing a key technical challenge: automatically crafting an adversarial sample from a known phishing website in a way that can mislead classifiers. To launch attacks in the white‐ and gray‐box scenarios, we also propose a sample‐based collision attack to gain the knowledge of the target classifier. We demonstrate the efficacy of our evasion attacks on the state‐of‐the‐art, Google's phishing page filter, achieved 100% attack success rate in less than one second per website. Moreover, the transferability attack on BitDefender's industrial phishing page classifier, TrafficLight, achieved up to 81.25% attack success rate. We further propose a similarity‐based method to mitigate such evasion attacks, Pelican, which compares the similarity of an unknown website with recently detected phishing websites. We demonstrate that Pelican can effectively detect evasion attacks, hence could be integrated into ML‐based classifiers. We also highlight two strategies of classification rule selection to enhance the robustness of classifiers. Our findings contribute to design more robust phishing website classifiers in practice.