Creativity research often relies on human raters to judge the novelty of participants' responses on openended tasks, such as the alternate uses task (AUT). Albeit useful, manual ratings are subjective and labor-intensive. To address these limitations, researchers increasingly use automatic scoring methods based on a natural language processing technique for quantifying the semantic distance between words. However, many methodological choices remain open on how to obtain semantic distance scores for ideas, which can significantly impact reliability and validity. In this project, we propose a new semantic distance-based method, maximum associative distance (MAD), for assessing response novelty in AUT. Within a response, MAD uses the semantic distance of the word that is maximally remote from the prompt word to reflect response novelty. We compare the results from MAD with other competing semantic distance-based methods, including element-wise multiplication-a commonly used compositional model-across three published datasets including a total of 447 participants. We found MAD to be more strongly correlated with human creativity ratings than the competing methods. In addition, MAD scores reliably predict external measures such as openness to experience. We further explored how idea elaboration affects the performance of various scoring methods and found that MAD is closely aligned with human raters in processing multiword responses. Thus, the MAD method improves the psychometrics of automatic creativity assessment, while also provides insights into what human raters perceive as creative about ideas.