Convergent progress in cognitive science and machine learning has given rise to a new approach to modeling human behavior, which can better account for its adaptive nature. This follows the theory of computational rationality, under which behavior adapts to maximize expected utility while constrained by the computational limits implied by their cognition. Building computational models of this theory involves three phases: i) specifying rewards and cognitive resources, ii) formulating a sequential decision-making problem to control such resources to maximize expected reward, and iii) using methods like deep reinforcement learning to approximate the optimal policy solution. However, despite their success in reproducing human-like adaptive behavior across everyday tasks, these models are not easy to build. The design of psychological assumptions and machine learning-decisions concerning policy optimization, parameter inference, and model selection, are all tangled processes rife with pitfalls that hinder the development of valid and effective models. The article, drawing from a decade of work on this approach, lays out a workflow for tackling this challenge and is accompanied with detailed discussion of pros and cons at key decision points.