Approximate linear programs (ALPs) are well-known models based on value function approximations (VFAs) to obtain heuristic policies and lower bounds on the optimal policy cost of Markov decision processes (MDPs).The ALP VFA is a linear combination of predefined basis functions that are chosen using domain knowledge and updated heuristically if the ALP optimality gap is large. We side-step the need for such basis function engineering in ALP -an implementation bottleneck -by proposing a sequence of ALPs that embed increasing numbers of random basis functions obtained via inexpensive sampling. We provide a sampling guarantee and show that the VFAs from this sequence of models converge to the exact value function. Nevertheless, the performance of the ALP policy can fluctuate significantly as more basis functions are sampled. To mitigate these fluctuations, we "self-guide" our convergent sequence of ALPs using past VFA information such that a worstcase measure of policy performance is improved. We perform numerical experiments on perishable inventory control and generalized joint replenishment applications, which, respectively, give rise to challenging discounted-cost MDPs and average-cost semi-MDPs. We find that self-guided ALPs (i) significantly reduce policy cost fluctuations and improve the optimality gaps from an ALP approach that employs basis functions tailored to the former application, and (ii) deliver optimality gaps that are comparable to a known adaptive basis function generation approach targeting the latter application. More broadly, our methodology provides application-agnostic policies and lower bounds to benchmark approaches that exploit application structure. Bhat N, Farias V, Moallemi CC (2012) Non-parametric approximate dynamic programming via the kernel method. Advances in Neural Information Processing Systems, 386-394. Blado D, Toriello A (2019) Relaxation analysis for the dynamic knapsack problem with stochastic item sizes. SIAM Journal on Optimization 29(1):1-30. Chen X, Pang Z, Pan L (2014) Coordinating inventory control and pricing strategies for perishable products. Operations Research 62(2):284-300. De Farias DP, Van Roy B (2003) The linear programming approach to approximate dynamic programming. Operations Research 51(6):850-865. De Farias DP, Van Roy B (2004) On constraint sampling in the linear programming approach to approximate dynamic programming. Mathematics of Operations Research 29(3):462-478. Desai VV, Farias VF, Moallemi CC (2012) Approximate dynamic programming via a smoothed linear program. Operations Research 60(3):655-674. Folland GB (1999) Real Analysis: Modern Techniques and Their Applications (New York, NY: John Wiley & Sons). Forsell N, Sabbadin R (2006) Approximate linear-programming algorithms for graph-based Markov decision processes. Karaesmen IZ, Scheller-Wolf A, Deniz B (2011) Managing perishable and aging inventories: review and future research directions, 393-436 (New York, NY: Springer). Klabjan D, Adelman D (2007) An infinite-dimensional linear programming al...