Given the maturity of the data integration field it is surprising that rigorous empirical evaluations of research ideas are so scarce. We identify one major roadblock for empirical work-the lack of comprehensive metadata generators that can be used to create benchmarks for different integration tasks. This makes it difficult to compare integration solutions, understand their generality, and understand their performance. We present iBench, the first metadata generator that can be used to evaluate a wide-range of integration tasks (data exchange, mapping creation, mapping composition, schema evolution, among many others). iBench permits control over the size and characteristics of the metadata it generates (schemas, constraints, and mappings). We show that iBench can be used to create very large, complex, yet realistic scenarios. Our evaluation of iBench demonstrates that it can efficiently generate large scenarios with different characteristics. We also present an evaluation of two mapping creation systems using iBench and show that the intricate control that iBench provides over metadata scenarios can reveal new and important empirical insights into integration solutions. iBench is an open-source, extensible tool that we are providing to the community. We believe it will raise the bar for empirical evaluation and comparison of data integration systems.