One of the main noise sources in current gravitational wave detectors is the thermal noise of the high-reflectivity coatings on the main interferometer optics.
Coating thermal noise is dominated by the mechanical loss of the high-refractive index material within the coating stacks, Ta2O5 mixed with TiO2. For upgrades to room-temperature detectors, a mixture of GeO2 and TiO2 is an interesting alternative candidate coating material. While the rather low refractive index of GeO2 increases with increasing TiO2 content, a higher TiO2 content results in a lower threshold temperature before heat treatment leads to crystallisation, and potentially to a degradation of optical properties. For future cryogenic detectors, on the other hand, a higher TiO2 content is beneficial as the TiO2 suppresses the low-temperature mechanical loss peak of GeO2. In this paper, we present the optical properties of coatings -- produced by plasma-assisted ion-beam evaporation -- with high TiO2 content at 1550nm, a laser wavelength considered for cryogenic gravitational-wave detectors, as a function of heat-treatment temperature. For comparison, the absorption was also measured of pure GeO2. Furthermore, results at the currently-used wavelength of 1064nm are presented.