Systra mi, du är bäst! Ikväll ska vi tokdansa med Ronny och Ragge på bordet tills byxorna spricker så att mamma får dåndimpen och pappa ramlar ur soffan.
AbstractPose (position and orientation) tracking in room-scaled environments is an enabling technique for many applications. Today, virtual reality (vr) and augmented reality (ar) are two examples of such applications, receiving high interest both from the public and the research community. Accurate pose tracking of the vr or ar equipment, often a camera or a headset, or of different body parts is crucial to trick the human brain and make the virtual experience realistic. Pose tracking in room-scaled environments is also needed for reference tracking and metrology. This thesis focuses on an application to metrology. In this application, photometric models of a photo studio are needed to perform realistic scene reconstruction and image synthesis. Pose tracking of a dedicated sensor enables creation of these photometric models. The demands on the tracking system used in this application is high. It must be able to provide sub-centimeter and sub-degree accuracy and at same time be easy to move and install in new photo studios.The focus of this thesis is to investigate and develop methods for a pose tracking system that satisfies the requirements of the intended metrology application. The Bayesian filtering framework is suggested because of its firm theoretical foundation in informatics and because it enables straightforward fusion of measurements from several sensors. Sensor fusion is in this thesis seen as a way to exploit complementary characteristics of different sensors to increase tracking accuracy and robustness. Four different types of measurements are considered; inertialmeasurements, images from a camera, range (time-of-flight) measurements from ultra wide band (uwb) radio signals, and range and velocity measurements from echoes of transmitted acoustic signals.A simulation study and a study of the Cramér-Rao lower filtering bound (crlb) show that an inertial-camera system has the potential to reach the required tracking accuracy. It is however assumed that known fiducial markers, that can be detected and recognized in images, are deployed in the environment. The study shows that many markers are required. This makes the solution more of a stationary solution and the mobility requirement is not fulfilled. A simultaneous localization and mapping (slam) solution, where naturally occurring features are used instead of known markers, are suggested solve this problem. Evaluation using real data shows that the provided inertial-camera slam filter suffers from drift but that support from uwb range measurements eliminates this drift. The slam solution is then only dependent on knowing the position of very few stationary uwb transmitters compared to a large number of known fiducial markers. As a last step, to increase the accuracy of the slam filter, it is investigated if and how range measurements can be complemented with velocity measurement obtained as a resu...