Augmented reality (AR) is reshaping scientific visualization by offering immersive, interactive environments. This project explores the development of an AR molecular visualization system for the Meta Quest 3 that simulates atomic orbitals and hybridization processes in real time. Traditional 2D models often fail to convey the spatial complexity of molecular structures, making AR a promising alternative.
The primary research question is: How can AR technology improve interactive molecular visualization on the Meta Quest 3? The hypothesis is that AR systems can provide greater spatial understanding and engagement compared to static models.
Using Unreal Engine 5.4 and the Oculus XR Plugin, the system enables users to spawn, manipulate, and combine s, p, and d orbitals into hybrid orbitals such as sp, sp², and sp³. Hand tracking and passthrough mode allow users to interact directly with orbitals, while the system provides feedback on correct combinations and spatial configurations.
Expected outcomes include insights into the usability and technical performance of AR-based molecular simulations. This research contributes to the growing field of AR applications in scientific visualization, with potential implications for chemistry education, industrial design, and scientific research.
Poster submission was presented at the New Westminster campus on April 10, 2025, for Student Research Days 2025.
The main objective of this applied research project is to develop a computer tool to capture a full map of the human body, then overlay it onto video footage to record accurate measurements of joint angles and linear displacement in real-time. The completed tool will give Douglas College Sports Science students and instructors a real-time tracking tool for joint movement, providing accurate data for making graphs and enhancing their ability to learn and apply concepts discussed in classes. While the current application of the tool involves motion analysis of the body, it has potential future applications in different research areas as well. The tool works by asking participants to perform specific movements, such as throwing a ball, allowing the system to record joint angles and linear displacements. This data is then stored in an Excel sheet, allowing the user to extract relevant insights from the data by making graphs. The integration of a calibration tool gives users a way to manually select two points for distance calibration, ensuring the accuracy of the real-time measurements. Poster submission presented at the New Westminster campus on April 11, 2024, for Student Research Days 2024.