PROFACTOR showcases skill-adaptive projection guidance that slashes training time and re-work effort.
Researchers at PROFACTOR, working within the EU-funded RaRe² project, have introduced a multi-modal guidance system that
projects context-aware instructions directly onto real parts—eliminating paper notes, pens and slow, static training methods.
The system switches seamlessly between two user-centred modes:
In laboratory trials on automotive door panels the approach boosted task efficiency, kept users more engaged and allowed repeated use
of training parts thanks to marker-less interaction . A ceiling-mounted Orbbec Astra 2 camera combined with a Panasonic PT-MZ880
projector delivered sub-5 cm annotation accuracy under calibrated conditions. Built on a modular web stack (React + Three.js), the platform lets engineers add new parts and defect types in minutes and is already
being prepared for electronics and aerospace pilots. Planned upgrades include automatic RGB-D calibration, glove-robust hand tracking
and optional voice commands to keep operators hands-free.
This research is financed by Horizon Europe RaRe² (Grant 101092073) and the Upper Austrian Government’s DemoDatenPro initiative.
📄 For a deeper dive into the research, read the full paper:
• Skill-Based Adaptation through Intuitive Interfaces: Multi-Modal Guidance Systems for Industrial Environments