Skip to content

Samsung's Facesense Recognizes Facial Expressions For VR Navigation

Samsung's Facesense Recognizes Facial Expressions For VR Navigation

With body language, humans are able to express a range of feelings with only the most the subtle shifts in demeanor. Our facial expressions alone can show off a huge range of responses. Interaction with VR is something various companies continue to try to nail down, attempting to enable users to manipulate these worlds while breaking immersion as little as possible. Samsung’s experimental Facesense is a new attempt at changing the game, harnessing the power of our facial expression for hands-free VR interaction.


April 14th – 15th at the VRLA Expo, Samsung showed a new creation from their  C-Lab division. C-Lab cultivates ideas that are more experimental and this particular one provides a new way to navigate within VR. Detailed in an announcement, Facesense tracks electric signals that are created any time we speak, change our expression, or shift our gaze. Those signals are then used for navigation input along with a few spoken commands.


We tried the technology at VRLA briefly and it was a very early concept. It likely won’t be something that dominates as a primary means of interaction within VR, but it could be a complement to VR controllers. We’ll have to see if Facesense can find a degree of consistency with its input across all users. It could also serve as an option for those that cannot use VR controllers, opening up accessibility to virtual technology in a big way.

Weekly Newsletter

See More