With technological advancements in voice command programs, such as Amazon鈥檚 Alexa, Google Assistant and Apple Siri, it鈥檚 become a new normal for people to speak into their devices to operate their home appliances and gadgets. But a Canadian technology start-up is working on remoting devices without users needing to do anything at all.

, a Montreal-based computer and electronics manufacturing company, says it has built a headset that translates brain and bio-signals into commands for devices, enabling various appliances to understand the needs of a user without any physical or vocal prompts.

鈥淲e are building brain computing interfaces,鈥 Aavaa CTO and founder Naeem Komeilipoor told CTVNews.ca. 鈥淭his technology can understand the attention and intent of users. It tracks their head movement, eye movement, facial gestures like blinking or clenching, and uses this as an input to control their devices.鈥

Komeilipoor, a neuroscientist who previously worked at the University of Montreal, says Aavaa devices, including earpieces, glasses and headsets, allow people to operate home appliances simply by moving their eyes.

Yet, beyond the seemingly-telekinetic ability to turn on your TV just by looking at it, Komeilipoor pointed out that this tech could have significant medical applications, such as allowing paralyzed people to steer wheelchairs, or monitoring the biometrics of intensive care patients more thoroughly.

And, he said, the devices can also help those with speech and hearing impairments.

鈥淚 grew up with grandparents who were hearing-aid users,鈥 Komeilipoor said, standing before his display at the Hardware Tech and Founders Showcase, an exhibit in Toronto hosted by founder support organization last week.

鈥淭he problem with hearing aids is that in a noisy environment like this they often fail because they don鈥檛 know what sounds you are paying attention to.鈥

The concept Komeilipoor is referring to is called 鈥渁uditory scene analysis,鈥 or 鈥渁uditory stream segregation鈥 -- the process by which the human auditory system processes and organizes different sounds. Hearing aid technology is often able to reduce background noise, but the equipment is not yet capable of understanding the intention or attention of the listener.

To understand this, he said, 鈥淵ou need to monitor their brain and bio signals, or their head movement, eye movement.鈥

Through attention tracking, speech enhancement and wearable sensors, Aavaa鈥檚 devices are an attempt to tap into cognitive processes that have remained beyond the reach of computers, he said.

鈥淯ntil now, it has been us trying to make machines understand us. With this type of technology, machines could easily understand our intention and provide service for us.鈥