How would you use a touch screen mobile device if you couldn’t use your hands to touch the device? With support from Google.org, the Neil Squire Society released the LipSync, a mouth controlled input device which enables people with little or no hand movement to operate a touchscreen device.

The Lipsync is a mouth operated joystick that allows a person to control a computer cursor with minimal head and neck movement. All the electronics are housed in the ‘head’ of the device so there are no additional control boxes, making the LipSync a good candidate for portable, wheelchair-mounted applications. The mouthpiece is attached to a precision miniature joystick sensor that requires very slight pressure in order to move a cursor on the screen.

An estimated 1,000,000 people in Canada and the United States have limited or no use of their arms—meaning they are unable to use touchscreen devices that could provide access to helpful apps and services.

While solutions exist for desktop computers, they can cost up to $1,500 USD and do not work well on mobile devices.

We are releasing the project open source so it can be affordably made at the community level by makers, engineers, tinkerers, and hobbyists. The total cost will be less than $300 to source and assemble 3D printed parts and can be built as a weekend project.

It currently works with Android and Windows phones and tablets, and all desktops and laptops. iOS currently does not have cursor support so iPhones and iPads are incompatible.


Makers have the option of acquiring the parts individually or through Makers Making Change.