Has anyone gotten Unity to work with multitouch on Linux? The OS captures the touch just fine, but Unity doesn't see them as either mouse or touch events. So far I've tried Fedora and Ubuntu.
Sorry, touch input is not currently supported for Linux standalone. This is on our roadmap, but there's currently no ETA.
It looks like Unity 2018.2 fixed something related to touch "Linux: Fix Linux touch input for mouse events (947049)" at https://unity3d.com/unity/beta/unity2018.2.0b2 I'm very curious if multi-touch works yet because I've been holding off on buying a multi-touch monitor for a project until it does.
Any updates on this? I'm curious as what the technical limitation is that prevents Unity from processing touch events on Linux?
We've gotten a working prototype using TouchScript with TUIO input and mtdev2tuio or UPDD TUIO server as the backend. It's hacky because we basically have to convert the touches to OSC messages and then consume those in our app, but it functions.
Hey @UnityBLI You should check Reactivision http://reactivision.sourceforge.net/ It has a working linux version. How did you get Touchsript to listen the TUIO messages ?
Im using a custom driver for that purpose... its like a Touch driver working like a Mouse Driver... and the touch gestures can be passed to the kernel like (Mouse Events...) ie. Pinch-Expand gesture = Mouse Scroll UP/DOWN... , the option 2 is... create a Plugin Between Linux and Unity .