On 3/18/18 10:36 PM, Carlos Garnacho wrote:

>>>> 2. the touchscreen seems not to understand multitouch at all so it also
>>>> does not allow scrolling most of the time. If it is firefox or a terminal -
>>>> moving finger over the window selects text. No two finger scrolling is
>>>> possible (enabled and works with the touchpad).
>>>
>>> Ultimately, reaction to (multi)touch is implemented by each specific
>>> application. Any issues there should thus be reported on each
>>> application bug reporting system. NB, such issues are very probably
>>> already filed, so look out for duplicates.
>>
>> Is this different from touchpad? I am asking as the touchpad settings are
>> system wide and it works the same everywhere but there is no such control
>> over the touchscreen behaviour and I wonder why... I expected touchpad and
>> touchscreen to share same settings or each to have own settings and neither
>> seems to be the case.
> 
> It is a whole lot different :). From the perspective of the
> application, a touchpad is the essentially same than a mouse, and
> applications have been able to deal with those for many years. As the
> actions apply on a single concrete point (where you see the mouse
> cursor), the windowing (xorg/wayland) is able to send specific events
> of what happened (moved to these coordinates, pressed mouse button,
> scrolled by this much).
> 
> Multitouch is peculiar in that the application is made known of basic
> information (new touchpoint at these coordinates, touchpoint moved
> somewhere else, touchpoint was lifted) for *every* touchpoint. The
> application is in charge of seeing where do each apply and giving
> those a meaning, some examples:
> - A touchpoint happening in a browser tab might eventually tap some
> button in the webpage, or be used by the browser to start scrolling if
> it moves past a distance (and effectively cancelling other actions)
> - Say you actually meant to zoom on the page, so a second touchpoint
> on the webpage arrives a couple milliseconds later. All previous
> actions should be rendered ineffective, and the browser should see how
> far away are those touchpoints in order to start zooming based on that
> initial distance, using the center of those 2 points as the zooming
> anchor.
> - But say you missed the webpage in that second touch, and it ends up
> on the location bar. What should the app do in that case? I dunno, but
> what it shouldn't do is zooming :).

Oh. Ok. Seems weird as touchpad seems to have to solve exact same problems
but ok :)


> Toolkits like GTK+ may help to some extent, but nothing enforced apps
> to using that fully/correctly. And in the case of Firefox, I think
> their use of GTK+ is a lot more superficial than that, so there's more
> on their plate to handle.
> 
> TL;DR: Touch input is highly location and context dependent, there's
> no way "correct" multitouch-y behavior can be done for free without
> app involvement.

Got it. Thanks for the explanations.


-- 
Alexey
_______________________________________________
gnome-shell-list mailing list
gnome-shell-list@gnome.org
https://mail.gnome.org/mailman/listinfo/gnome-shell-list

Reply via email to