ubuntu-phone team mailing list archive
-
ubuntu-phone team
-
Mailing list archive
-
Message #14763
Re: Idea for better user experience on train/bus/car
I believe visual input plays a game changing role here. The proof is that
most people feel a lot better when they look at distant landscape.
I also recognize that our brain is compensating for the movement (otherwise
it would get ridiculously hard to write this email) but the problem lies
exactly there. Same thing happens when an astigmatic eye looks at a dense
parallel pattern. The brain compensates optical error without you ever
finding out but it comes at a cost of headache and dizziness.
2015年8月8日(土) 19:30 Ed Kapitein <ed@xxxxxxxxxxxx>:
>
> On 08/05/15 18:51, Arash wrote:
>
> Many people feel a little dizzy or nauseous when reading on a train or
> car. This happens because your eyes see that the phone (or book) moves up
> and down but your body doesn't feel that movement. (more info
> <http://www.nytimes.com/2008/07/15/science/15qna.html>)
>
> This is a pretty common situation. Particularly for me as a student I
> spend hours of commuting time daily reading or browsing on my phone.
>
> Now that Ubuntu has a goal of improving user experience on mobile and
> desktop, I was thinking about what we can do to help people feel better in
> these kinds of situation.
>
> The logical solution to this is compensating phone movements by reading
> input from accelerometer sensors and moving screen buffer a little bit to
> compensate that movement.
>
> Scenario goes like this:
>
> 1. Every few seconds when phone is awake, it checks for vertical
> vibrations with frequencies in order of 1-2Hz. If we keep record of one
> second of data from sensor, measuring this is easy with a discrete Fourier
> transform or any other similar method.
> 2. If the amount of vibrations passed a threshold and stayed more than
> that amount for a couple of seconds, phone shrinks the screen vertically
> for a small amount (a couple of millimeters)
> 3. Every couple of frames phone reads from accelerometer and moves the
> screen vertically to compensate that movement.
> 4. When phone detected that the vibrations have come down below the
> threshold, it will grow the screen back to its full size.
>
>
> I have made some some assumptions here:
>
> - That reading constantly from accelerometer does not have a drastic
> effect on battery usage (compared to battery usage of screen, because this
> needs to happen only when screen is on).
> - That we can shrink the screen vertically a millimeter or two on the
> fly when device detects these movements or the feature is enabled.
> - That moving visual output on scale of millimeters is actually enough
> to make user feel less dizzy.
> - That we can read input and move buffer fast enough. If we don't act
> fast enough, we might be making it worse for the user :)
> - And that moving vertically is enough to help users. Maybe we should
> move horizontally too.
>
> It is not something that is totally impossible to implement, for example
> iOS has an eye-candy that it moves background picture a little bit (in real
> time) when you tilt the device. (The difference is that we need to react to
> movement and not angular movement.)
>
> It's just an idea :) Let's talk about it.
>
> Arash
>
>
> Hi Arash,
>
> It is more complicated then that. All your "sensors" play a part in the
> discomfort you are experiencing. you are feeling vibrations you don't see,
> you are hearing sounds that are not part of what you are seeing, so you
> brain has to compensate all off that.
> And on top of that, your body compensates things on its own. if you excel
> you notice, but when at speed, you have no notice of the speed you are
> travelling at. if you tilt, you'll notice, but after a while, you body
> compensates.
>
> without offering any solution i hope you will find what has the most
> effect, sound, feel, smell or light.
>
> Kind regards,
> Ed
>
>
References