hybrid-graphics-linux team mailing list archive
-
hybrid-graphics-linux team
-
Mailing list archive
-
Message #01022
Re: some hybrid linux hack (rather dirty)
just spit-balling here, but has anyone looked at the new xf86-video-nested code?
http://lists.x.org/archives/xorg-devel/2011-May/022007.html
On Fri, May 6, 2011 at 10:28 AM, Joakim Gebart <joakim.gebart@xxxxxx> wrote:
> I've been using this with great success the last couple of days. On my
> i3 2.4GHz I can run it at 60fps (the usleep in the code changed to
> 16000) with around 30% cpu usage on a single core.
> In my opinion the other method using virtualgl and passing data as jpeg
> is not the optimal solution to the optimus problem and feels a bit too
> complex. Using the quick and dirty solution, sharing an RGB buffer
> between two servers, should be able to get better framerate after some
> optimizations since we don't need to do any encoding of the data.
> For comparison: I get around 4000fps in glxgears on my GF GT310 using
> this method, although I only display 60fps on the intel card.
> I've been playing with the idea of running the windump program as a root
> window on the intel X display and then running Gnome or whatever on only
> the nvidia display. Another thing I've been wanting to test is to use
> the intel card's opengl for rendering and putting the XShm image in a
> texture. Also syncing with the intel card's refresh rate should be easy
> using opengl.
>
> Florian: You didn't mention any license on the code you posted, would
> you like to release it under the GPL?
> I'm interested in working some more on this program, but I want to be
> able to release my modifications if I manage to achieve anything. Like
> markc said, github is excellent for this kind of project.
>
> Best regards,
> Joakim
>
> 2011-05-01 16:30, Joaquín Ignacio Aramendía skrev:
>> El dom, 01-05-2011 a las 12:13 +1000, Pigeon escribió:
>>> Hi Florian,
>>>
>>>> i recently bought a laptop (msi cx640) using nvidia optimus
>>>> unknowingly, that it was pretty unusable to me on linux.
>>>> however i found a (rather dirty) hack to be able to at least use my
>>>> nvidia graphics card, so that i dont have to throw my laptop into the
>>>> trashbin.
>>> I've been experimenting the same kind of approach by using vnc
>>> (x11vnc + vncviewer). I'm running two X servers on the same host, one
>>> intel and one nvidia. I have to force the two X servers to be on the
>>> same vt to trick both X to "render" at the same time, rather dirty too.
>>>
>>> Performance isn't great as expected. I got quite a lot of
>>> tearing when I was testing with some GL apps/games. But it's usable at
>>> least.
>>>
>>> I haven't tried your windump app yet. I believe you can get
>>> some performance gain by using the Xdamage protocol, which x11vnc uses
>>> by default. Though I'm not an X expert either.
>>>
>>> This is on an Alienware M11x-R2 btw.
>>>
>>>
>>> Pigeon.
>>
>> Small steps for you... big ones for hybrid graphics I think :)
>
> _______________________________________________
> Mailing list: https://launchpad.net/~hybrid-graphics-linux
> Post to : hybrid-graphics-linux@xxxxxxxxxxxxxxxxxxx
> Unsubscribe : https://launchpad.net/~hybrid-graphics-linux
> More help : https://help.launchpad.net/ListHelp
>
References