just spit-balling here, but has anyone looked at the new xf86-video-nested code? http://lists.x.org/archives/xorg-devel/2011-May/022007.html
On Fri, May 6, 2011 at 10:28 AM, Joakim Gebart <[email protected]> wrote: > I've been using this with great success the last couple of days. On my > i3 2.4GHz I can run it at 60fps (the usleep in the code changed to > 16000) with around 30% cpu usage on a single core. > In my opinion the other method using virtualgl and passing data as jpeg > is not the optimal solution to the optimus problem and feels a bit too > complex. Using the quick and dirty solution, sharing an RGB buffer > between two servers, should be able to get better framerate after some > optimizations since we don't need to do any encoding of the data. > For comparison: I get around 4000fps in glxgears on my GF GT310 using > this method, although I only display 60fps on the intel card. > I've been playing with the idea of running the windump program as a root > window on the intel X display and then running Gnome or whatever on only > the nvidia display. Another thing I've been wanting to test is to use > the intel card's opengl for rendering and putting the XShm image in a > texture. Also syncing with the intel card's refresh rate should be easy > using opengl. > > Florian: You didn't mention any license on the code you posted, would > you like to release it under the GPL? > I'm interested in working some more on this program, but I want to be > able to release my modifications if I manage to achieve anything. Like > markc said, github is excellent for this kind of project. > > Best regards, > Joakim > > 2011-05-01 16:30, Joaquín Ignacio Aramendía skrev: >> El dom, 01-05-2011 a las 12:13 +1000, Pigeon escribió: >>> Hi Florian, >>> >>>> i recently bought a laptop (msi cx640) using nvidia optimus >>>> unknowingly, that it was pretty unusable to me on linux. >>>> however i found a (rather dirty) hack to be able to at least use my >>>> nvidia graphics card, so that i dont have to throw my laptop into the >>>> trashbin. >>> I've been experimenting the same kind of approach by using vnc >>> (x11vnc + vncviewer). I'm running two X servers on the same host, one >>> intel and one nvidia. I have to force the two X servers to be on the >>> same vt to trick both X to "render" at the same time, rather dirty too. >>> >>> Performance isn't great as expected. I got quite a lot of >>> tearing when I was testing with some GL apps/games. But it's usable at >>> least. >>> >>> I haven't tried your windump app yet. I believe you can get >>> some performance gain by using the Xdamage protocol, which x11vnc uses >>> by default. Though I'm not an X expert either. >>> >>> This is on an Alienware M11x-R2 btw. >>> >>> >>> Pigeon. >> >> Small steps for you... big ones for hybrid graphics I think :) > > _______________________________________________ > Mailing list: https://launchpad.net/~hybrid-graphics-linux > Post to : [email protected] > Unsubscribe : https://launchpad.net/~hybrid-graphics-linux > More help : https://help.launchpad.net/ListHelp > _______________________________________________ Mailing list: https://launchpad.net/~hybrid-graphics-linux Post to : [email protected] Unsubscribe : https://launchpad.net/~hybrid-graphics-linux More help : https://help.launchpad.net/ListHelp

