Is this can help? http://www.phoronix.com/scan.php?page=news_item&px=OTQwNw
> I've been using this with great success the last couple of days. On my > i3 2.4GHz I can run it at 60fps (the usleep in the code changed to > 16000) with around 30% cpu usage on a single core. > In my opinion the other method using virtualgl and passing data as jpeg > is not the optimal solution to the optimus problem and feels a bit too > complex. Using the quick and dirty solution, sharing an RGB buffer > between two servers, should be able to get better framerate after some > optimizations since we don't need to do any encoding of the data. > For comparison: I get around 4000fps in glxgears on my GF GT310 using > this method, although I only display 60fps on the intel card. > I've been playing with the idea of running the windump program as a root > window on the intel X display and then running Gnome or whatever on only > the nvidia display. Another thing I've been wanting to test is to use > the intel card's opengl for rendering and putting the XShm image in a > texture. Also syncing with the intel card's refresh rate should be easy > using opengl. > > Florian: You didn't mention any license on the code you posted, would > you like to release it under the GPL? > I'm interested in working some more on this program, but I want to be > able to release my modifications if I manage to achieve anything. Like > markc said, github is excellent for this kind of project. > > Best regards, > Joakim > > 2011-05-01 16:30, Joaquín Ignacio Aramendía skrev: > > El dom, 01-05-2011 a las 12:13 +1000, Pigeon escribió: > >> Hi Florian, > >> > >>> i recently bought a laptop (msi cx640) using nvidia optimus > >>> unknowingly, that it was pretty unusable to me on linux. > >>> however i found a (rather dirty) hack to be able to at least use my > >>> nvidia graphics card, so that i dont have to throw my laptop into the > >>> trashbin. > >> I've been experimenting the same kind of approach by using vnc > >> (x11vnc + vncviewer). I'm running two X servers on the same host, one > >> intel and one nvidia. I have to force the two X servers to be on the > >> same vt to trick both X to "render" at the same time, rather dirty too. > >> > >> Performance isn't great as expected. I got quite a lot of > >> tearing when I was testing with some GL apps/games. But it's usable at > >> least. > >> > >> I haven't tried your windump app yet. I believe you can get > >> some performance gain by using the Xdamage protocol, which x11vnc uses > >> by default. Though I'm not an X expert either. > >> > >> This is on an Alienware M11x-R2 btw. > >> > >> > >> Pigeon. > > > > Small steps for you... big ones for hybrid graphics I think :) > > _______________________________________________ > Mailing list: https://launchpad.net/~hybrid-graphics-linux > Post to : [email protected] > Unsubscribe : https://launchpad.net/~hybrid-graphics-linux > More help : https://help.launchpad.net/ListHelp _______________________________________________ Mailing list: https://launchpad.net/~hybrid-graphics-linux Post to : [email protected] Unsubscribe : https://launchpad.net/~hybrid-graphics-linux More help : https://help.launchpad.net/ListHelp

