That's cool then - i have written a shellscript
which automatically does the xhost stuff for all
nodes in my hostfile :)

On Mon, Jan 26, 2009 at 9:25 PM, Ralph Castain <r...@lanl.gov> wrote:
>
> On Jan 26, 2009, at 1:20 PM, jody wrote:
>
>> Hi Brian
>>
>>>
>>> I would rather not have mpirun doing an xhost command - I think that is
>>> beyond our comfort zone. Frankly, if someone wants to do this, it is up
>>> to
>>> them to have things properly setup on their machine - as a rule, we don't
>>> mess with your machine's configuration. Makes sys admins upset :-)
>>
>> So what you mean is that the user must do the xhost before using the
>> xceren feature?
>> If not, how else can i have xterms from another machine display locally?
>
> That is correct. I don't think that is -too- odious a requirement - I'm just
> not comfortable modifying access controls from within OMPI since xhost
> persists after OMPI is done with the job.
>
>>
>>>
>>> However, I can check to ensure that the DISPLAY value is locally set and
>>> automatically export it for you (so you don't have to do the -x DISPLAY
>>> option). What I have done is provided a param whereby you can tell us
>>> what
>>> command to use to generate the new screen, with it defaulting to "xterm
>>> -e".
>>> I also allow you to specify which ranks you want displayed this way - you
>>> can specify "all" by giving it a "-1".
>>
>> Cool!
>>
>>> Will hopefully have this done today or tomorrow. It will be in the OMPI
>>> trunk repo for now. I'll send out a note pointing to it so people can
>>> check
>>> all these options out - I would really appreciate the help to ensure
>>> things
>>> are working across as many platforms as possible before we put it in the
>>> official release!
>>
>> I'll be happy to test these new features!
>>
>> Jody
>>
>>>> Hi
>>>> I have written some shell scripts which ease the output
>>>> to an xterm for each processor for normal execution(run_sh.sh),
>>>> gdb (run_gdb.sh), and valgrind (run_vg.sh).
>>>>
>>>> In order for the xterms to be shown on your machine,
>>>> you have to set the DISPLAY variable on every host
>>>> (if this is not done by ssh)
>>>> export DISPLAY=myhost:0.0
>>>>
>>>> on myhost you may have to allow access:
>>>> do
>>>> xhost +<host-name>
>>>> for each machine in your hostfile.
>>>>
>>>> Then start
>>>> mpirun -np 12 -x DISPLAY run_gdb.sh myApp arg1 arg2 arg3
>>>>
>>>> I've attached these little scripts to this mail.
>>>> Feel free to use them.
>>>>
>>>> I've started working on my "complicated" way, i.e.
>>>> wrappers redirecting output via sockets to a server.
>>>>
>>>> Jody
>>>>
>>>> On Sun, Jan 25, 2009 at 1:20 PM, Ralph Castain <r...@lanl.gov> wrote:
>>>>>
>>>>> For those of you following this thread:
>>>>>
>>>>> I have been impressed by the various methods used to grab the output
>>>>> from
>>>>> processes. Since this is clearly something of interest to a broad
>>>>> audience,
>>>>> I would like to try and make this easier to do by adding some options
>>>>> to
>>>>> mpirun. Coming in 1.3.1 will be --tag-output, which will automatically
>>>>> tag
>>>>> each line of output with the rank of the process - this was already in
>>>>> the
>>>>> works, but obviously doesn't meet the needs expressed here.
>>>>>
>>>>> I have done some prelim work on a couple of options based on this
>>>>> thread:
>>>>>
>>>>> 1. spawn a screen and redirect process output to it, with the ability
>>>>> to
>>>>> request separate screens for each specified rank. Obviously, specifying
>>>>> all
>>>>> ranks would be the equivalent of replacing "my_app" on the mpirun cmd
>>>>> line
>>>>> with "xterm my_app". However, there are cases where you only need to
>>>>> see
>>>>> the
>>>>> output from a subset of the ranks, and that is the intent of this
>>>>> option.
>>>>>
>>>>> 2. redirect output of specified processes to files using the provided
>>>>> filename appended with ".rank". You can do this for all ranks, or a
>>>>> specified subset of them.
>>>>>
>>>>> 3. timestamp output
>>>>>
>>>>> Is there anything else people would like to see?
>>>>>
>>>>> It is also possible to write a dedicated app such as Jody described,
>>>>> but
>>>>> that is outside my purview for now due to priorities. However, I can
>>>>> provide
>>>>> technical advice to such an effort, so feel free to ask.
>>>>>
>>>>> Ralph
>>>>>
>>>>>
>>>>> On Jan 23, 2009, at 12:19 PM, Gijsbert Wiesenekker wrote:
>>>>>
>>>>>> jody wrote:
>>>>>>>
>>>>>>> Hi
>>>>>>> I have a small cluster consisting of 9 computers (8x2 CPUs, 1x4
>>>>>>> CPUs).
>>>>>>> I would like to be able to observe the output of the processes
>>>>>>> separately during an mpirun.
>>>>>>>
>>>>>>> What i currently do is to apply the mpirun to a shell script which
>>>>>>> opens a xterm for each process,
>>>>>>> which then starts the actual application.
>>>>>>>
>>>>>>> This works, but is a bit complicated, e.g. finding the window you're
>>>>>>> interested in among 19 others.
>>>>>>>
>>>>>>> So i was wondering is there a possibility to capture the processes'
>>>>>>> outputs separately, so
>>>>>>> i can make an application in which i can switch between the different
>>>>>>> processor outputs?
>>>>>>> I could imagine that could be done by wrapper applications which
>>>>>>> redirect the output over a TCP
>>>>>>> socket to a server application.
>>>>>>>
>>>>>>> But perhaps there is an easier way, or something like this alread
>>>>>>> does
>>>>>>> exist?
>>>>>>>
>>>>>>> Thank You
>>>>>>> Jody
>>>>>>> _______________________________________________
>>>>>>> users mailing list
>>>>>>> us...@open-mpi.org
>>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>>>
>>>>>>>
>>>>>> For C I use a printf wrapper function that writes the output to a
>>>>>> logfile.
>>>>>> I derive the name of the logfile from the mpi_id. It prefixes the
>>>>>> lines
>>>>>> with
>>>>>> a time-stamp, so you also get some basic profile information. I can
>>>>>> send
>>>>>> you
>>>>>> the source code if you like.
>>>>>>
>>>>>> Regards,
>>>>>> Gijsbert
>>>>>>
>>>>>>
>>>>>> _______________________________________________
>>>>>> users mailing list
>>>>>> us...@open-mpi.org
>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>
>>>>> _______________________________________________
>>>>> users mailing list
>>>>> us...@open-mpi.org
>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>
>>>>
>>>>
>>>> <run_gdb.sh><run_vg.sh><run_sh.sh>_______________________________________________
>>>> users mailing list
>>>> us...@open-mpi.org
>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>
>>> _______________________________________________
>>> users mailing list
>>> us...@open-mpi.org
>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>
>> _______________________________________________
>> users mailing list
>> us...@open-mpi.org
>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>
> _______________________________________________
> users mailing list
> us...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/users
>

Reply via email to