Typo there: "xceren" stands for "screen" - sorry :)

On Mon, Jan 26, 2009 at 9:20 PM, jody <jody....@gmail.com> wrote:
> Hi Brian
>
>>
>> I would rather not have mpirun doing an xhost command - I think that is
>> beyond our comfort zone. Frankly, if someone wants to do this, it is up to
>> them to have things properly setup on their machine - as a rule, we don't
>> mess with your machine's configuration. Makes sys admins upset :-)
> So what you mean is that the user must do the xhost before using the
> xceren feature?
> If not, how else can i have xterms from another machine display locally?
>>
>> However, I can check to ensure that the DISPLAY value is locally set and
>> automatically export it for you (so you don't have to do the -x DISPLAY
>> option). What I have done is provided a param whereby you can tell us what
>> command to use to generate the new screen, with it defaulting to "xterm -e".
>> I also allow you to specify which ranks you want displayed this way - you
>> can specify "all" by giving it a "-1".
> Cool!
>
>> Will hopefully have this done today or tomorrow. It will be in the OMPI
>> trunk repo for now. I'll send out a note pointing to it so people can check
>> all these options out - I would really appreciate the help to ensure things
>> are working across as many platforms as possible before we put it in the
>> official release!
> I'll be happy to test these new features!
>
> Jody
>
>>> Hi
>>> I have written some shell scripts which ease the output
>>> to an xterm for each processor for normal execution(run_sh.sh),
>>> gdb (run_gdb.sh), and valgrind (run_vg.sh).
>>>
>>> In order for the xterms to be shown on your machine,
>>> you have to set the DISPLAY variable on every host
>>> (if this is not done by ssh)
>>>  export DISPLAY=myhost:0.0
>>>
>>> on myhost you may have to allow access:
>>> do
>>>  xhost +<host-name>
>>> for each machine in your hostfile.
>>>
>>> Then start
>>>  mpirun -np 12 -x DISPLAY run_gdb.sh myApp arg1 arg2 arg3
>>>
>>> I've attached these little scripts to this mail.
>>> Feel free to use them.
>>>
>>> I've started working on my "complicated" way, i.e.
>>> wrappers redirecting output via sockets to a server.
>>>
>>> Jody
>>>
>>> On Sun, Jan 25, 2009 at 1:20 PM, Ralph Castain <r...@lanl.gov> wrote:
>>>>
>>>> For those of you following this thread:
>>>>
>>>> I have been impressed by the various methods used to grab the output from
>>>> processes. Since this is clearly something of interest to a broad
>>>> audience,
>>>> I would like to try and make this easier to do by adding some options to
>>>> mpirun. Coming in 1.3.1 will be --tag-output, which will automatically
>>>> tag
>>>> each line of output with the rank of the process - this was already in
>>>> the
>>>> works, but obviously doesn't meet the needs expressed here.
>>>>
>>>> I have done some prelim work on a couple of options based on this thread:
>>>>
>>>> 1. spawn a screen and redirect process output to it, with the ability to
>>>> request separate screens for each specified rank. Obviously, specifying
>>>> all
>>>> ranks would be the equivalent of replacing "my_app" on the mpirun cmd
>>>> line
>>>> with "xterm my_app". However, there are cases where you only need to see
>>>> the
>>>> output from a subset of the ranks, and that is the intent of this option.
>>>>
>>>> 2. redirect output of specified processes to files using the provided
>>>> filename appended with ".rank". You can do this for all ranks, or a
>>>> specified subset of them.
>>>>
>>>> 3. timestamp output
>>>>
>>>> Is there anything else people would like to see?
>>>>
>>>> It is also possible to write a dedicated app such as Jody described, but
>>>> that is outside my purview for now due to priorities. However, I can
>>>> provide
>>>> technical advice to such an effort, so feel free to ask.
>>>>
>>>> Ralph
>>>>
>>>>
>>>> On Jan 23, 2009, at 12:19 PM, Gijsbert Wiesenekker wrote:
>>>>
>>>>> jody wrote:
>>>>>>
>>>>>> Hi
>>>>>> I have a small cluster consisting of 9 computers (8x2 CPUs, 1x4 CPUs).
>>>>>> I would like to be able to observe the output of the processes
>>>>>> separately during an mpirun.
>>>>>>
>>>>>> What i currently do is to apply the mpirun to a shell script which
>>>>>> opens a xterm for each process,
>>>>>> which then starts the actual application.
>>>>>>
>>>>>> This works, but is a bit complicated, e.g. finding the window you're
>>>>>> interested in among 19 others.
>>>>>>
>>>>>> So i was wondering is there a possibility to capture the processes'
>>>>>> outputs separately, so
>>>>>> i can make an application in which i can switch between the different
>>>>>> processor outputs?
>>>>>> I could imagine that could be done by wrapper applications which
>>>>>> redirect the output over a TCP
>>>>>> socket to a server application.
>>>>>>
>>>>>> But perhaps there is an easier way, or something like this alread does
>>>>>> exist?
>>>>>>
>>>>>> Thank You
>>>>>> Jody
>>>>>> _______________________________________________
>>>>>> users mailing list
>>>>>> us...@open-mpi.org
>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>>
>>>>>>
>>>>> For C I use a printf wrapper function that writes the output to a
>>>>> logfile.
>>>>> I derive the name of the logfile from the mpi_id. It prefixes the lines
>>>>> with
>>>>> a time-stamp, so you also get some basic profile information. I can send
>>>>> you
>>>>> the source code if you like.
>>>>>
>>>>> Regards,
>>>>> Gijsbert
>>>>>
>>>>>
>>>>> _______________________________________________
>>>>> users mailing list
>>>>> us...@open-mpi.org
>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>
>>>> _______________________________________________
>>>> users mailing list
>>>> us...@open-mpi.org
>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>
>>>
>>> <run_gdb.sh><run_vg.sh><run_sh.sh>_______________________________________________
>>> users mailing list
>>> us...@open-mpi.org
>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>
>> _______________________________________________
>> users mailing list
>> us...@open-mpi.org
>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>
>

Reply via email to