Troy,

I've been able to reproduce this. Should have this
corrected shortly.

Thanks,
Tim

> On Mon, 14 Nov 2005 10:38:03 -0700, Troy Telford
> <ttelf...@linuxnetworx.com> wrote:
>
>> My mvapi config is using the Mellanox IB Gold 1.8 IB software release.
>> Kernel 2.6.5-7.201 (SLES 9 SP2)
>>
>> When I ran IMB using mvapi, I received the following error:
>> ***
>> [0,1,2][btl_mvapi_component.c:637:mca_btl_mvapi_component_progress]
>> error
>> in pod
>> [0,1,3][btl_mvapi_component.c:637:mca_btl_mvapi_component_progress]
>> error
>> in pod
>> [0,1,2][btl_mvapi_component.c:637:mca_btl_mvapi_component_progress]
>> error
>> in pod
>> ***
>>
>> Execution (for the mvapi test) is started with:
>> mpirun --prefix $MPI_HOME --mca btl mvapi,self -np 8 -machinefile
>> $work_dir/node.gen1 $work_dir/IMB-MPI1
>
> A few clarifications:  here's the output, by program:
>
> Error when Executing Presta's 'com' test on MVAPI:
> [0,1,1][btl_mvapi_component.c:637:mca_btl_mvapi_component_progress]
> [0,1,0][btld
> error in posting pending send
>
> Error for the 'allred' rest:
> [btl_mvapi_component.c:637:mca_btl_mvapi_component_progress] error in
> posting pending send
> [0,1,5][btl_mvapi_component.c:637:mca_btl_mvapi_component_progress] error
> in posting pending send
> [0,1,1][btl_mvapi_component.c:637:mca_btl_mvapi_component_progress] error
> in posting pending send
> [0,1,6][btl_mvapi_component.c:637:mca_btl_mvapi_component_progress] error
> in posting pending send
>
> For 'Globalop':
> [0,1,2][btl_mvapi_component.c:637:mca_btl_mvapi_component_progress] error
> in posting pending send
> [n54:12267] *** An error occurred in MPI_Reduce
> [n54:12267] *** on communicator MPI_COMM_WORLD
> [n54:12267] *** MPI_ERR_OTHER: known error not in list
> [n54:12267] *** MPI_ERRORS_ARE_FATAL (goodbye)
>
> For IMB:
> [0,1,3][btl_mvapi_component.c:637:mca_btl_mvapi_component_progress]
> [0,1,2][btl_mvapi_component.c:637:mca_btl_mvapi_component_progress] error
> in posting pending send
> error in posting pending send
> [0,1,3][btl_mvapi_component.c:637:mca_btl_mvapi_component_progress] error
> in posting pending send
>
> mvapi did run HPL successfully, but it hasn't finished running HPCC just
> yet.
>
> Also, I can say that I've been successful in running HPL and HPCC over GM
> (in fact, I've been able to run IMB, Presta, HPCC, and HPL with no issues
> using GM.  This pleases me)
>
> I've just finished a build of RC7, so I'll go give that a whirl and
> report.
> --
> Troy Telford
> Linux Networx
> ttelf...@linuxnetworx.com
> (801) 649-1356
> _______________________________________________
> users mailing list
> us...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/users
>

Reply via email to