Hello Matt,
I have attached the output with mat_view for 8 and 40 processors.
I am unsure what is meant by the matrix communicator and the partitioning. I am
using the default behaviour in every case. How can I find this information?
I have attached the log view as well if that helps.
Thanks,
Matt
On 23 Jul 2024, at 9:24 PM, Matthew Knepley <[email protected]> wrote:
You don't often get email from [email protected]<mailto:[email protected]>.
Learn why this is
important<https://urldefense.us/v3/__https://aka.ms/LearnAboutSenderIdentification__;!!G_uCfscf7eWS!defNF55JDHADXFMCPrlWVnASGb8l1sxXg5-10IVx4Ff5FFmO2N003z0BQ80cCU3clrwdPmEGeMWVUhdzckDhFG0VKlPdvbvDJrA$
>
Also, you could run with
-mat_view ::ascii_info_detail
and send the output for both cases. The storage of matrix values is not
redundant, so something else is
going on. First, what communicator do you use for the matrix, and what
partitioning?
Thanks,
Matt
On Mon, Jul 22, 2024 at 10:27 PM Barry Smith
<[email protected]<mailto:[email protected]>> wrote:
This Message Is From an External Sender
This message came from outside your organization.
Send the code.
On Jul 22, 2024, at 9:18 PM, Matthew Thomas via petsc-users
<[email protected]<mailto:[email protected]>> wrote:
This Message Is From an External Sender
This message came from outside your organization.
Hello,
I am using petsc and slepc to solve an eigenvalue problem for sparse matrices.
When I run my code with double the number of processors, the memory usage also
doubles.
I am able to reproduce this behaviour with ex1 of slepc’s hands on exercises.
The issue is occurring with petsc not with slepc as this still occurs when I
remove the solve step and just create and assemble the petsc matrix.
With n=100000, this uses ~1Gb with 8 processors, but ~5Gb with 40 processors.
This was done with petsc 3.21.3, on linux compiled with Intel using Intel-MPI
Is this the expected behaviour? If not, how can I bug fix this?
Thanks,
Matt
--
What most experimenters take for granted before they begin their experiments is
infinitely more interesting than any results to which their experiments lead.
-- Norbert Wiener
https://urldefense.us/v3/__https://www.cse.buffalo.edu/*knepley/__;fg!!G_uCfscf7eWS!defNF55JDHADXFMCPrlWVnASGb8l1sxXg5-10IVx4Ff5FFmO2N003z0BQ80cCU3clrwdPmEGeMWVUhdzckDhFG0VKlPduQ6gjvc$
<https://urldefense.us/v3/__http://www.cse.buffalo.edu/*knepley/__;fg!!G_uCfscf7eWS!defNF55JDHADXFMCPrlWVnASGb8l1sxXg5-10IVx4Ff5FFmO2N003z0BQ80cCU3clrwdPmEGeMWVUhdzckDhFG0VKlPdbv0ojtA$
>
Mat Object: 8 MPI processes
type: mpiaij
rows=100000, cols=100000
total: nonzeros=299998, allocated nonzeros=299998
total number of mallocs used during MatSetValues calls=0
[0] Local rows 12500 nz 37499 nz alloced 37499 mem 0., not using I-node
routines
[0] on-diagonal part: nz 37498
[0] off-diagonal part: nz 1
[1] Local rows 12500 nz 37500 nz alloced 37500 mem 0., not using I-node
routines
[1] on-diagonal part: nz 37498
[1] off-diagonal part: nz 2
[2] Local rows 12500 nz 37500 nz alloced 37500 mem 0., not using I-node
routines
[2] on-diagonal part: nz 37498
[2] off-diagonal part: nz 2
[3] Local rows 12500 nz 37500 nz alloced 37500 mem 0., not using I-node
routines
[3] on-diagonal part: nz 37498
[3] off-diagonal part: nz 2
[4] Local rows 12500 nz 37500 nz alloced 37500 mem 0., not using I-node
routines
[4] on-diagonal part: nz 37498
[4] off-diagonal part: nz 2
[5] Local rows 12500 nz 37500 nz alloced 37500 mem 0., not using I-node
routines
[5] on-diagonal part: nz 37498
[5] off-diagonal part: nz 2
[6] Local rows 12500 nz 37500 nz alloced 37500 mem 0., not using I-node
routines
[6] on-diagonal part: nz 37498
[6] off-diagonal part: nz 2
[7] Local rows 12500 nz 37499 nz alloced 37499 mem 0., not using I-node
routines
[7] on-diagonal part: nz 37498
[7] off-diagonal part: nz 1
Information on VecScatter used in matrix-vector product:
PetscSF Object: 8 MPI processes
type: basic
[0] Number of roots=12500, leaves=1, remote ranks=1
[0] 0 <- (1,0)
[1] Number of roots=12500, leaves=2, remote ranks=2
[1] 0 <- (0,12499)
[1] 1 <- (2,0)
[2] Number of roots=12500, leaves=2, remote ranks=2
[2] 0 <- (1,12499)
[2] 1 <- (3,0)
[3] Number of roots=12500, leaves=2, remote ranks=2
[3] 0 <- (2,12499)
[3] 1 <- (4,0)
[4] Number of roots=12500, leaves=2, remote ranks=2
[4] 0 <- (3,12499)
[4] 1 <- (5,0)
[5] Number of roots=12500, leaves=2, remote ranks=2
[5] 0 <- (4,12499)
[5] 1 <- (6,0)
[6] Number of roots=12500, leaves=2, remote ranks=2
[6] 0 <- (5,12499)
[6] 1 <- (7,0)
[7] Number of roots=12500, leaves=1, remote ranks=1
[7] 0 <- (6,12499)
[0] Roots referenced by my leaves, by rank
[0] 1: 1 edges
[0] 0 <- 0
[1] Roots referenced by my leaves, by rank
[1] 0: 1 edges
[1] 0 <- 12499
[1] 2: 1 edges
[1] 1 <- 0
[2] Roots referenced by my leaves, by rank
[2] 1: 1 edges
[2] 0 <- 12499
[2] 3: 1 edges
[2] 1 <- 0
[3] Roots referenced by my leaves, by rank
[3] 2: 1 edges
[3] 0 <- 12499
[3] 4: 1 edges
[3] 1 <- 0
[4] Roots referenced by my leaves, by rank
[4] 3: 1 edges
[4] 0 <- 12499
[4] 5: 1 edges
[4] 1 <- 0
[5] Roots referenced by my leaves, by rank
[5] 4: 1 edges
[5] 0 <- 12499
[5] 6: 1 edges
[5] 1 <- 0
[6] Roots referenced by my leaves, by rank
[6] 5: 1 edges
[6] 0 <- 12499
[6] 7: 1 edges
[6] 1 <- 0
[7] Roots referenced by my leaves, by rank
[7] 6: 1 edges
[7] 0 <- 12499
MultiSF sort=rank-order
======================================================================================
Resource Usage on 2024-07-24 09:36:33:
Job Id: 121558447.gadi-pbs
Project: y08
Exit Status: 0
Service Units: 0.04
NCPUs Requested: 8 NCPUs Used: 8
CPU Time Used: 00:00:42
Memory Requested: 8.0GB Memory Used: 919.55MB
Walltime requested: 00:05:00 Walltime Used: 00:00:08
JobFS requested: 100.0MB JobFS used: 0B
======================================================================================
Mat Object: 40 MPI processes
type: mpiaij
rows=100000, cols=100000
total: nonzeros=299998, allocated nonzeros=299998
total number of mallocs used during MatSetValues calls=0
[0] Local rows 2500 nz 7499 nz alloced 7499 mem 0., not using I-node
routines
[0] on-diagonal part: nz 7498
[0] off-diagonal part: nz 1
[1] Local rows 2500 nz 7500 nz alloced 7500 mem 0., not using I-node
routines
[1] on-diagonal part: nz 7498
[1] off-diagonal part: nz 2
[2] Local rows 2500 nz 7500 nz alloced 7500 mem 0., not using I-node
routines
[2] on-diagonal part: nz 7498
[2] off-diagonal part: nz 2
[3] Local rows 2500 nz 7500 nz alloced 7500 mem 0., not using I-node
routines
[3] on-diagonal part: nz 7498
[3] off-diagonal part: nz 2
[4] Local rows 2500 nz 7500 nz alloced 7500 mem 0., not using I-node
routines
[4] on-diagonal part: nz 7498
[4] off-diagonal part: nz 2
[5] Local rows 2500 nz 7500 nz alloced 7500 mem 0., not using I-node
routines
[5] on-diagonal part: nz 7498
[5] off-diagonal part: nz 2
[6] Local rows 2500 nz 7500 nz alloced 7500 mem 0., not using I-node
routines
[6] on-diagonal part: nz 7498
[6] off-diagonal part: nz 2
[7] Local rows 2500 nz 7500 nz alloced 7500 mem 0., not using I-node
routines
[7] on-diagonal part: nz 7498
[7] off-diagonal part: nz 2
[8] Local rows 2500 nz 7500 nz alloced 7500 mem 0., not using I-node
routines
[8] on-diagonal part: nz 7498
[8] off-diagonal part: nz 2
[9] Local rows 2500 nz 7500 nz alloced 7500 mem 0., not using I-node
routines
[9] on-diagonal part: nz 7498
[9] off-diagonal part: nz 2
[10] Local rows 2500 nz 7500 nz alloced 7500 mem 0., not using I-node
routines
[10] on-diagonal part: nz 7498
[10] off-diagonal part: nz 2
[11] Local rows 2500 nz 7500 nz alloced 7500 mem 0., not using I-node
routines
[11] on-diagonal part: nz 7498
[11] off-diagonal part: nz 2
[12] Local rows 2500 nz 7500 nz alloced 7500 mem 0., not using I-node
routines
[12] on-diagonal part: nz 7498
[12] off-diagonal part: nz 2
[13] Local rows 2500 nz 7500 nz alloced 7500 mem 0., not using I-node
routines
[13] on-diagonal part: nz 7498
[13] off-diagonal part: nz 2
[14] Local rows 2500 nz 7500 nz alloced 7500 mem 0., not using I-node
routines
[14] on-diagonal part: nz 7498
[14] off-diagonal part: nz 2
[15] Local rows 2500 nz 7500 nz alloced 7500 mem 0., not using I-node
routines
[15] on-diagonal part: nz 7498
[15] off-diagonal part: nz 2
[16] Local rows 2500 nz 7500 nz alloced 7500 mem 0., not using I-node
routines
[16] on-diagonal part: nz 7498
[16] off-diagonal part: nz 2
[17] Local rows 2500 nz 7500 nz alloced 7500 mem 0., not using I-node
routines
[17] on-diagonal part: nz 7498
[17] off-diagonal part: nz 2
[18] Local rows 2500 nz 7500 nz alloced 7500 mem 0., not using I-node
routines
[18] on-diagonal part: nz 7498
[18] off-diagonal part: nz 2
[19] Local rows 2500 nz 7500 nz alloced 7500 mem 0., not using I-node
routines
[19] on-diagonal part: nz 7498
[19] off-diagonal part: nz 2
[20] Local rows 2500 nz 7500 nz alloced 7500 mem 0., not using I-node
routines
[20] on-diagonal part: nz 7498
[20] off-diagonal part: nz 2
[21] Local rows 2500 nz 7500 nz alloced 7500 mem 0., not using I-node
routines
[21] on-diagonal part: nz 7498
[21] off-diagonal part: nz 2
[22] Local rows 2500 nz 7500 nz alloced 7500 mem 0., not using I-node
routines
[22] on-diagonal part: nz 7498
[22] off-diagonal part: nz 2
[23] Local rows 2500 nz 7500 nz alloced 7500 mem 0., not using I-node
routines
[23] on-diagonal part: nz 7498
[23] off-diagonal part: nz 2
[24] Local rows 2500 nz 7500 nz alloced 7500 mem 0., not using I-node
routines
[24] on-diagonal part: nz 7498
[24] off-diagonal part: nz 2
[25] Local rows 2500 nz 7500 nz alloced 7500 mem 0., not using I-node
routines
[25] on-diagonal part: nz 7498
[25] off-diagonal part: nz 2
[26] Local rows 2500 nz 7500 nz alloced 7500 mem 0., not using I-node
routines
[26] on-diagonal part: nz 7498
[26] off-diagonal part: nz 2
[27] Local rows 2500 nz 7500 nz alloced 7500 mem 0., not using I-node
routines
[27] on-diagonal part: nz 7498
[27] off-diagonal part: nz 2
[28] Local rows 2500 nz 7500 nz alloced 7500 mem 0., not using I-node
routines
[28] on-diagonal part: nz 7498
[28] off-diagonal part: nz 2
[29] Local rows 2500 nz 7500 nz alloced 7500 mem 0., not using I-node
routines
[29] on-diagonal part: nz 7498
[29] off-diagonal part: nz 2
[30] Local rows 2500 nz 7500 nz alloced 7500 mem 0., not using I-node
routines
[30] on-diagonal part: nz 7498
[30] off-diagonal part: nz 2
[31] Local rows 2500 nz 7500 nz alloced 7500 mem 0., not using I-node
routines
[31] on-diagonal part: nz 7498
[31] off-diagonal part: nz 2
[32] Local rows 2500 nz 7500 nz alloced 7500 mem 0., not using I-node
routines
[32] on-diagonal part: nz 7498
[32] off-diagonal part: nz 2
[33] Local rows 2500 nz 7500 nz alloced 7500 mem 0., not using I-node
routines
[33] on-diagonal part: nz 7498
[33] off-diagonal part: nz 2
[34] Local rows 2500 nz 7500 nz alloced 7500 mem 0., not using I-node
routines
[34] on-diagonal part: nz 7498
[34] off-diagonal part: nz 2
[35] Local rows 2500 nz 7500 nz alloced 7500 mem 0., not using I-node
routines
[35] on-diagonal part: nz 7498
[35] off-diagonal part: nz 2
[36] Local rows 2500 nz 7500 nz alloced 7500 mem 0., not using I-node
routines
[36] on-diagonal part: nz 7498
[36] off-diagonal part: nz 2
[37] Local rows 2500 nz 7500 nz alloced 7500 mem 0., not using I-node
routines
[37] on-diagonal part: nz 7498
[37] off-diagonal part: nz 2
[38] Local rows 2500 nz 7500 nz alloced 7500 mem 0., not using I-node
routines
[38] on-diagonal part: nz 7498
[38] off-diagonal part: nz 2
[39] Local rows 2500 nz 7499 nz alloced 7499 mem 0., not using I-node
routines
[39] on-diagonal part: nz 7498
[39] off-diagonal part: nz 1
Information on VecScatter used in matrix-vector product:
PetscSF Object: 40 MPI processes
type: basic
[0] Number of roots=2500, leaves=1, remote ranks=1
[0] 0 <- (1,0)
[1] Number of roots=2500, leaves=2, remote ranks=2
[1] 0 <- (0,2499)
[1] 1 <- (2,0)
[2] Number of roots=2500, leaves=2, remote ranks=2
[2] 0 <- (1,2499)
[2] 1 <- (3,0)
[3] Number of roots=2500, leaves=2, remote ranks=2
[3] 0 <- (2,2499)
[3] 1 <- (4,0)
[4] Number of roots=2500, leaves=2, remote ranks=2
[4] 0 <- (3,2499)
[4] 1 <- (5,0)
[5] Number of roots=2500, leaves=2, remote ranks=2
[5] 0 <- (4,2499)
[5] 1 <- (6,0)
[6] Number of roots=2500, leaves=2, remote ranks=2
[6] 0 <- (5,2499)
[6] 1 <- (7,0)
[7] Number of roots=2500, leaves=2, remote ranks=2
[7] 0 <- (6,2499)
[7] 1 <- (8,0)
[8] Number of roots=2500, leaves=2, remote ranks=2
[8] 0 <- (7,2499)
[8] 1 <- (9,0)
[9] Number of roots=2500, leaves=2, remote ranks=2
[9] 0 <- (8,2499)
[9] 1 <- (10,0)
[10] Number of roots=2500, leaves=2, remote ranks=2
[10] 0 <- (9,2499)
[10] 1 <- (11,0)
[11] Number of roots=2500, leaves=2, remote ranks=2
[11] 0 <- (10,2499)
[11] 1 <- (12,0)
[12] Number of roots=2500, leaves=2, remote ranks=2
[12] 0 <- (11,2499)
[12] 1 <- (13,0)
[13] Number of roots=2500, leaves=2, remote ranks=2
[13] 0 <- (12,2499)
[13] 1 <- (14,0)
[14] Number of roots=2500, leaves=2, remote ranks=2
[14] 0 <- (13,2499)
[14] 1 <- (15,0)
[15] Number of roots=2500, leaves=2, remote ranks=2
[15] 0 <- (14,2499)
[15] 1 <- (16,0)
[16] Number of roots=2500, leaves=2, remote ranks=2
[16] 0 <- (15,2499)
[16] 1 <- (17,0)
[17] Number of roots=2500, leaves=2, remote ranks=2
[17] 0 <- (16,2499)
[17] 1 <- (18,0)
[18] Number of roots=2500, leaves=2, remote ranks=2
[18] 0 <- (17,2499)
[18] 1 <- (19,0)
[19] Number of roots=2500, leaves=2, remote ranks=2
[19] 0 <- (18,2499)
[19] 1 <- (20,0)
[20] Number of roots=2500, leaves=2, remote ranks=2
[20] 0 <- (19,2499)
[20] 1 <- (21,0)
[21] Number of roots=2500, leaves=2, remote ranks=2
[21] 0 <- (20,2499)
[21] 1 <- (22,0)
[22] Number of roots=2500, leaves=2, remote ranks=2
[22] 0 <- (21,2499)
[22] 1 <- (23,0)
[23] Number of roots=2500, leaves=2, remote ranks=2
[23] 0 <- (22,2499)
[23] 1 <- (24,0)
[24] Number of roots=2500, leaves=2, remote ranks=2
[24] 0 <- (23,2499)
[24] 1 <- (25,0)
[25] Number of roots=2500, leaves=2, remote ranks=2
[25] 0 <- (24,2499)
[25] 1 <- (26,0)
[26] Number of roots=2500, leaves=2, remote ranks=2
[26] 0 <- (25,2499)
[26] 1 <- (27,0)
[27] Number of roots=2500, leaves=2, remote ranks=2
[27] 0 <- (26,2499)
[27] 1 <- (28,0)
[28] Number of roots=2500, leaves=2, remote ranks=2
[28] 0 <- (27,2499)
[28] 1 <- (29,0)
[29] Number of roots=2500, leaves=2, remote ranks=2
[29] 0 <- (28,2499)
[29] 1 <- (30,0)
[30] Number of roots=2500, leaves=2, remote ranks=2
[30] 0 <- (29,2499)
[30] 1 <- (31,0)
[31] Number of roots=2500, leaves=2, remote ranks=2
[31] 0 <- (30,2499)
[31] 1 <- (32,0)
[32] Number of roots=2500, leaves=2, remote ranks=2
[32] 0 <- (31,2499)
[32] 1 <- (33,0)
[33] Number of roots=2500, leaves=2, remote ranks=2
[33] 0 <- (32,2499)
[33] 1 <- (34,0)
[34] Number of roots=2500, leaves=2, remote ranks=2
[34] 0 <- (33,2499)
[34] 1 <- (35,0)
[35] Number of roots=2500, leaves=2, remote ranks=2
[35] 0 <- (34,2499)
[35] 1 <- (36,0)
[36] Number of roots=2500, leaves=2, remote ranks=2
[36] 0 <- (35,2499)
[36] 1 <- (37,0)
[37] Number of roots=2500, leaves=2, remote ranks=2
[37] 0 <- (36,2499)
[37] 1 <- (38,0)
[38] Number of roots=2500, leaves=2, remote ranks=2
[38] 0 <- (37,2499)
[38] 1 <- (39,0)
[39] Number of roots=2500, leaves=1, remote ranks=1
[39] 0 <- (38,2499)
[0] Roots referenced by my leaves, by rank
[0] 1: 1 edges
[0] 0 <- 0
[1] Roots referenced by my leaves, by rank
[1] 0: 1 edges
[1] 0 <- 2499
[1] 2: 1 edges
[1] 1 <- 0
[2] Roots referenced by my leaves, by rank
[2] 1: 1 edges
[2] 0 <- 2499
[2] 3: 1 edges
[2] 1 <- 0
[3] Roots referenced by my leaves, by rank
[3] 2: 1 edges
[3] 0 <- 2499
[3] 4: 1 edges
[3] 1 <- 0
[4] Roots referenced by my leaves, by rank
[4] 3: 1 edges
[4] 0 <- 2499
[4] 5: 1 edges
[4] 1 <- 0
[5] Roots referenced by my leaves, by rank
[5] 4: 1 edges
[5] 0 <- 2499
[5] 6: 1 edges
[5] 1 <- 0
[6] Roots referenced by my leaves, by rank
[6] 5: 1 edges
[6] 0 <- 2499
[6] 7: 1 edges
[6] 1 <- 0
[7] Roots referenced by my leaves, by rank
[7] 6: 1 edges
[7] 0 <- 2499
[7] 8: 1 edges
[7] 1 <- 0
[8] Roots referenced by my leaves, by rank
[8] 7: 1 edges
[8] 0 <- 2499
[8] 9: 1 edges
[8] 1 <- 0
[9] Roots referenced by my leaves, by rank
[9] 8: 1 edges
[9] 0 <- 2499
[9] 10: 1 edges
[9] 1 <- 0
[10] Roots referenced by my leaves, by rank
[10] 9: 1 edges
[10] 0 <- 2499
[10] 11: 1 edges
[10] 1 <- 0
[11] Roots referenced by my leaves, by rank
[11] 10: 1 edges
[11] 0 <- 2499
[11] 12: 1 edges
[11] 1 <- 0
[12] Roots referenced by my leaves, by rank
[12] 11: 1 edges
[12] 0 <- 2499
[12] 13: 1 edges
[12] 1 <- 0
[13] Roots referenced by my leaves, by rank
[13] 12: 1 edges
[13] 0 <- 2499
[13] 14: 1 edges
[13] 1 <- 0
[14] Roots referenced by my leaves, by rank
[14] 13: 1 edges
[14] 0 <- 2499
[14] 15: 1 edges
[14] 1 <- 0
[15] Roots referenced by my leaves, by rank
[15] 14: 1 edges
[15] 0 <- 2499
[15] 16: 1 edges
[15] 1 <- 0
[16] Roots referenced by my leaves, by rank
[16] 15: 1 edges
[16] 0 <- 2499
[16] 17: 1 edges
[16] 1 <- 0
[17] Roots referenced by my leaves, by rank
[17] 16: 1 edges
[17] 0 <- 2499
[17] 18: 1 edges
[17] 1 <- 0
[18] Roots referenced by my leaves, by rank
[18] 17: 1 edges
[18] 0 <- 2499
[18] 19: 1 edges
[18] 1 <- 0
[19] Roots referenced by my leaves, by rank
[19] 18: 1 edges
[19] 0 <- 2499
[19] 20: 1 edges
[19] 1 <- 0
[20] Roots referenced by my leaves, by rank
[20] 19: 1 edges
[20] 0 <- 2499
[20] 21: 1 edges
[20] 1 <- 0
[21] Roots referenced by my leaves, by rank
[21] 20: 1 edges
[21] 0 <- 2499
[21] 22: 1 edges
[21] 1 <- 0
[22] Roots referenced by my leaves, by rank
[22] 21: 1 edges
[22] 0 <- 2499
[22] 23: 1 edges
[22] 1 <- 0
[23] Roots referenced by my leaves, by rank
[23] 22: 1 edges
[23] 0 <- 2499
[23] 24: 1 edges
[23] 1 <- 0
[24] Roots referenced by my leaves, by rank
[24] 23: 1 edges
[24] 0 <- 2499
[24] 25: 1 edges
[24] 1 <- 0
[25] Roots referenced by my leaves, by rank
[25] 24: 1 edges
[25] 0 <- 2499
[25] 26: 1 edges
[25] 1 <- 0
[26] Roots referenced by my leaves, by rank
[26] 25: 1 edges
[26] 0 <- 2499
[26] 27: 1 edges
[26] 1 <- 0
[27] Roots referenced by my leaves, by rank
[27] 26: 1 edges
[27] 0 <- 2499
[27] 28: 1 edges
[27] 1 <- 0
[28] Roots referenced by my leaves, by rank
[28] 27: 1 edges
[28] 0 <- 2499
[28] 29: 1 edges
[28] 1 <- 0
[29] Roots referenced by my leaves, by rank
[29] 28: 1 edges
[29] 0 <- 2499
[29] 30: 1 edges
[29] 1 <- 0
[30] Roots referenced by my leaves, by rank
[30] 29: 1 edges
[30] 0 <- 2499
[30] 31: 1 edges
[30] 1 <- 0
[31] Roots referenced by my leaves, by rank
[31] 30: 1 edges
[31] 0 <- 2499
[31] 32: 1 edges
[31] 1 <- 0
[32] Roots referenced by my leaves, by rank
[32] 31: 1 edges
[32] 0 <- 2499
[32] 33: 1 edges
[32] 1 <- 0
[33] Roots referenced by my leaves, by rank
[33] 32: 1 edges
[33] 0 <- 2499
[33] 34: 1 edges
[33] 1 <- 0
[34] Roots referenced by my leaves, by rank
[34] 33: 1 edges
[34] 0 <- 2499
[34] 35: 1 edges
[34] 1 <- 0
[35] Roots referenced by my leaves, by rank
[35] 34: 1 edges
[35] 0 <- 2499
[35] 36: 1 edges
[35] 1 <- 0
[36] Roots referenced by my leaves, by rank
[36] 35: 1 edges
[36] 0 <- 2499
[36] 37: 1 edges
[36] 1 <- 0
[37] Roots referenced by my leaves, by rank
[37] 36: 1 edges
[37] 0 <- 2499
[37] 38: 1 edges
[37] 1 <- 0
[38] Roots referenced by my leaves, by rank
[38] 37: 1 edges
[38] 0 <- 2499
[38] 39: 1 edges
[38] 1 <- 0
[39] Roots referenced by my leaves, by rank
[39] 38: 1 edges
[39] 0 <- 2499
MultiSF sort=rank-order
======================================================================================
Resource Usage on 2024-07-24 09:34:31:
Job Id: 121558351.gadi-pbs
Project: y08
Exit Status: 0
Service Units: 0.07
NCPUs Requested: 40 NCPUs Used: 40
CPU Time Used: 00:00:36
Memory Requested: 8.0GB Memory Used: 4.59GB
Walltime requested: 00:05:00 Walltime Used: 00:00:03
JobFS requested: 100.0MB JobFS used: 0B
======================================================================================
****************************************************************************************************************************************************************
*** WIDEN YOUR WINDOW TO 160 CHARACTERS. Use
'enscript -r -fCourier9' to print this document
***
****************************************************************************************************************************************************************
------------------------------------------------------------------ PETSc
Performance Summary:
------------------------------------------------------------------
##########################################################
# #
# WARNING!!! #
# #
# This code was compiled with a debugging option. #
# To get timing results run ./configure #
# using --with-debugging=no, the performance will #
# be generally two or three times faster. #
# #
##########################################################
/home/149/mt3516/island_damping/petsc_test/slepc_test/ex1 on a
arch-linux-c-debug named gadi-cpu-clx-0898.gadi.nci.org.au with 8 processes, by
mt3516 on Wed Jul 24 09:51:44 2024
Using Petsc Release Version 3.21.1, unknown
Max Max/Min Avg Total
Time (sec): 3.080e-02 1.000 3.080e-02
Objects: 0.000e+00 0.000 0.000e+00
Flops: 0.000e+00 0.000 0.000e+00 0.000e+00
Flops/sec: 0.000e+00 0.000 0.000e+00 0.000e+00
Memory (bytes): 1.953e+06 1.000 1.953e+06 1.562e+07
MPI Msg Count: 4.000e+00 2.000 3.500e+00 2.800e+01
MPI Msg Len (bytes): 1.600e+01 2.000 4.000e+00 1.120e+02
MPI Reductions: 8.100e+01 1.000
Flop counting convention: 1 flop = 1 real number operation of type
(multiply/divide/add/subtract)
e.g., VecAXPY() for real vectors of length N --> 2N
flops
and VecAXPY() for complex vectors of length N -->
8N flops
Summary of Stages: ----- Time ------ ----- Flop ------ --- Messages --- --
Message Lengths -- -- Reductions --
Avg %Total Avg %Total Count %Total
Avg %Total Count %Total
0: Main Stage: 3.0757e-02 99.9% 0.0000e+00 0.0% 2.800e+01 100.0%
4.000e+00 100.0% 6.200e+01 76.5%
------------------------------------------------------------------------------------------------------------------------
See the 'Profiling' chapter of the users' manual for details on interpreting
output.
Phase summary info:
Count: number of times phase was executed
Time and Flop: Max - maximum over all processors
Ratio - ratio of maximum to minimum over all processors
Mess: number of messages sent
AvgLen: average message length (bytes)
Reduct: number of global reductions
Global: entire computation
Stage: stages of a computation. Set stages with PetscLogStagePush() and
PetscLogStagePop().
%T - percent time in this phase %F - percent flop in this phase
%M - percent messages in this phase %L - percent message lengths in
this phase
%R - percent reductions in this phase
Total Mflop/s: 10e-6 * (sum of flop over all processors)/(max time over all
processors)
------------------------------------------------------------------------------------------------------------------------
##########################################################
# #
# WARNING!!! #
# #
# This code was compiled with a debugging option. #
# To get timing results run ./configure #
# using --with-debugging=no, the performance will #
# be generally two or three times faster. #
# #
##########################################################
Event Count Time (sec) Flop
--- Global --- --- Stage ---- Total
Max Ratio Max Ratio Max Ratio Mess AvgLen Reduct
%T %F %M %L %R %T %F %M %L %R Mflop/s
------------------------------------------------------------------------------------------------------------------------
--- Event Stage 0: Main Stage
BuildTwoSided 3 1.0 2.1506e-04 1.4 0.00e+00 0.0 1.4e+01 4.0e+00
6.0e+00 1 0 50 50 7 1 0 50 50 10 0
BuildTwoSidedF 2 1.0 1.9223e-04 1.2 0.00e+00 0.0 0.0e+00 0.0e+00
4.0e+00 1 0 0 0 5 1 0 0 0 6 0
MatAssemblyBegin 1 1.0 2.5808e-04 1.0 0.00e+00 0.0 0.0e+00 0.0e+00
6.0e+00 1 0 0 0 7 1 0 0 0 10 0
MatAssemblyEnd 1 1.0 1.0028e-02 1.0 0.00e+00 0.0 2.8e+01 4.0e+00
3.7e+01 33 0 100 100 46 33 0 100 100 60 0
SFSetGraph 1 1.0 1.6253e-05 1.5 0.00e+00 0.0 0.0e+00 0.0e+00
2.0e+00 0 0 0 0 2 0 0 0 0 3 0
SFSetUp 1 1.0 2.0841e-04 1.0 0.00e+00 0.0 2.8e+01 4.0e+00
4.0e+00 1 0 100 100 5 1 0 100 100 6 0
------------------------------------------------------------------------------------------------------------------------
Object Type Creations Destructions. Reports information only for
process 0.
--- Event Stage 0: Main Stage
Matrix 3 3
Vector 2 2
Index Set 2 2
Star Forest Graph 1 1
========================================================================================================================
Average time to get PetscTime(): 3.6601e-08
Average time for MPI_Barrier(): 1.48194e-05
Average time for zero size MPI_Send(): 6.92962e-06
#PETSc Option Table entries:
-log_view # (source: command line)
#End of PETSc Option Table entries
Compiled without FORTRAN kernels
Compiled with full precision matrices (default)
sizeof(short) 2 sizeof(int) 4 sizeof(long) 8 sizeof(void*) 8
sizeof(PetscScalar) 16 sizeof(PetscInt) 4
Configure options: --with-scalar-type=complex --with-mkl_cpardiso
-----------------------------------------
Libraries compiled on 2024-07-19 04:07:25 on gadi-login-09.gadi.nci.org.au
Machine characteristics:
Linux-4.18.0-513.24.1.el8.nci.x86_64-x86_64-with-centos-8.9-Green_Obsidian
Using PETSc directory: /home/149/mt3516/island_damping/matrix_packages/petsc
Using PETSc arch: arch-linux-c-debug
-----------------------------------------
Using C compiler: mpicc -fPIC -Wall -Wwrite-strings -Wno-unknown-pragmas
-Wno-lto-type-mismatch -fstack-protector -fvisibility=hidden -g3 -O0
Using Fortran compiler: mpif90 -fPIC -Wall -ffree-line-length-none
-ffree-line-length-0 -Wno-lto-type-mismatch -Wno-unused-dummy-argument -g -O0
-----------------------------------------
Using include paths:
-I/home/149/mt3516/island_damping/matrix_packages/petsc/include
-I/home/149/mt3516/island_damping/matrix_packages/petsc/arch-linux-c-debug/include
-----------------------------------------
Using C linker: mpicc
Using Fortran linker: mpif90
Using libraries:
-Wl,-rpath,/home/149/mt3516/island_damping/matrix_packages/petsc/arch-linux-c-debug/lib
-L/home/149/mt3516/island_damping/matrix_packages/petsc/arch-linux-c-debug/lib
-lpetsc -Wl,-rpath,/apps/intel-ct/2022.2.0/mkl/lib/intel64
-L/apps/intel-ct/2022.2.0/mkl/lib/intel64 -lmkl_intel_lp64 -lmkl_core
-lmkl_gnu_thread -lmkl_blacs_intelmpi_lp64 -lgomp -ldl -lpthread -lm -lX11
-lgfortran -lstdc++ -lquadmath
-----------------------------------------
##########################################################
# #
# WARNING!!! #
# #
# This code was compiled with a debugging option. #
# To get timing results run ./configure #
# using --with-debugging=no, the performance will #
# be generally two or three times faster. #
# #
##########################################################
======================================================================================
Resource Usage on 2024-07-24 09:51:48:
Job Id: 121558958.gadi-pbs
Project: y08
Exit Status: 0
Service Units: 0.01
NCPUs Requested: 8 NCPUs Used: 8
CPU Time Used: 00:00:03
Memory Requested: 8.0GB Memory Used: 1.0GB
Walltime requested: 00:05:00 Walltime Used: 00:00:03
JobFS requested: 100.0MB JobFS used: 0B
======================================================================================