[deal.II] Modeling Stokes flow in and around a droplet using the approach of step-46

2016-10-28 Thread Oded Yaakobi
Hello, I am working on modeling a phenomenon that is similar to Stokes flow in and around a droplet. In my model there are two sets of Stokes equations that govern the flow in each one of the media that are made of different materials – one set for the flow in the droplet, and the other f

Re: [deal.II] Re: post processing a vector valued solution

2016-10-28 Thread thomas stephens
Your suggestions worked exactly as advertised. I've copy-pasta's my updated code below for completeness. > Thomas, > your code looks essentially correct, though it can be improved: > > > Now, for my attempt at this: > > > > | > > template(int spacedim> > > void data_dot_normal (Vector

Re: [deal.II] Multiplie mpi instances of code

2016-10-28 Thread Timo Heister
> I must reinstall Petsc and deal.ii again. Is there a need to reinstall P4est > as well ? p4est also uses MPI so yes. By the way: after you install deal.II, you should run "make test" in the build directory. I am pretty sure it would have detected this problem. -- Timo Heister http://www.math.

Re: [deal.II] Multiplie mpi instances of code

2016-10-28 Thread RAJAT ARORA
Respected Professor, Yes, I remember this discussion regarding not using --download mpich tag while installing Petsc happening on github as well. Yes, I think I should now do it. Although it still remains a mystery to me, how and why openmpi got installed and is now clashing with mpich. Anyway

Re: [deal.II] Multiplie mpi instances of code

2016-10-28 Thread Wolfgang Bangerth
On 10/28/2016 11:49 AM, RAJAT ARORA wrote: I think that openmpi and Mpich are clashing now. (I didn't even know that both were installed upto now). I think, if I can disable openmpi somehow, I will be in good shape. This is because, I compiled deal.ii with mpich library which was installed by p

Re: [deal.II] Multiplie mpi instances of code

2016-10-28 Thread RAJAT ARORA
I am surprised why it is not working. I can't recall what has changed. I don't remember installing any new libraries. It was working till Monday. Also, I have installed petsc with --download mpich tag but donot have any other mpi installation. and more importantly, it was working until monday.

Re: [deal.II] Deal.ii works on Linux subsystem in Windows 10

2016-10-28 Thread Timo Heister
I also played with bash for windows last week and got the same conclusion: it is super easy to get it to work. I was too busy to write up an FAQ entry, but I agree that this is probably the best option on windows right now. > I would guess that one can also just use Eclipse, which is free and wil

Re: [deal.II] Multiplie mpi instances of code

2016-10-28 Thread Timo Heister
>> Should I reinstall my mpi library ? Does deal.ii will also need to be >> recompiled then ? > > I don't know. You say that it suddenly stopped working. Did it work > correctly before? What has changed in the meantime? This can happen if you have conflicting mpi libraries installed (for example l

Re: [deal.II] Multiplie mpi instances of code

2016-10-28 Thread Wolfgang Bangerth
On 10/28/2016 09:55 AM, RAJAT ARORA wrote: I am just calling dealii::Utilities::MPI::MPI_InitFinalize mpi_initialization(argc, argv, 1) once. I am not making an explicit call to MPI_Init anywhere. Yes, calling MPI_InitFinalize is enough. Should I reinstall my mpi library ? Does deal.ii wil

Re: [deal.II] Multiplie mpi instances of code

2016-10-28 Thread RAJAT ARORA
Respected Prof. Bangerth, I am just calling dealii::Utilities::MPI::MPI_InitFinalize mpi_initialization(argc, argv, 1) once. I am not making an explicit call to MPI_Init anywhere. Should I reinstall my mpi library ? Does deal.ii will also need to be recompiled then ? On Friday, October 28, 2

Re: [deal.II] Multiplie mpi instances of code

2016-10-28 Thread Wolfgang Bangerth
On 10/28/2016 09:47 AM, RAJAT ARORA wrote: Yes, you are right :). When running with 4 mpi processes, all threads print their rank as 0. Why is it broken all of a sudden. What is this problem called ? How do I correct it ? I don't know. Are you calling MPI_Init, for example? W. -- ---

Re: [deal.II] Multiplie mpi instances of code

2016-10-28 Thread RAJAT ARORA
Respected Prof. Bangerth, Yes, you are right :). When running with 4 mpi processes, all threads print their rank as 0. Why is it broken all of a sudden. What is this problem called ? How do I correct it ? On Friday, October 28, 2016 at 11:39:33 AM UTC-4, Wolfgang Bangerth wrote: > > On 10/28/20

Re: [deal.II] Imposing distributed pressure on the curved edge of a circle(sphere in 3D)

2016-10-28 Thread Wolfgang Bangerth
On 10/24/2016 01:41 PM, benhour.amiria...@gmail.com wrote: for (typename Triangulation ::active_cell_iterator cell=triangulation.begin_active(); cell!=triangulation.end(); ++cell) *for (unsigned int f=0; f::faces_per_cell; ++f)* *if (cell->face(f)->at_bo

Re: [deal.II] Multiplie mpi instances of code

2016-10-28 Thread Wolfgang Bangerth
On 10/28/2016 09:34 AM, RAJAT ARORA wrote: To run the code on n processes, I used to run the command mpirun -np ./ But, now, when I run it using this command, it runs n programs with n mpi processes each. It is like I have executed mpirun -np ./ command n times and the compiler is then run

[deal.II] Multiplie mpi instances of code

2016-10-28 Thread RAJAT ARORA
Hello all, I am working on a 3D solid mechanics problem using deal.ii. To run the code on n processes, I used to run the command mpirun -np ./ But, now, when I run it using this command, it runs n programs with n mpi processes each. It is like I have executed mpirun -np ./ command n times

Re: [deal.II] Re: what do SolverControl do for Direct Solvers?

2016-10-28 Thread Bruno Turcksin
Hamed, 2016-10-28 10:36 GMT-04:00 Hamed Babaei : > If I solve a problem with one million DOFs on a cluster of 64 processors, > every processor will be dealing with around 15000 DOFs. > So you mean even in that case direct solver would be slow? I thought direct > solver would solve the problem only

Re: [deal.II] Re: boundary conditions

2016-10-28 Thread Retired Replicant.
Many thanks Wolfgang and Daneil. I think Wolfgang's answer is illuminating. On Friday, October 28, 2016 at 3:13:56 PM UTC+2, Daniel Arndt wrote: > > so the uniform flux from left and right of the rectangle implies periodic >> boundary condition. But the K ( hydraulic conductivity) is a funct

Re: [deal.II] Re: what do SolverControl do for Direct Solvers?

2016-10-28 Thread Hamed Babaei
Dear Bruno, If I solve a problem with one million DOFs on a cluster of 64 processors, every processor will be dealing with around 15000 DOFs. So you mean even in that case direct solver would be slow? I thought direct solver would solve the problem only on those DOFs that it owns. Best, Hamed

Re: [deal.II] Re: boundary conditions

2016-10-28 Thread Daniel Arndt
> > so the uniform flux from left and right of the rectangle implies periodic > boundary condition. But the K ( hydraulic conductivity) is a function of > (x,y). > If we want to enforce the periodic boundary condition, should we expect to > have a condition on K ? should K be periodic as well f

Re: [deal.II] Re: boundary conditions

2016-10-28 Thread Wolfgang Bangerth
On 10/28/2016 06:18 AM, Retired Replicant. wrote: I am sorry I can not formulate my problem more clearly. This is another try. so the uniform flux from left and right of the rectangle implies periodic boundary condition. But the K ( hydraulic conductivity) is a function of (x,y). If we want to

Re: [deal.II] Re: what do SolverControl do for Direct Solvers?

2016-10-28 Thread Bruno Turcksin
Hamed, 2016-10-27 19:03 GMT-04:00 Hamed Babaei : > I've been using the default one, Amesos_KLU, and I checked > Amesos_Superludist but trilinos doesn't recognize it. > when we installed dealii and its dependencies on the cluster, I think we > didn't uncomment the "#PACKAGES="${PACKAGES} once:super

Re: [deal.II] Re: boundary conditions

2016-10-28 Thread Retired Replicant.
I am sorry I can not formulate my problem more clearly. This is another try. so the uniform flux from left and right of the rectangle implies periodic boundary condition. But the K ( hydraulic conductivity) is a function of (x,y). If we want to enforce the periodic boundary condition, should we