Search found 40 matches
- Yesterday, 17:14
- Forum: ElmerSolver
- Topic: Variable boundary conditon
- Replies: 2
- Views: 23
Re: Variable boundary conditon
I knew I was doing something stupid! Thanks!
- Yesterday, 01:32
- Forum: ElmerSolver
- Topic: Variable boundary conditon
- Replies: 2
- Views: 23
Variable boundary conditon
I thought my first test of a variable boundary worked, but I can not replicate it. I have the program loaded in the sif using Boundary Condition 1 Target Boundaries(3) = 1 2 3 Variable Potential Real Procedure "timestep" "timestep" End And it does get loaded when fed to ElmerSolv...
- 27 Apr 2024, 18:44
- Forum: Bug reports
- Topic: High order infinite loop during partition/ parallel run fails
- Replies: 3
- Views: 330
Re: High order infinite loop during partition/ parallel run fails
The p-elements do not work with StatElecSolve.F90:
I would have fun helping to make that work, but I doubt I'd be very efficient at it.
Code: Select all
IF( Found ) THEN
IF(str(1:2) == 'p:') CALL Fatal('StatElecSolver_init','No support for p-elements in solver!')
END IF
- 25 Apr 2024, 22:53
- Forum: Bug reports
- Topic: High order infinite loop during partition/ parallel run fails
- Replies: 3
- Views: 330
High order infinite loop during partition/ parallel run fails
I have a mesh which is hexahedral with 27 points per element. The object is a hollow cylinder with 2 units radius and 40 units tall. Attempting to make it a parallel using command ElmerGrid 2 2 <directory> -partition 1 1 16 caused a seg fault, and a little debugging found it is in egnative.c when at...
- 25 Apr 2024, 22:27
- Forum: ElmerSolver
- Topic: Parallel grid seg fault
- Replies: 10
- Views: 554
Re: Parallel grid seg fault
I like higher order because it allows a cruder mesh. I will post it as a bug report. Thank you for helping me get around it!
- 25 Apr 2024, 18:07
- Forum: ElmerSolver
- Topic: Parallel grid seg fault
- Replies: 10
- Views: 554
Re: Parallel grid seg fault
That works! I will make note of this command and ignore the fun of debugging!
Thanks, I'm really happy that gets me back to physics.
Thanks, I'm really happy that gets me back to physics.
- 25 Apr 2024, 17:30
- Forum: ElmerSolver
- Topic: Parallel grid seg fault
- Replies: 10
- Views: 554
Re: Parallel grid seg fault
Yes, it does work for linear elements. Saving mesh for 16 partitions part elements nodes shared bc elems 1 51300 55233 1083 9102 2 51300 54150 2166 7050 3 51300 54150 2166 7050 4 51300 54150 2166 7050 5 51300 54150 2166 7050 6 51300 54150 2166 7050 7 51300 54150 2166 7050 8 51300 54150 2166 7050 9 5...
- 25 Apr 2024, 00:30
- Forum: ElmerSolver
- Topic: Parallel grid seg fault
- Replies: 10
- Views: 554
Re: Parallel grid seg fault
On the element which causes the infinite loop, there are 4 faces in the boundary file: From mesh.boundary file: 46851 34 14398 820794 409 168389 3583 10 753 189569 3983 770 189567 189570 116219 4 820794 0 409 417324 3583 10 5842 462484 3983 5879 462482 462485 116903 2 820794 0 409 10 753 463115 5842...
- 24 Apr 2024, 23:50
- Forum: ElmerSolver
- Topic: Parallel grid seg fault
- Replies: 10
- Views: 554
Re: Parallel grid seg fault
Looks like an infinite loop, I assume a pointer is going out of bounds somewhere. I put in printfs in egnative.c: GetElementSide(elemind,side,normal,data,&sideind2[0],&sideelemtype2); printf("after getelements side = %d %d %d\n", side, sideelemtype2, sideelemtype); fflush(stdout); ...
- 24 Apr 2024, 23:11
- Forum: ElmerSolver
- Topic: Parallel grid seg fault
- Replies: 10
- Views: 554
Re: Parallel grid seg fault
Same error: $ ElmerGrid 2 2 deuteron_chamber -partdual -metiskway 16 Starting program Elmergrid, compiled on Mar 10 2024 Elmergrid reading in-line arguments Using dual (elemental) graph in partitioning. The mesh will be partitioned with Metis to 16 partitions. Output will be saved to file deuteron_c...