750 likes | 1.13k Views
Overview. Discretization IssuesCell versus Vertex BasedGrid Alignment ProblemsReconstruction for 2nd order accuracyGradient reconstruction issuesArtificial DissipationLimitersViscous DiscretizationsChoice of element typeFull NS termsGrid resolution issuesSolvers and Scalability Conclusi
E N D
1. Unstructured Mesh Discretizations and Solvers for Computational Aerodynamics Dimitri J. Mavriplis
University of Wyoming
2. Overview
Discretization Issues
Cell versus Vertex Based
Grid Alignment Problems
Reconstruction for 2nd order accuracy
Gradient reconstruction issues
Artificial Dissipation
Limiters
Viscous Discretizations
Choice of element type
Full NS terms
Grid resolution issues
Solvers and Scalability
Conclusions
3. Cell Centered vs Vertex-Based Tetrahedral Mesh contains 5 to 6 times more cells than vertices
Hexahedral meshes contain same number of cells and vertices (excluding boundary effects)
Prismatic meshes: cells = 2X vertices
Tetrahedral cells : 4 neighbors
Vertices: 14 neighbors on average
4. Cell Centered vs Vertex-Based On given mesh:
Cell centered discretization: Higher accuracy
Vertex discretization: Lower cost
Equivalent Accuracy-Cost Comparisons Difficult
Often based on equivalent numbers of surface unknowns (2:1 for tet meshes)
Levy (1999)
Yields advantage for vertex-discretization
5. Example: DLR-F4 Wing-body (AIAA Drag Prediction Workshop)
6. DLRF4-F6 Test Cases (DPW) Wing-Body Configuration
Transonic Flow
Mach=0.75, Incidence = 0 degrees, Reynolds number=3,000,000
7. Illustrative Example: DLR-F4 NSU3D: vertex-based discretization
Grid : 48K boundary pts, 1.65M pts (9.6M cells)
USM3D: cell-centered discretization
Grid : 50K boundary cells, 2.4M cells (414K pts)
Uses wall functions
NSU3D: on cell centered type grid
Grid: 46K boundary cells, 2.7M cells (470K pts)
8. Cell versus Vertex Discretizations Similar Lift for both codes on cell-centered grid
Baseline NSU3D (finer vertex grid) has lower lift
9. Cell versus Vertex Discretizations Pressure drag
Wall treatment discrepancies
NSU3D : cell centered grid
High drag, (10 to 20 counts)
Grid too coarse for NSU3D
Inexpensive computation
USM3D on cell-centered grid closer to NSU3D on vertex grid
10. Vertex vs. Cell-Centered Discretizations For tetrahedral mesh :
N vertices
6N cells
7N edges
12N Faces (triangles)
Cell centered approach has 6 times more d.o.f. on same grid as vertex scheme
But vertex scheme on 6 times finer grid is more accurate than cell-centered scheme
Vertex scheme has 7 fluxes per cv
Cell centered scheme has 12/6=2 fluxes per cv
Differences less pronounced on mixed element grids
AIAA Drag Prediction Workshop Practice:
“Equivalent” vertex grid ~ 3 times finer than cell centered grid (not 6 times)
Computational overheads favor vertex approach (in our opinion)
Cell centered schemes have other advantages
Easier grid generation and file transfer/archiving
Longer term objective:
Single code can be run cell or vertex centered (graph based)
11. Boundary Conditions Element of boundary is a face (not a vertex)
Unambiguous BC prescription requires face based implementation
Weak form for vertex discretizations
12. Grid Resolution and Discretization Issues
Choice of discretization and effect of dissipation (intricately linked)
Cells versus points
Discretization formulations
Grid resolution requirements
Choice of element type
Grid resolution issues
Grid convergence
13. Spatial Discretization Mixed Element Meshes
Tetrahedra, Prisms, Pyramids, Hexahedra
Control Volume Based on Median Duals
Fluxes based on edges
Single edge-based data-structure represents all element types
14. Mixed-Element Discretizations Edge-based data structure
Building block for all element types
Reduces memory requirements
Minimizes indirect addressing / gather-scatter
Graph of grid = Discretization stencil
Implications for solvers, Partitioners
15. Alignment Problems Structured grids can be aligned with flow features
Specific examples of unstructured grid alignment
Prismatic layers in boundary layer
More difficult for shocks/shear layers
Mesh generation based on structured mesh methods
Quad/hex/prism element types
Mesh adaptation through point movement
Possibly adjoint based
16. Adjoint Driven Shock Fitting Optimization of Mesh Based on Minimizing Error from “Exact Solution”
17. Upwind Discretization
18. Matrix Artificial Dissipation
19. Entropy Fix L matrix: diagonal with eigenvalues:
u, u, u, u+c, u-c
Robustness issues related to vanishing eigenvalues
Limit smallest eigenvalues as fraction of largest eigenvalue: |u| + c
u = sign(u) * max(|u|, d(|u|+c))
u+c = sign(u+c) * max(|u+c|, d(|u|+c))
u – c = sign(u -c) * max(|u-c|, d(|u|+c))
20. Entropy Fix u = sign(u) * max(|u|, d(|u|+c))
u+c = sign(u+c) * max(|u+c|, d(|u|+c))
u – c = sign(u -c) * max(|u-c|, d(|u|+c))
d = 0.1 : typical value for enhanced robustness
d = 1.0 : Scalar dissipation
- L becomes scaled identity matrix
T |L| T-1 becomes scalar quantity
Simplified (lower cost) dissipation operator
Applicable to upwind and art. dissipation schemes
21. Discretization Formulations Examine effect of discretization type and parameter variations on drag prediction
Effect on drag polars for DLR-F4:
Matrix artificial dissipation
Dissipation levels
Entropy fix
Low order blending
Upwind schemes
Gradient reconstruction
Entropy fix
Limiters
22. Effect of Artificial Dissipation Level Increased accuracy through lower dissipation coef.
Potential loss of robustness
23. Effect of Entropy Fix for Artificial Dissipation Scheme Insensitive to small values of d=0.1, 0.2
High drag values for large d and scalar scheme
24. Effect of Low-Order Dissipation Blending for Shock Capturing Lift and drag relatively insensitive
Generally not recommended for transonics
25. Effect of Artificial Dissipation
26. Comparison of Discretization Formulation (Art. Dissip vs. Grad. Rec.) Least squares approach slightly more diffusive (?)
Extremely sensitive to entropy fix value
Unweighted LS Gradient extremely inaccurate in BL region
AIAA Paper 2003-3986: Revisiting the LS Gradient
27. Unweighted Least-Squares Gradient Accuracy compromised in regions of high-stretching and moderate curvature
28. Unweighted Least-Squares Gradient Accuracy compromised in regions of high-stretching and moderate curvature
Distance Weighting cures problem but lacks robustness
29. Effect of Limiters on Upwind Discretization Limiters reduces accuracy, increase robustness
Less sensitive to non-monotone limiters
30. Effect of Discretization Type
31. Effect of Element Type Right angle tetrahedra produced in boundary layer regions
Highly stretched elements for efficiency
Non obtuse angle requirement for accuracy
Diagonal edge has face not aligned with flow features
Problematic ?
De-emphasize diagonal edges: containment dual cv
32. Effect of Element Type Alternate strategy: Remove culprit edges
Mesh generation task
Semi-structured tetrahedra combinable into prisms
Prism elements of lower complexity (fewer edges)
No significant accuracy benefit (Aftosmis et. al. 1994 in 2D)
33. Effect of Element Type in BL Region Little overall effect on accuracy
Potential differences between two codes
Further grid refinement shows increased discrepancies (Lee-Rausch et al. (2003, 2004)
34. Grid Convergence Study (DLR-F4)
35. Viscous Term Formulation Vertex-based: Linear Galerkin Finite Elements
Extra stencil pts on hybrid elements
Edge data-structure insufficient
Exact Jacobian construction
36. Viscous Term Formulation Gradients of Gradients:
Extended stencil (neighbors of neighbors)
Odd-even decoupling (stencil 2h)
Multi-dimensional thin layer
Laplacian of velocity: Incompressible NS
Inconsistent Laplacian on edge-data-structure
Consistent for orthogonal prismatic BL grids
Hybrid approach:
Laplacian on edges
Gradients of gradients for remaining terms
Relieves odd-even coupling problem
Retains extended stencil (inexact Jacobian)
37. Sensitivity to Navier-Stokes Terms DPW2 Wing-Body
Mach=0.75, Incidence=0 degrees, Re=3 million
Regions of separated flow
Differences not significant
38. Grid Resolution Issues Possibly greatest impediment to reliable RANS drag prediction
Promise of adaptive meshing held back by development of adequate error estimators
Unstructured mesh requirement similar to structured mesh requirements
200 to 500 vertices chordwise (cruise)
Lower optimal spanwise resolution
Y+ of order 1 required in BL
39. Effect of Normal Spacing in BL Inadequate resolution under-predicts skin friction
Direct influence on drag prediction
40. Effect of Normal Resolution for High-Lift (c/o Anderson et. AIAA J. Aircraft, 1995) Indirect influence on drag prediction
Easily mistaken for poor flow physics modeling
41. DPW3 Wing1-Wing2 Cases
42. W1-W2 Grid Convergence Study
43. W1-W2 Grid Convergence Study
44. W1-W2 Results Discrepancy between UW and Cessna Results
Importance of consistent family of grids
45. W1-W2 Results Removing effect of lift-induced drag :
Results on both grid families converge consistently
46. DPW2/3 Configurations Up to 72M point meshes
47. Sensitivity to Dissipation Levels Drag is grid converging
Sensitivity to dissipation decreases as expected
48. 65M pt mesh Results 10% drop in CL at AoA=0o: closer to experiment
Drop in CD: further from experiment
Same trends at Mach=0.3
Little sensitivity to dissipation
49. Grid Specifications 65 million pt grid
50. Grid Convergence Grid convergence apparent using self-similar family of grids
Large discrepancies possible across grid families
Sensitive areas
Separation, Trailing edge
Pathological cases ?
Would grid families converge to same result limit of infinite resolution ?
i.e. Do we have consistency ?
Due to element types ?
51. Structured vs Unstructured Drag Prediction (AIAA workshop results) Similar predictive ability for both approaches
More scatter for structured methods
More submissions/variations for structured methods
53. Unstructured vs Structured (Transonics) Considerable scatter in both cases
No clear advantage of one method over the other in terms of accuracy
DPW3 Observation:
Core set of codes which:
Agree remarkably well with each other
Span all types of grids
Structured, Overset, Unstructured
Have been developed and used extensively for transonic aerodynamics
54. Solution Methodologies Explicit no-longer acceptable (40M pt grids)
Implicit
Locally or globally
Multigrid
Linear or non-linear (FAS)
Preconditioned Newton-Krylov
Preconditioners are key
Any of above iterative methods
Matrix based (ILU)
55. Solution Methodology To solve R(w) = 0 (steady or unsteady residual)
Newton’s method:
Requires storage/inversion of Jacobian (too big for 2nd order scheme)
Replace with 1st order Jacobian
Stored as block Diagonals [D] (for each vertex) and off-diagonals [O] (2 for each edge)
Use block Jacobi or Gauss-Seidel to invert Jacobian at each Newton iteration using subiteration k:
56. Solution Methodology Corresponds to linear Jacobi/Gauss-Seidel in many unstructured mesh solvers
Alternately, replace Jacobian simply by [D] (i.e. drop [O] terms) (Point implicit)
Non-linear residual must now be updated at every iteration (no subiterations)
Corresponds to non-linear Jacobi/Gauss-Seidel
57. Solution Methodologies In almost all applications, reduced Jacobian is used for linear and/or non-linear solvers
Nearest neighbor stencil
Reduced memory footprint
Can be viewed as:
Defect correction scheme
Preconditioning strategy
Preconditioner = 1st order Jacobian
1st order multigrid coarse level discretizations
Inherent limit on convergence efficiency
58. Non-Linear vs Linear Solvers
Expense of non-linear solver dominated by residual evaluation (non-linear term)
Expense of linear solver determined only by stencil topology (once Jacobian has been constructed)
Memory requirements of linear solver can be considerably higher
1st order Jacobian: ~350 words per vertex
Point implicit: 25 words per vertex
59. Non-Linear vs Linear Solvers In most cases:
Linear solver faster per iteration
Non-linear solver requires much less memory
In asymptotic limit, both deliver identical convergence rates
Linear solver can be less robust at startup
Prefer :
Non-linear solver for steady-state
Linear solver for unsteady time-implicit problems
60. Preconditioned AMG Solver Point or line-implicit solver
Reduces stiffness due to anisotropy
Managable memory overhead (in non-linear form)
Agglomeration multigrid
Convergence rate independent of grid resolution (approximately)
Can be implemented as linear or non-linear solver or preconditioner for GMRES
61. Method of Solution Line-implicit solver
62. Agglomeration Multigrid Agglomeration Multigrid solvers for unstructured meshes
Coarse level meshes constructed by agglomerating fine grid cells/equations
63. Agglomeration Multigrid
64. Agglomeration Multigrid
65. Agglomeration Multigrid
66. Agglomeration Multigrid
67. Parallelization through Domain Decomposition Intersected edges resolved by ghost vertices
Generates communication between original and ghost vertex
Handled using MPI and/or OpenMP (Hybrid implementation)
Local reordering within partition for cache-locality
Multigrid levels partitioned independently
Match levels using greedy algorithm
Optimize intra-grid communication vs inter-grid communication
68. Partitioning (Block) Tridiagonal Lines solver inherently sequential
Contract graph along implicit lines
Weight edges and vertices
Partition contracted graph
Decontract graph
Guaranteed lines never broken
Possible small increase in imbalance/cut edges
69. Partitioning Example 32-way partition of 30,562 point 2D grid
Unweighted partition: 2.6% edges cut, 2.7% lines cut
Weighted partition: 3.2% edges cut, 0% lines cut
70. Partitioning Example 32-way partition of 30,562 point 2D grid
Unweighted partition: 2.6% edges cut, 2.7% lines cut
Weighted partition: 3.2% edges cut, 0% lines cut
71. Line Solver Multigrid Convergence
72. (Multigrid) Preconditioned Newton Krylov Mesh independent property of Multigrid
GMRES effective (in asymptotic range) but requires extra memory
73. Scalability Near ideal speedup for 72M pt grid on 2008 cpus of NASA Columbia Machine
Homogeneous Data-Structure
Near perfect load balancing
Near Optimal Partitioners
74. Conclusions For transonics
Equivalent accuracy on equivalent grids
Equivalent or superior solution technology
Superior scalability
Indirect addressing and memory overheads
Problems remain particularly for high-speed flows
Alignment
Robustness/Accuracy
Gradient reconstruction
Viscous terms and element type
Grid Convergence and consistency questions
75. Discretization Governing Equations: Reynolds Averaged Navier-Stokes Equations
Conservation of Mass, Momentum and Energy
Single Equation turbulence model (Spalart-Allmaras)
Convection-Diffusion – Production
Vertex-Based Discretization
2nd order upwind finite-volume scheme
6 variables per grid point
Flow equations fully coupled (5x5)
Turbulence equation uncoupled