Skip to content
Merged
Show file tree
Hide file tree
Changes from 22 commits
Commits
Show all changes
25 commits
Select commit Hold shift + click to select a range
d310f06
build: Add PyTwin to doc dependencies
EDCarman Jun 5, 2025
be4e277
docs: Initial static ROM example
EDCarman Jun 7, 2025
33fd387
Merge branch 'ansys:main' into docs/rom_data_generation_example
EDCarman Jun 7, 2025
a256d7b
docs: refactor example
EDCarman Jun 7, 2025
8ed05f3
docs: remove local .db file, refactor script
EDCarman Jun 10, 2025
5b7fa76
docs: separate MAPDL DoE and export runs in example
EDCarman Jun 15, 2025
91897f7
docs: add Pytwin to links and intersphinx mapping
EDCarman Jun 15, 2025
708e520
docs: add Static ROM example description
EDCarman Jun 15, 2025
f6a149b
docs: undo commenting out filename_pattern
EDCarman Jun 15, 2025
75570b9
Merge branch 'ansys:main' into docs/rom_data_generation_example
EDCarman Jun 15, 2025
e2b75a4
build: Undo addition of PyTwin to doc dependencies
EDCarman Jun 15, 2025
28d3d99
docs: fixed code-block vailing Vale.Spelling
EDCarman Jun 16, 2025
253e3f3
docs: add corrections suggested by sourcery-ai
EDCarman Jun 16, 2025
133efa6
docs: sort scoping IDs
EDCarman Jun 16, 2025
684bcc4
Merge branch 'main' into docs/rom_data_generation_example
germa89 Jun 16, 2025
c7fd971
chore: adding changelog file 4011.miscellaneous.md [dependabot-skip]
pyansys-ci-bot Jun 16, 2025
e7afbd0
chore: adding changelog file 4011.miscellaneous.md [dependabot-skip]
pyansys-ci-bot Jun 16, 2025
84c6432
docs: address sourcery-ai corrections
EDCarman Jun 16, 2025
0e3685c
Merge branch 'migration/pr-4009' of https://github.com/ansys/pymapdl …
EDCarman Jun 16, 2025
49954d3
chore: adding changelog file 4011.miscellaneous.md [dependabot-skip]
pyansys-ci-bot Jun 16, 2025
4196026
chore: adding changelog file 4011.miscellaneous.md [dependabot-skip]
pyansys-ci-bot Jun 16, 2025
75c1e2f
docs: Apply suggestions from code review
EDCarman Jun 17, 2025
e119012
docs: Apply modified suggestions from code review
EDCarman Jun 17, 2025
8ed861c
docs: fix failing Vale and pre-commit check after corrections.
EDCarman Jun 17, 2025
9f8db5f
Merge branch 'main' into migration/pr-4009
EDCarman Jun 17, 2025
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
1 change: 1 addition & 0 deletions doc/changelog.d/4011.miscellaneous.md
Original file line number Diff line number Diff line change
@@ -0,0 +1 @@
Chore: (migrated pr #4009) docs: static rom training data generation example
1 change: 1 addition & 0 deletions doc/source/conf.py
Original file line number Diff line number Diff line change
Expand Up @@ -124,6 +124,7 @@
"ansys-dpf-core": ("https://dpf.docs.pyansys.com/version/stable/", None),
"ansys-math-core": ("https://math.docs.pyansys.com/version/stable/", None),
"ansys-tools-path": ("https://path.tools.docs.pyansys.com/version/stable/", None),
"pytwin": ("https://twin.docs.pyansys.com/version/stable/", None),
}

suppress_warnings = ["label.*", "design.fa-build", "config.cache"]
Expand Down
12 changes: 8 additions & 4 deletions doc/source/examples/extended_examples/index.rst
Original file line number Diff line number Diff line change
Expand Up @@ -19,13 +19,16 @@ with other programs, libraries, and features in development.
+------------------------------------------------------+--------------------------------------------------------------------------------------------+
| :ref:`dash_example` | Demonstrates the use of PyMAPDL with Dash. |
+------------------------------------------------------+--------------------------------------------------------------------------------------------+
| :ref:`executable_example` | Demonstrates how to create your own Python app with a command-line interface. |
| :ref:`executable_example` | Demonstrates how to create your own Python app with a command-line interface. |
+------------------------------------------------------+--------------------------------------------------------------------------------------------+
| :ref:`gui_example` | Demonstrates how to create your own GUI app with Python and PySide6. |
| :ref:`gui_example` | Demonstrates how to create your own GUI app with Python and PySide6. |
+------------------------------------------------------+--------------------------------------------------------------------------------------------+
| :ref:`hpc_ml_ga_example` | Demonstrates how to use PyMAPDL in a high-performance computing system managed by SLURM. |
| :ref:`hpc_ml_ga_example` | Demonstrates how to use PyMAPDL in a high-performance computing system managed by SLURM. |
+------------------------------------------------------+--------------------------------------------------------------------------------------------+
| :ref:`stochastic_fem_example` | Demonstrates using PyMAPDL for stochastic FEA using Monte Carlo simulation. |
| :ref:`stochastic_fem_example` | Demonstrates using PyMAPDL for stochastic FEA using Monte Carlo simulation. |
+------------------------------------------------------+--------------------------------------------------------------------------------------------+
| :ref:`static_rom_data_generation` | Demonstrates using PyMAPDL and PyDPF to generate training data for an Ansys Twin Builder |
| | static reduced order model (ROM). |
+------------------------------------------------------+--------------------------------------------------------------------------------------------+


Expand All @@ -42,4 +45,5 @@ with other programs, libraries, and features in development.
gui/executable.rst
hpc/hpc_ml_ga.rst
sfem/stochastic_fem.rst
static_rom/static_rom_data_generation.rst

Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
Original file line number Diff line number Diff line change
@@ -0,0 +1,302 @@
# Copyright (C) 2016 - 2025 ANSYS, Inc. and/or its affiliates.
# SPDX-License-Identifier: MIT
#
#
# Permission is hereby granted, free of charge, to any person obtaining a copy
# of this software and associated documentation files (the "Software"), to deal
# in the Software without restriction, including without limitation the rights
# to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
# copies of the Software, and to permit persons to whom the Software is
# furnished to do so, subject to the following conditions:
#
# The above copyright notice and this permission notice shall be included in all
# copies or substantial portions of the Software.
#
# THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
# IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
# FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
# AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
# LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
# OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
# SOFTWARE.

"""
Creating training data for a 3D Static reduced order model (ROM)
----------------------------------------------------------------

This example shows how to run a parametric sweep on a MAPDL model and export the output displacement
and stress data into the format required to build a static ROM with Ansys Twin Builder.
"""

import csv
import json
from pathlib import Path
import tempfile

from ansys.dpf import core as dpf
import numpy as np
from pytwin import write_binary

from ansys.mapdl.core import launch_mapdl
from ansys.mapdl.core.examples.downloads import download_example_data


def compress_id_list(id_list: np.ndarray):
"""
Compress array of consecutive IDs.

Compress array by replacing runs of three or more consecutive integers with ``start, -1, end``.

Example
-------
>>> input = np.array([0, 1, 2, 3, 4, 5, 6, 28, 29, 30, 31, 13, 15, 17, 18, 19, 20])
>>> compress_id_list(input)
[0, -1, 6, 28, -1, 31, 13, 15, 17, -1, 20]
"""
if id_list.size == 0:
return []

# Find breaks in consecutive sequences.
breaks = np.where(np.diff(id_list) != 1)[0]

# Add endpoints to form run boundaries
run_starts = np.insert(breaks + 1, 0, 0)
run_ends = np.append(breaks, len(id_list) - 1)

result = []
for start, end in zip(run_starts, run_ends):
length = end - start + 1
if length >= 3:
result.extend([int(id_list[start]), -1, int(id_list[end])])
else:
result.extend(id_list[start : end + 1].tolist())
return result


def write_settings(
path: str | Path, field: dpf.Field, name: str, is_deformation: bool = False
):
"""Write the ``settings.json`` file."""

if field.component_count in [1, 3]:
dimensionality = [field.component_count]
symmetricalDim = False
elif field.component_count == 6:
dimensionality = [3, 3]
symmetricalDim = True
else:
raise ValueError(f"Unsupported field dimensionality {field.component_count}")

settings = {
"pointsCoordinates": False,
"ids": compress_id_list(field.scoping.ids),
"location": "Nodal",
"unit": field.unit,
"unitDimension": {},
"unitFactor": 1.0,
"name": name,
"deformation": is_deformation,
"dimensionality": dimensionality,
"symmetricalDim": symmetricalDim,
"namedSelections": {},
}

with open(Path(path).joinpath("settings.json"), "w") as fw:
# Set default to convert Numpy int to int
json.dump(settings, fw, default=int, indent=4)


def get_scoping(model: dpf.Model):
"""Return scoping of unique node IDs connected to elements in model."""
op = dpf.operators.scoping.connectivity_ids(
mesh_scoping=model.metadata.meshed_region.elements.scoping,
mesh=model.metadata.meshed_region,
take_mid_nodes=True,
)
# Get output data
connected_nodes_scoping = op.outputs.mesh_scoping()
# Compress the list to only keep unique IDs
connected_nodes_scoping.ids = sorted(list(set(connected_nodes_scoping.ids)))
return connected_nodes_scoping


def write_points(model: dpf.Model, scoping: dpf.Scoping, output_folder: str | Path):
"""Write the ``points.bin`` file."""
nodes = model.metadata.meshed_region.nodes
scoped_node_indices, _ = nodes.map_scoping(scoping)
points_coordinates = nodes.coordinates_field.data[scoped_node_indices]
write_binary(Path(output_folder).joinpath("points.bin"), points_coordinates)


def write_doe_headers(output_folder: str | Path, name: str, parameters: dict):
"""Write a blank ``doe.csv`` file with headers."""
with open(Path(output_folder).joinpath("doe.csv"), "w", newline="") as fw:
writer = csv.writer(fw)
writer.writerow([name] + list(parameters.keys()))


def write_doe_entry(output_folder: str | Path, snapshot_name: str, parameters: dict):
"""Write an entry to the ``doe.csv`` file."""
with open(Path(output_folder).joinpath("doe.csv"), "a", newline="") as fw:
writer = csv.writer(fw)
writer.writerow([snapshot_name] + list(parameters.values()))


def export_static_ROM_variation(
model: dpf.Model,
scoping: dpf.Scoping,
name: str,
output_folder: str | Path,
parameters: dict,
snap_idx: int = 0,
new_metadata: bool = False,
):
"""
Export static ROM data for one parameter variation.

Parameters
----------
model : dpf.Model
DPF model with results data loaded.
scoping : dpf.Scoping
DPF nodal scoping for result export.
name : str
Result quantity to export. Options are ``displacement`` and ``stress``.
output_folder : str|Path
exported data will be stored in this folder. Use separate folders for each physics type.
parameters : dict
Dictionary of name-value pairs for the input parameters used to generate the current
results.
snap_idx : int, default = 0
Unique ID for the current results.
new_metadata : bool, default = False
Whether to trigger the creation of the following files for a given
data generation run, overwriting any existing ones: ``points.bin``,
``settings.json``, and ``doe.csv``.
"""
# Create the output folder
output_folder = Path(output_folder)
output_folder.mkdir(parents=True, exist_ok=True)

# Modify this section to export additional result types
is_deformation = False
if name == "displacement":
result = model.results.displacement
is_deformation = True
elif name == "stress":
result = model.results.stress
else:
raise ValueError(f"Unsupported result type: {name}")

# Retrieve displacement and stress at last result set
scoped_result = result.on_last_time_freq.on_mesh_scoping(scoping)

# Result must be sorted by scoping to ensure consistency across outputs
sorted_result = dpf.operators.logic.ascending_sort_fc(
scoped_result, sort_by_scoping=True
)
result_field = sorted_result.outputs.fields_container()[0]

if new_metadata:
write_points(model, scoping, output_folder)
write_doe_headers(output_folder, name, parameters)
write_settings(output_folder, result_field, name, is_deformation=is_deformation)

# Write snapshots
snapshot_folder = output_folder.joinpath("snapshots")
snapshot_folder.mkdir(parents=True, exist_ok=True)
snap_name = f"file{snap_idx}.bin"
write_doe_entry(output_folder, snap_name, parameters)
write_binary(snapshot_folder.joinpath(snap_name), result_field.data)


def export_static_ROM_data(
mapdl_results: list[tuple[str, dict]], output_folder: str | Path
):
"""
Export static ROM data to output folder.

Parameters
----------
mapdl_results: list[tuple[str, dict]]
list of tuples of MAPDL result file path and the parameter values for each variation solved.
output_folder: str|Path
Path to the folder to store ROM output data in.
"""
for idx, (rst_path, parameters) in enumerate(mapdl_results):
# Load the results to DPF and create scoping.
model = dpf.Model(rst_path)
scoping = get_scoping(model)

# Only create 'points.bin' and 'settings.json' files on first design point
new_metadata = idx == 0

# Export displacement and stress data
for name in ["displacement", "stress"]:
data_folder = Path(output_folder).joinpath(name)
export_static_ROM_variation(
model,
scoping,
name,
data_folder,
parameters=parameters,
snap_idx=idx,
new_metadata=new_metadata,
)


def run_mapdl_variations():
"""
Run the MAPDL model parametric variations.

Returns
-------
list[tuple[str, dict]]
list of tuples of MAPDL result file path (on the platform where MAPDL was executed) and the
parameter values for each variation solved.
"""
# Specify the force load variations
forces = [250, 500, 750, 1000]

# Start MAPDL and disable all but error messages
mapdl = launch_mapdl(loglevel="ERROR")

# Download the example database: ``notch_file`` is the path to the downloaded file.
notch_file = download_example_data(
filename="3d_notch.db", directory="pymapdl/static_ROM_data_generation"
)

mapdl.resume(notch_file, mute=True)

# Initialize the outputs
outputs = []

# Solve the parameter variations
for idx, force_load in enumerate(forces):
# Rename the job, change log, and error log files
mapdl.filname(f"variation_{idx}")
mapdl.run("/SOLU")
mapdl.cmsel("S", "load_node", "NODE")
mapdl.fdele("ALL", "FX")
mapdl.f("ALL", "FX", force_load)
mapdl.allsel()
mapdl.antype("STATIC")
mapdl.solve()
mapdl.finish(mute=True)
rst_path = mapdl.result_file
outputs.append((rst_path, {"force[N]": force_load}))
print(f"MAPDL run in: {mapdl.directory}")
mapdl.exit()
return outputs


def run():
# Define a folder for output
rom_folder = Path(tempfile.gettempdir()).joinpath("ansys_pymapdl_Static_ROM")
mapdl_results = run_mapdl_variations()
export_static_ROM_data(mapdl_results, rom_folder)
print(f"ROM data exported to: {rom_folder}")


if __name__ == "__main__":
run()
Loading
Loading