This system requires one or more
Detached Controller Host stations. The physics
solvers execute on the HPC (High Performance Computing) nodes by means of a
3DOrchestrate
transient station (SMAExeTranstation).
These simulations include Abaqus solver FEA simulations submitted from the physics apps—Linear Structural Validation, Structural Scenario Creation, Mechanical Scenario Creation, Fluid Scenario Creation, etc.—as well as topology (shape) optimizations and parametric optimizations
submitted from the Functional Generative Design (GDE) app. After you have configured the network infrastructure, these remote solver
jobs execute automatically through a built-in simulation process.
The diagram below shows the numbered sequence of actions that occur
through the system when an end user submits a simulation for execution on an
HPC node.
Icons are © Microsoft Corporation
|
End user submits the simulation for remote execution from the
3DEXPERIENCE
native client app. If
3DOrchestrate
run-as security is enabled, the user enters username/password OS credentials
for Windows and/or Linux.
|
|
Detached Controller Host is a Windows computer with the
NativeApps_3DEXP media installed, including
the PLM batch monitor and
3DOrchestrate
regular (public) station.
|
|
Detached Controller Host generates solver input data, updates
mesh, and checks out solver license.
|
|
Solver inputs are written to the shared folder specified in
the
fiper.file.fs.simshared property of the
3DOrchestrate
station on the detached controller host.
|
|
DRM job scheduler inputs received from open DRM custom UI in
client app are sent to the open DRM submission script (as command-line
arguments). Submission script runs on the
3DOrchestrate
server.
If run-as security is enabled, the command runs using the
end-user-provided OS credentials. On Linux, the command runs via a
setuid or
sudo SMAExePlaunch binary which validates
the credentials using the authentication mechanism configured for the operating
system.
The command defined by the open DRM server property
fiper.system.<drm-name>.transStationCommand
(executable command for
3DOrchestrate
transient station) is submitted to the job scheduler. The
3DOrchestrate
server is typically set up as a scheduler submission host or it connects to a
valid submission host.
|
|
DRM job scheduler dispatches the execution job to an HPC (High
Performance Computing) node.
|
|
When the DRM scheduler dispatches the job to an HPC node(s), a
3DOrchestrate
transient station starts (only on the primary node). The transient station is
passed a workitem ID as part of its command-line arguments. When the transient
station starts, it claims that workitem from the
3DOrchestrate
server and then uses its adapters to execute the process.
TCP socket communication occurs between the transient station
machine (HPC node) and the Detached Controller Host station machine. This
communication occurs over the ports defined by the station property
fiper.station.sim.portrange . There is also
communication to the DSLS license server. Additional HTTP and JMS traffic to
the
3DOrchestrate
server occurs as well. This status communication will typically be less than 2
MB for the duration of the complete simulation.
|
|
(Optional) You can use a custom file transfer script for your
open DRM system; see
Custom File Transfer Script.
In this case, the
fiper.station.sim.stageFileScript pulls input
from
fiper.file.fs.simshared to the HPC node
transient station tempdir (SMP) or the local
simshared folder (DMP). This is optional if
HPC can directly access the on-premises
simshared folder.
|
|
The HPC nodes must have the
SIMULIA_ComputeServices media installed
locally or available from a network shared directory. This media provides the
Abaqus
solvers and other services:
Abaqus/Standard,
Abaqus/Explicit,
Flow (Fluid/CFD), optimization, etc. The HPC nodes must also have a
3DOrchestrate
public (regular) station installed locally or available from a network shared
directory—this allows the HPC node to start a transient station (transtation)
when it is needed.
|
|
(Optional) Custom file transfer script pushes results back to
the station
fiper.file.fs.simshared .
|
Configuring a Detached Controller Host Station
One or more Detached Controller Hosts must be configured to achieve
the complete physics execution system described above.
A Detached Controller Host must have the
NativeApps_3DEXP
media installed—this includes the
native (rich client) simulation apps as well as a public (regular)
3DOrchestrate
station. In the station properties file, you must specify the location of the
native apps installation in the following property:
fiper.station.sim.clientservices=
You must also set the following station property to open TCP/IP ports
for communication between the executables running on different machines.
fiper.station.sim.portrange=
You can specify a port number range such as 56000-57000, a
comma-separated list of individual ports numbers such as 56010,56012,56050, or
a combination of both. If this property is not set, the simulation executables
will use any available open ports.
In the complete physics execution system, the Detached Controller Host
performs the following operations: generate solver input data, update mesh, and
check out solver license. The Detached Controller Host does not run the physics
simulation solvers—that happens on the DRM-HPC compute nodes.
The Detached Controller Host must make use of a shared disk/file
system on your local area network. This folder/directory must be specified in
the
3DOrchestrate
station property
fiper.file.fs.simshared
:
fiper.file.fs.simshared=$SHARED_DRIVE
You must either define the environment variable SHARED_DRIVE on the
Detached Controller Host or set the property to an absolute path; for example,
the Windows UNC path
\\server1\sharedsim\.
The
simshared
directory exists as a waypoint between the
Windows Detached Controller Host and the remote HPC compute node (either Linux
or Windows). When executing a physics simulation or topology (shape)
optimization (from the
Functional Generative Design
app), the Detached Controller Host puts the solver input files in the
simshared
directory where they are visible to the HPC
compute node.
Sizing Considerations for a Detached Controller Host
The memory requirements for your Detached Controller Host station(s)
depend on the number of concurrent jobs that will be executing. Consider the
following issues when planning the computer that will be a Detached Controller
Host station for
3DOrchestrate: