class F06Writer(OP2_F06_Common): def __init__(self): OP2_F06_Common.__init__(self) self.card_count = {} self.additional_matrices = {} self.subcase_key = defaultdict(list) self.end_options = {} self._results = ResultSet(self.get_all_results(), self.op2_results.get_sum_objects_map(), self.log) def get_all_results(self) -> List[str]: all_results = [ 'stress', 'strain', 'element_forces', 'constraint_forces', 'thermal_load' ] + self.get_table_types() return all_results def clear_results(self) -> None: self._results.clear() def add_results(self, results: Union[str, List[str]]) -> None: if isinstance(results, str): results = [results] all_results = self.get_all_results() for result in results: result = str(result) if result not in all_results: raise RuntimeError( '%r is not a valid result to remove; all_results=%s' % (result, all_results)) if result == 'stress': stress_results = [] for result in all_results: if 'stress' in result.lower(): stress_results.append(result) #stress_results = [result if 'stress' in result.lower() for result in all_results] self._results.update(stress_results) elif result == 'strain': strain_results = [] for result in all_results: if 'strain' in result.lower(): strain_results.append(result) #strain_results = [result if 'strain' in result.lower() for result in all_results] self._results.update(strain_results) elif 'stress' in result.lower(): self._results.add('stress') elif 'strain' in result.lower(): self._results.add('strain') elif result in ('spc_forces', 'mpc_forces', 'constraint_forces'): self._results.add('constraint_forces') elif 'force' in result.lower(): # could use more validation... self._results.add('element_forces') # thermalLoad_VU_3D, thermalLoad_1D, conv_thermal_load, thermalLoad_2D_3D self._results.add(result) def set_results(self, results: Union[str, List[str]]) -> None: self.clear_results() self.add_results(results) def remove_results(self, results: Union[str, List[str]]) -> None: self._results.remove(results) def make_f06_header(self): """If this class is inherited, the F06 Header may be overwritten""" return make_f06_header() def make_stamp(self, title, today, build=None): """If this class is inherited, the PAGE stamp may be overwritten""" return make_stamp(title, today, build=None) def make_grid_point_singularity_table(self, failed): """ creates a grid point singularity table Parameters ---------- failed : List[(nid, component), ...] defines failed degrees of freedom """ msg = '' if failed: msg += '0 G R I D P O I N T S I N G U L A R I T Y T A B L E\n' msg += '0 POINT TYPE FAILED STIFFNESS OLD USET NEW USET\n' msg += ' ID DIRECTION RATIO EXCLUSIVE UNION EXCLUSIVE UNION\n' for (nid, dof) in failed: msg += ' %8s G %s 0.00E+00 B F SB SB *\n' % ( nid, dof) else: msg += 'No constraints have been applied...\n' page_stamp = self.make_stamp(self.title, self.date) msg += page_stamp % self.page_num self.page_num += 1 return msg def _write_summary(self, f06_file, card_count=None): """writes the F06 card summary table""" summary_header = ' M O D E L S U M M A R Y\n\n' summary = '' self.cards_to_read = { # rigid elements 'RBAR', 'RBAR1', 'RBE1', 'RBE2', 'RBE3', # spc/mpc constraints 'SPC', 'SPCADD', 'SPC1', 'SPCD', 'SPCAX', 'MPC', 'MPCADD', 'SUPORT', 'SUPORT1', # aero cards 'CAERO1', 'CAERO2', 'CAERO3', 'CAERO4', 'CAERO5', # temperature cards 'CHBDYE', 'CHBDYG', 'CHBDYP', 'CONV', } blocks = [ ['POINTS', [ 'GRID', 'GRDSET', ]], ['ENTRIES', ['SPOINT']], [ 'ELEMENTS', [ # these are sorted # elements 'CONM1', 'CONM2', 'CMASS1', 'CMASS2', 'CMASS3', 'CMASS4', # springs 'CELAS1', 'CELAS2', 'CELAS3', 'CELAS4', 'CELAS5', # bushings 'CBUSH', 'CBUSH1D', 'CBUSH2D', # dampers 'CDAMP1', 'CDAMP2', 'CDAMP3', 'CDAMP4', 'CDAMP5', # bar flags 'BAROR', 'CBARAO', # bars 'CBAR', 'CROD', 'CTUBE', 'BEAMOR', 'CBEAM', 'CBEAM3', 'CONROD', 'CBEND', # shells 'CTRIA3', 'CTRIA6', 'CTRIAR', 'CTRIAX', 'CTRIAX6', 'CQUAD4', 'CQUAD8', 'CQUADR', 'CQUADX', 'CQUAD', # solids 'CTETRA', 'CPENTA', 'CHEXA', # other 'CSHEAR', 'CVISC', 'CRAC2D', 'CRAC3D', 'CGAP', 'CFAST', 'RBE2', 'RBE3', # thermal 'CHBDYP', 'CHBDYG', 'CONV', ] ], ] #print("self.card_count", self.card_count) if card_count is None: card_count = self.card_count for block in blocks: block_name, keys = block key_count = 0 for key in sorted(keys): try: value = card_count[key] summary += ' NUMBER OF %-8s %-8s = %8s\n' % ( key, block_name, value) key_count += 1 except KeyError: pass if key_count: summary += ' \n' if summary: f06_file.write(summary_header) f06_file.write(summary) page_stamp = self.make_stamp(self.title, self.date) f06_file.write(page_stamp % self.page_num) self.page_num += 1 def write_f06(self, f06_outname: str, matrix_filename: Optional[str] = None, is_mag_phase: bool = False, is_sort1: bool = True, delete_objects: bool = True, end_flag: bool = False, quiet: bool = True, repr_check: bool = False, close: bool = True) -> None: """ Writes an F06 file based on the data we have stored in the object Parameters ---------- f06_outname : str the name of the F06 file to write matrix_filename : str; default=None str : the name of the .mat file to write None : based on f06_outname is_mag_phase : bool; default=False should complex data be written using Magnitude/Phase instead of Real/Imaginary Real objects don't use this parameter is_sort1 : bool; default=True writes output in SORT1 format if the output is transient; ignored for static analyses delete_objects : bool; default=True should objects be deleted after they're written to reduce memory end_flag : bool; default=False should a dummy Nastran "END" table be made quiet : bool; default=False suppress print messages repr_check: bool; defualt=False calls the object repr as a validation test (prints nothing) close : bool; default=True close the f06 file """ if not quiet: print("F06:") if isinstance(f06_outname, str): if matrix_filename is None: matrix_filename = os.path.splitext(f06_outname)[0] + '.mat' #print("matrix_filename =", matrix_filename) #mat = open(matrix_filename, 'wb') f06 = open(f06_outname, 'w') self._write_summary(f06) elif hasattr(f06_outname, 'read') and hasattr(f06_outname, 'write'): #f06 = f06_outname #else: #print('type(f06_outname) =', type(f06_outname)) #assert isinstance(f06_outname, file), 'type(f06_outname)= %s' % f06_outname f06 = f06_outname f06_outname = f06.name if matrix_filename is None: matrix_filename = os.path.splitext(f06_outname)[0] + '.mat' if not quiet: print('f06_outname =', f06_outname) page_stamp = self.make_stamp(self.title, self.date) if self.grid_point_weight: if not quiet: print(" grid_point_weight") for key, weight in self.grid_point_weight.items(): self.page_num = weight.write_f06(f06, page_stamp, self.page_num) if repr_check: str(self.grid_point_weight) assert isinstance(self.page_num, int), self.grid_point_weight.__class__.__name__ if self.oload_resultant is not None: self.page_num = self.oload_resultant.write_f06( f06, page_stamp, self.page_num) if repr_check: str(self.oload_resultant) assert isinstance(self.page_num, int), self.oload_resultant.__class__.__name__ # writes all results for self._write_f06_subcase_based(f06, page_stamp, delete_objects=delete_objects, is_mag_phase=is_mag_phase, is_sort1=is_sort1, quiet=quiet, repr_check=repr_check) self.op2_results.psds.write_f06(f06) #self._write_f06_time_based(f06, page_stamp) self.write_matrices(f06, matrix_filename, page_stamp, self.page_num, quiet=quiet) f06.write(make_end(end_flag, self.end_options)) if close: f06.close() def write_matrices(self, f06, matrix_filename, page_stamp, page_num, quiet=True): """writes the f06 matrices""" if len(self.matrices): if hasattr(self, 'monitor1'): page_num = self.monitor1.write(f06, page_stamp=page_stamp, page_num=page_num) self.log.debug('MONPNT1 from [PMRF, PERF, PFRF, AGRF]') with open(matrix_filename, 'wb') as mat: for name, matrix in self.matrices.items(): if name == 'MP3F': page_num = self.monitor3.write(f06, page_stamp=page_stamp, page_num=page_num) self.log.debug('MONPNT3 from MP3F') elif name in ['PMRF', 'PERF', 'PFRF', 'AGRF']: pass else: if not quiet: print(matrix) matrix.write(mat) def _write_f06_subcase_based(self, f06, page_stamp, delete_objects=True, is_mag_phase=False, is_sort1=True, quiet=False, repr_check=False): """ Helper function for ``write_f06`` that does the real work Parameters ---------- f06 : file the opened file object page_stamp : str the format string stamp is the ending to every F06 page that contains the version, date, and page number (e.g., 'pyNastran 0.8 1/1/2016 PAGE %%i') is_mag_phase : bool; default=False should complex data be written using Magnitude/Phase instead of Real/Imaginary Real objects don't use this parameter is_sort1 : bool; default=True writes output in SORT1 format if the output is transient; ignored for static analyses delete_objects : bool; default=True should objects be deleted after they're written to reduce memory quiet : bool; default=False suppress print messages repr_check: bool; defualt=False calls the object repr as a validation test (prints nothing) """ is_failed = False header = [ ' DEFAULT \n', '\n', '' ] # eigenvalues are written first f06.write(page_stamp % self.page_num) self.page_num += 1 for ikey, result in sorted(self.eigenvalues.items()): if not quiet: print('%-18s case=%r' % (result.__class__.__name__, ikey)) self.page_num = result.write_f06(f06, header, page_stamp, page_num=self.page_num) if repr_check: str(result) assert isinstance(self.page_num, int), 'pageNum=%r' % str(self.page_num) if delete_objects: del result self.page_num += 1 # then eigenvectors # has a special header # isubcases = sorted(self.isubcase_name_map.keys()) # TODO: superelement version...need the nominal... res_keys_subcase = self.subcase_key if len(res_keys_subcase) == 0: self.log.warning('no cases to write...self.subcase_key=%r' % self.subcase_key) return for isubcase, res_keys in sorted(res_keys_subcase.items()): for res_key in res_keys: if isinstance(res_key, tuple): pass #is_compressed = False else: # int #is_compressed = True isubcase = res_key if res_key not in self.eigenvectors: continue result = self.eigenvectors[res_key] if repr_check: str(result) subtitle = result.subtitle header[0] = ' %s\n' % subtitle header[ 1] = '0 SUBCASE %i\n' % isubcase #header[2] = complex/nonlinear res_length = 18 res_format = '*%%-%is SUBCASE=%%i' % res_length res_format_vectorized = ' %%-%is SUBCASE=%%i SUBTITLE=%%s' % res_length class_name = result.__class__.__name__ if hasattr(result, 'data'): if not quiet: print(res_format_vectorized % (class_name, isubcase, subtitle)) else: print(res_format % (class_name, isubcase)) self.page_num = result.write_f06(f06, header, page_stamp, self.page_num, is_mag_phase=is_mag_phase, is_sort1=True) assert isinstance(self.page_num, int), 'pageNum=%r' % str(self.page_num) if delete_objects: del result self.page_num += 1 # finally, we writte all the other tables # nastran puts the tables in order of the Case Control deck, # but we're lazy so we just hardcode the order # subcase name, subcase ID, transient word & value header_old = [ ' DEFAULT \n', '\n', ' \n' ] header = copy.deepcopy(header_old) unallowed_results = [ 'eigenvectors', 'eigenvalues', 'params', 'gpdt', 'bgpdt', 'eqexin', 'grid_point_weight', 'psds' ] res_types = list( self.get_result(table_type) for table_type in sorted(self.get_table_types()) if table_type not in unallowed_results and not table_type.startswith('responses.')) for isubcase, res_keys in sorted(res_keys_subcase.items()): for res_key in res_keys: #if isinstance(res_key, tuple): #is_compressed = False #else: #is_compressed = True res_length = self._get_result_length(res_types, res_key) if res_length == 0: # skipped subcase; no saved results continue res_format = '*%%-%is SUBCASE=%%i%%s' % res_length res_format_vectorized = ' %%-%is SUBCASE=%%i SUBTITLE=%%s %%s' % res_length for res_type in res_types: if res_key not in res_type: continue result = res_type[res_key] if repr_check: str(result) subtitle = result.subtitle label = result.label header = ['', ''] header[0] = ' %-126s\n' % subtitle header[ 1] = '0 %-32s SUBCASE %-15i\n \n' % ( label, isubcase) if result.nonlinear_factor is not None: header.append('') try: class_name = result.__class__.__name__ element_name = '' if hasattr(result, 'element_name'): element_name = ' - ' + result.element_name is_ignored = 'StrainEnergy' not in class_name and 'GridPointForces' not in class_name has_nnodes = not hasattr(result, 'nnodes_per_element') if has_nnodes and is_ignored and getpass.getuser( ) == 'sdoyle': self.log.error( f'{class_name} is missing nnodes_per_element' ) if hasattr(result, 'data'): if not quiet: print(res_format_vectorized % (class_name, isubcase, subtitle, element_name)) else: print(res_format % (class_name, isubcase, element_name)) result.is_complex result.is_real try: self.page_num = result.write_f06( f06, header, page_stamp, page_num=self.page_num, is_mag_phase=is_mag_phase, is_sort1=is_sort1) except Exception as error: print_exc(file=sys.stdout) print(''.join(result.get_stats())) raise #assert 'table_name=' in ''.join(result.get_stats()) assert isinstance( self.page_num, int), 'result=%s pageNum=%r' % (result, str(self.page_num)) except: #print("result name = %r" % result.name()) raise if delete_objects: del result self.page_num += 1
class F06Writer(OP2_F06_Common): def __init__(self): OP2_F06_Common.__init__(self) self.card_count = {} self.additional_matrices = {} self.matrices = {} self.subcase_key = defaultdict(list) self.end_options = {} self._results = ResultSet(self.get_all_results()) def get_all_results(self): all_results = [ 'stress', 'strain', 'element_forces', 'constraint_forces' ] + self.get_table_types() return all_results def clear_results(self): self._results.clear() def add_results(self, results): if isinstance(results, string_types): results = [results] all_results = self.get_all_results() for result in results: result = str(result) if result not in all_results: raise RuntimeError( '%r is not a valid result to remove; all_results=%s' % (result, all_results)) if result == 'stress': stress_results = [] for result in all_results: if 'stress' in result.lower(): stress_results.append(result) #stress_results = [result if 'stress' in result.lower() for result in all_results] self._results.update(stress_results) elif result == 'strain': strain_results = [] for result in all_results: if 'strain' in result.lower(): strain_results.append(result) #strain_results = [result if 'strain' in result.lower() for result in all_results] self._results.update(strain_results) elif 'stress' in result.lower(): self._results.add('stress') elif 'strain' in result.lower(): self._results.add('strain') elif result in ('spc_forces', 'mpc_forces', 'constraint_forces'): self._results.add('constraint_forces') elif 'force' in result.lower(): # could use more validation... self._results.add('element_forces') # thermalLoad_VU_3D, thermalLoad_1D, conv_thermal_load, thermalLoad_2D_3D self._results.add(result) def set_results(self, results): if isinstance(results, string_types): results = [results] self.clear_results() self.add_results(results) def remove_results(self, results): self._results.remove(results) def make_f06_header(self): """If this class is inherited, the F06 Header may be overwritten""" return make_f06_header() def make_stamp(self, title, today): """If this class is inherited, the PAGE stamp may be overwritten""" return make_stamp(title, today) def make_grid_point_singularity_table(self, failed): """ creates a grid point singularity table Parameters ---------- failed : List[(nid, component), ...] defines failed degrees of freedom """ msg = '' if failed: msg += '0 G R I D P O I N T S I N G U L A R I T Y T A B L E\n' msg += '0 POINT TYPE FAILED STIFFNESS OLD USET NEW USET\n' msg += ' ID DIRECTION RATIO EXCLUSIVE UNION EXCLUSIVE UNION\n' for (nid, dof) in failed: msg += ' %8s G %s 0.00E+00 B F SB SB *\n' % ( nid, dof) else: msg += 'No constraints have been applied...\n' page_stamp = self.make_stamp(self.title, self.date) msg += page_stamp % self.page_num self.page_num += 1 return msg def _write_summary(self, f06, card_count=None): """writes the F06 card summary table""" summary_header = ' M O D E L S U M M A R Y\n\n' summary = '' self.cards_to_read = set([ # rigid elements 'RBAR', 'RBAR1', 'RBE1', 'RBE2', 'RBE3', # spc/mpc constraints 'SPC', 'SPCADD', 'SPC1', 'SPCD', 'SPCAX', 'MPC', 'MPCADD', 'SUPORT', 'SUPORT1', # aero cards 'CAERO1', 'CAERO2', 'CAERO3', 'CAERO4', 'CAERO5', # temperature cards 'CHBDYE', 'CHBDYG', 'CHBDYP', 'CONV', ]) blocks = [ ['POINTS', [ 'GRID', 'GRDSET', ]], ['ENTRIES', ['SPOINT']], [ 'ELEMENTS', [ # these are sorted # elements 'CONM1', 'CONM2', 'CMASS1', 'CMASS2', 'CMASS3', 'CMASS4', # springs 'CELAS1', 'CELAS2', 'CELAS3', 'CELAS4', 'CELAS5', # bushings 'CBUSH', 'CBUSH1D', 'CBUSH2D', # dampers 'CDAMP1', 'CDAMP2', 'CDAMP3', 'CDAMP4', 'CDAMP5', # bar flags 'BAROR', 'CBARAO', # bars 'CBAR', 'CROD', 'CTUBE', 'BEAMOR', 'CBEAM', 'CBEAM3', 'CONROD', 'CBEND', # shells 'CTRIA3', 'CTRIA6', 'CTRIAR', 'CTRIAX', 'CTRIAX6', 'CQUAD4', 'CQUAD8', 'CQUADR', 'CQUADX', 'CQUAD', # solids 'CTETRA', 'CPENTA', 'CHEXA', # other 'CSHEAR', 'CVISC', 'CRAC2D', 'CRAC3D', 'CGAP', 'CFAST', 'RBE2', 'RBE3', # thermal 'CHBDYP', 'CHBDYG', 'CONV', ] ], ] #print("self.card_count", self.card_count) if card_count is None: card_count = self.card_count for block in blocks: block_name, keys = block key_count = 0 for key in sorted(keys): try: value = card_count[key] summary += ' NUMBER OF %-8s %-8s = %8s\n' % ( key, block_name, value) key_count += 1 except KeyError: pass if key_count: summary += ' \n' if summary: f06.write(summary_header) f06.write(summary) page_stamp = self.make_stamp(self.title, self.date) f06.write(page_stamp % self.page_num) self.page_num += 1 def write_f06(self, f06_outname, matrix_filename=None, is_mag_phase=False, is_sort1=True, delete_objects=True, end_flag=False, quiet=True, repr_check=False, close=True): """ Writes an F06 file based on the data we have stored in the object Parameters ---------- f06_outname : str the name of the F06 file to write matrix_filename : str; default=None str : the name of the .mat file to write None : based on f06_outname is_mag_phase : bool; default=False should complex data be written using Magnitude/Phase instead of Real/Imaginary Real objects don't use this parameter is_sort1 : bool; default=True writes output in SORT1 format if the output is transient; ignored for static analyses delete_objects : bool; default=True should objects be deleted after they're written to reduce memory end_flag : bool; default=False should a dummy Nastran "END" table be made quiet : bool; default=False suppress print messages repr_check: bool; defualt=False calls the object repr as a validation test (prints nothing) close : bool; default=True close the f06 file """ if not quiet: print("F06:") if isinstance(f06_outname, str): if matrix_filename is None: matrix_filename = os.path.splitext(f06_outname)[0] + '.mat' #print("matrix_filename =", matrix_filename) #mat = open(matrix_filename, 'wb') if PY2: f06 = open(f06_outname, 'wb') else: f06 = open(f06_outname, 'w') self._write_summary(f06) elif hasattr(f06_outname, 'read') and hasattr(f06_outname, 'write'): #f06 = f06_outname #else: #print('type(f06_outname) =', type(f06_outname)) #assert isinstance(f06_outname, file), 'type(f06_outname)= %s' % f06_outname f06 = f06_outname f06_outname = f06.name if matrix_filename is None: matrix_filename = os.path.splitext(f06_outname)[0] + '.mat' if not quiet: print('f06_outname =', f06_outname) page_stamp = self.make_stamp(self.title, self.date) if self.grid_point_weight.reference_point is not None: if not quiet: print(" grid_point_weight") self.page_num = self.grid_point_weight.write_f06( f06, page_stamp, self.page_num) if repr_check: str(self.grid_point_weight) assert isinstance(self.page_num, int), self.grid_point_weight.__class__.__name__ if self.oload_resultant is not None: self.page_num = self.oload_resultant.write_f06( f06, page_stamp, self.page_num) if repr_check: str(self.oload_resultant) assert isinstance(self.page_num, int), self.oload_resultant.__class__.__name__ # writes all results for self._write_f06_subcase_based(f06, page_stamp, delete_objects=delete_objects, is_mag_phase=is_mag_phase, is_sort1=is_sort1, quiet=quiet, repr_check=repr_check) #self._write_f06_time_based(f06, page_stamp) self.write_matrices(f06, matrix_filename, page_stamp, self.page_num, quiet=quiet) f06.write(make_end(end_flag, self.end_options)) if close: f06.close() def write_matrices(self, f06, matrix_filename, page_stamp, page_num, quiet=True): """writes the f06 matrices""" if len(self.matrices): if hasattr(self, 'monitor1'): page_num = self.monitor1.write(f06, page_stamp=page_stamp, page_num=page_num) print('MONPNT1 from [PMRF, PERF, PFRF, AGRF]') with open(matrix_filename, 'wb') as mat: for name, matrix in iteritems(self.matrices): if name == 'MP3F': page_num = self.monitor3.write(f06, page_stamp=page_stamp, page_num=page_num) print('MONPNT3 from MP3F') elif name in ['PMRF', 'PERF', 'PFRF', 'AGRF']: pass else: if not quiet: print(matrix) matrix.write(mat) def _write_f06_subcase_based(self, f06, page_stamp, delete_objects=True, is_mag_phase=False, is_sort1=True, quiet=False, repr_check=False): """ Helper function for ``write_f06`` that does the real work Parameters ---------- f06 : file the opened file object page_stamp : str the format string stamp is the ending to every F06 page that contains the version, date, and page number (e.g., 'pyNastran 0.8 1/1/2016 PAGE %%i') is_mag_phase : bool; default=False should complex data be written using Magnitude/Phase instead of Real/Imaginary Real objects don't use this parameter is_sort1 : bool; default=True writes output in SORT1 format if the output is transient; ignored for static analyses delete_objects : bool; default=True should objects be deleted after they're written to reduce memory quiet : bool; default=False suppress print messages repr_check: bool; defualt=False calls the object repr as a validation test (prints nothing) """ is_failed = False header = [ ' DEFAULT \n', '\n', '' ] # eigenvalues are written first f06.write(page_stamp % self.page_num) self.page_num += 1 for ikey, result in sorted(iteritems(self.eigenvalues)): if not quiet: print('%-18s case=%r' % (result.__class__.__name__, ikey)) self.page_num = result.write_f06(f06, header, page_stamp, page_num=self.page_num) if repr_check: str(result) assert isinstance(self.page_num, int), 'pageNum=%r' % str(self.page_num) if delete_objects: del result self.page_num += 1 # then eigenvectors # has a special header # isubcases = sorted(self.isubcase_name_map.keys()) # TODO: superelement version...need the nominal... res_keys_subcase = self.subcase_key if len(res_keys_subcase) == 0: self.log.warning('no cases to write...self.subcase_key=%r' % self.subcase_key) return for isubcase, res_keys in sorted(iteritems(res_keys_subcase)): for res_key in res_keys: if isinstance(res_key, tuple): is_compressed = False else: # int is_compressed = True isubcase = res_key if res_key not in self.eigenvectors: continue result = self.eigenvectors[res_key] if repr_check: str(result) subtitle = result.subtitle header[0] = ' %s\n' % subtitle header[ 1] = '0 SUBCASE %i\n' % isubcase #header[2] = complex/nonlinear res_length = 18 res_format = '*%%-%is SUBCASE=%%i' % res_length res_format_vectorized = ' %%-%is SUBCASE=%%i SUBTITLE=%%s' % res_length class_name = result.__class__.__name__ if hasattr(result, 'data'): if not quiet: print(res_format_vectorized % (class_name, isubcase, subtitle)) else: print(res_format % (class_name, isubcase)) self.page_num = result.write_f06(f06, header, page_stamp, self.page_num, is_mag_phase=is_mag_phase, is_sort1=True) assert isinstance(self.page_num, int), 'pageNum=%r' % str(self.page_num) if delete_objects: del result self.page_num += 1 # finally, we writte all the other tables # nastran puts the tables in order of the Case Control deck, # but we're lazy so we just hardcode the order # subcase name, subcase ID, transient word & value header_old = [ ' DEFAULT \n', '\n', ' \n' ] header = copy.deepcopy(header_old) unallowed_results = ['eigenvectors', 'eigenvalues'] res_types = list( self.get_result(table_type) for table_type in sorted(self.get_table_types()) if table_type not in unallowed_results) #ROUGV1 = self.op2_results.ROUGV1 #psd = self.op2_results.psd #ato = self.op2_results.ato #rms = self.op2_results.rms #crm = self.op2_results.crm #no = self.op2_results.no #modal_contribution = self.op2_results.modal_contribution #res_types = [ #self.displacements, ROUGV1.displacements, psd.displacements, ato.displacements, rms.displacements, crm.displacements, no.displacements, self.displacements_scaled, # ??? #self.accelerations, ROUGV1.accelerations, psd.accelerations, ato.accelerations, rms.accelerations, crm.accelerations, no.accelerations, #self.velocities, ROUGV1.velocities, psd.velocities, ato.velocities, rms.velocities, crm.velocities, no.velocities, #self.force_vectors, #self.load_vectors, ato.load_vectors, crm.load_vectors, psd.load_vectors, no.load_vectors, #self.temperatures, #self.eigenvectors, #self.eigenvectors_RADCONS, #self.eigenvectors_RADEFFM, #self.eigenvectors_RADEATC, #self.eigenvectors_ROUGV1, #self.mpc_forces, psd.mpc_forces, ato.mpc_forces, rms.mpc_forces, ##self.mpc_forces_RAQCONS, ##self.mpc_forces_RAQEATC, #self.spc_forces, psd.spc_forces, ato.spc_forces, rms.spc_forces, #self.thermal_load_vectors, ##self.strain_energy, #self.cquad4_strain_energy, self.cquad8_strain_energy, #self.cquadr_strain_energy, self.cquadx_strain_energy, #self.ctria3_strain_energy, self.ctria6_strain_energy, #self.ctriar_strain_energy, self.ctriax_strain_energy, #self.ctriax6_strain_energy, #self.ctetra_strain_energy, self.cpenta_strain_energy, #self.chexa_strain_energy, self.cpyram_strain_energy, #self.crod_strain_energy, self.ctube_strain_energy, #self.conrod_strain_energy, #self.cbar_strain_energy, self.cbeam_strain_energy, #self.cgap_strain_energy, self.cbush_strain_energy, #self.celas1_strain_energy, self.celas2_strain_energy, #self.celas3_strain_energy, self.celas4_strain_energy, #self.cdum8_strain_energy, self.dmig_strain_energy, #self.cbend_strain_energy, self.conm2_strain_energy, #self.genel_strain_energy, self.cshear_strain_energy, ##------------------------------------------ ## OEF - forces ## alphabetical order... ## bars #self.cbar_force, ato.cbar_force, crm.cbar_force, psd.cbar_force, rms.cbar_force, no.cbar_force, #self.cbar_force_10nodes, ## beam #self.cbend_force, #self.cbeam_force, ato.cbeam_force, crm.cbeam_force, psd.cbeam_force, rms.cbeam_force, no.cbeam_force, #self.cbeam_force_vu, ## alphabetical #self.celas1_force, #self.celas2_force, #self.celas3_force, #self.celas4_force, #self.cquad4_force, ato.cquad4_force, crm.cquad4_force, psd.cquad4_force, rms.cquad4_force, no.cquad4_force, #self.cquad8_force, ato.cquad8_force, crm.cquad8_force, psd.cquad8_force, rms.cquad8_force, no.cquad8_force, #self.cquadr_force, ato.cquadr_force, crm.cquadr_force, psd.cquadr_force, rms.cquadr_force, no.cquadr_force, #self.conrod_force, #self.crod_force, #self.cshear_force, #self.ctria3_force, ato.ctria3_force, crm.ctria3_force, psd.ctria3_force, rms.ctria3_force, no.ctria3_force, #self.ctria6_force, ato.ctria3_force, crm.ctria6_force, psd.ctria6_force, rms.ctria6_force, no.ctria6_force, #self.ctriar_force, ato.ctriar_force, crm.ctriar_force, psd.ctriar_force, rms.ctriar_force, no.ctriar_force, #self.ctube_force, ## springs #self.celas1_force, #self.celas2_force, #self.celas3_force, #self.celas4_force, ## dampers #self.cdamp1_force, #self.cdamp2_force, #self.cdamp3_force, #self.cdamp4_force, ## other #self.cbush_force, ato.cbush_force, crm.cbush_force, psd.cbush_force, rms.cbush_force, no.cbush_force, #self.cgap_force, #self.cvisc_force, #self.chexa_pressure_force, #self.cpenta_pressure_force, #self.ctetra_pressure_force, #self.coneax_force, ##------------------------------------------ ## OES - strain ## 1. cbar ## 2. cbeam ## 3. crod/ctube/conrod ## springs #self.celas1_strain, #self.celas2_strain, #self.celas3_strain, #self.celas4_strain, #self.nonlinear_celas1_stress, #self.nonlinear_celas3_stress, ## bars/beams #self.cbar_strain, ato.cbar_strain, crm.cbar_strain, psd.cbar_strain, rms.cbar_strain, no.cbar_strain, #self.cbar_strain_10nodes, #self.cbeam_strain, ## plates #self.cquad4_composite_strain, #self.cquad8_composite_strain, #self.cquadr_composite_strain, #self.ctria3_composite_strain, #self.ctria6_composite_strain, #self.ctriar_composite_strain, #self.nonlinear_ctria3_strain, #self.nonlinear_cquad4_strain, #self.ctriax_strain, ## rods #self.nonlinear_crod_strain, #self.nonlinear_ctube_strain, #self.nonlinear_conrod_strain, #self.chexa_strain, modal_contribution.chexa_strain, #self.conrod_strain, modal_contribution.conrod_strain, #self.cpenta_strain, modal_contribution.cpenta_strain, #self.cquad4_strain, modal_contribution.cquad4_strain, #self.cquad8_strain, modal_contribution.cquad8_strain, #self.cquadr_strain, modal_contribution.cquadr_strain, #self.crod_strain, modal_contribution_crod_strain, #self.cshear_strain, modal_contribution.cshear_strain, #self.ctetra_strain, modal_contribution.ctetra_strain, #self.ctria3_strain, modal_contribution.ctria3_strain, #self.ctria6_strain, modal_contribution.ctria6_strain, #self.ctriar_strain, modal_contribution.ctriar_strain, #self.ctube_strain, modal_contribution.ctube_strain, ## bush #self.cbush_strain, #self.nonlinear_cbush_stress, #self.cbush1d_stress_strain, ##------------------------------------------ ## cbars/cbeams #self.cbar_stress, ato.cbar_stress, crm.cbar_stress, psd.cbar_stress, rms.cbar_stress, no.cbar_stress, #self.cbar_stress_10nodes, #self.nonlinear_cbeam_stress, #self.cbeam_stress, ## bush #self.cbush_stress, ## rods #self.nonlinear_crod_stress, #self.nonlinear_ctube_stress, #self.nonlinear_conrod_stress, ## shear ## OES - stress #self.celas1_stress, modal_contribution.celas1_stress, #self.celas2_stress, modal_contribution.celas2_stress, #self.celas3_stress, modal_contribution.celas3_stress, #self.celas4_stress, modal_contribution.celas4_stress, #self.chexa_stress, #self.conrod_stress, #self.cpenta_stress, #self.cquad4_stress, #self.cquad8_stress, #self.cquadr_stress, #self.crod_stress, #self.cshear_stress, #self.ctetra_stress, #self.ctria3_stress, #self.ctria6_stress, #self.ctriar_stress, #self.ctube_stress, #self.cquad4_composite_stress, #self.cquad8_composite_stress, #self.cquadr_composite_stress, #self.ctria3_composite_stress, #self.ctria6_composite_stress, #self.ctriar_composite_stress, #self.nonlinear_ctria3_stress, #self.nonlinear_cquad4_stress, #self.ctriax_stress, #self.hyperelastic_cquad4_strain, ##------------------------------------------ ##OEF - Fluxes - tCode=4 thermal=1 #self.conv_thermal_load, ##self.thermalLoad_CHBDY, #self.chbdye_thermal_load, #self.chbdyg_thermal_load, #self.chbdyp_thermal_load, #self.chbdye_thermal_load_flux, #self.chbdyg_thermal_load_flux, #self.chbdyp_thermal_load_flux, ##self.thermalLoad_1D, #self.crod_thermal_load, #self.cbeam_thermal_load, #self.ctube_thermal_load, #self.conrod_thermal_load, #self.cbar_thermal_load, #self.cbend_thermal_load, #self.crod_thermal_load_flux, #self.cbeam_thermal_load_flux, #self.ctube_thermal_load_flux, #self.conrod_thermal_load_flux, #self.cbar_thermal_load_flux, #self.cbend_thermal_load_flux, ##self.thermalLoad_2D_3D, #self.cquad4_thermal_load, #self.ctriax6_thermal_load, #self.cquad8_thermal_load, #self.ctria3_thermal_load, #self.ctria6_thermal_load, #self.ctetra_thermal_load, #self.chexa_thermal_load, #self.cpenta_thermal_load, #self.cquad4_thermal_load_flux, #self.ctriax6_thermal_load_flux, #self.cquad8_thermal_load_flux, #self.ctria3_thermal_load_flux, #self.ctria6_thermal_load_flux, #self.ctetra_thermal_load_flux, #self.chexa_thermal_load_flux, #self.cpenta_thermal_load_flux, #self.thermalLoad_VU, #self.thermalLoad_VU_3D, #self.vu_beam_thermal_load, #self.vu_quad_force, #self.vu_tria_force, ##------------------------------------------ #self.grid_point_stresses, self.grid_point_volume_stresses, self.grid_point_forces, #] for isubcase, res_keys in sorted(iteritems(res_keys_subcase)): # print(res_keys) for res_key in res_keys: if isinstance(res_key, tuple): is_compressed = False else: is_compressed = True res_length = self._get_result_length(res_types, res_key) if res_length == 0: # skipped subcase; no saved results continue res_format = '*%%-%is SUBCASE=%%i%%s' % res_length res_format_vectorized = ' %%-%is SUBCASE=%%i SUBTITLE=%%s %%s' % res_length for res_type in res_types: if res_key not in res_type: continue result = res_type[res_key] if repr_check: str(result) subtitle = result.subtitle label = result.label header = ['', ''] header[0] = ' %-126s\n' % subtitle header[ 1] = '0 %-32s SUBCASE %-15i\n \n' % ( label, isubcase) if result.nonlinear_factor is not None: header.append('') try: element_name = '' if hasattr(result, 'element_name'): element_name = ' - ' + result.element_name class_name = result.__class__.__name__ if hasattr(result, 'data'): if not quiet: print(res_format_vectorized % (class_name, isubcase, subtitle, element_name)) else: print(res_format % (class_name, isubcase, element_name)) try: self.page_num = result.write_f06( f06, header, page_stamp, page_num=self.page_num, is_mag_phase=is_mag_phase, is_sort1=is_sort1) except Exception as error: print_exc(file=sys.stdout) print(''.join(result.get_stats())) raise #assert 'table_name=' in ''.join(result.get_stats()) assert isinstance( self.page_num, int), 'pageNum=%r' % str(self.page_num) except: #print("result name = %r" % result.name()) raise if delete_objects: del result self.page_num += 1