def floatX_convert(s): if s == "32": return "float32" elif s == "64": return "float64" elif s == "16": return "float16" else: return s AddConfigVar('floatX', "Default floating-point precision for python casts.\n" "\n" "Note: float16 support is experimental, use at your own risk.", EnumStr('float64', 'float32', 'float16', convert=floatX_convert,), ) AddConfigVar('warn_float64', "Do an action when a tensor variable with float64 dtype is" " created. They can't be run on the GPU with the current(old)" " gpu back-end and are slow with gamer GPUs.", EnumStr('ignore', 'warn', 'raise', 'pdb'), in_c_key=False, ) AddConfigVar('cast_policy', 'Rules for implicit type casting', EnumStr('custom', 'numpy+floatX', # The 'numpy' policy was originally planned to provide a # smooth transition from numpy. It was meant to behave the
BoolParam(True), in_c_key=False) AddConfigVar('NanGuardMode.inf_is_error', "Default value for inf_is_error", BoolParam(True), in_c_key=False) AddConfigVar('NanGuardMode.big_is_error', "Default value for big_is_error", BoolParam(True), in_c_key=False) AddConfigVar('NanGuardMode.action', "What NanGuardMode does when it finds a problem", EnumStr('raise', 'warn', 'pdb'), in_c_key=False) logger = logging.getLogger("theano.compile.nanguardmode") def flatten(l): """ Turns a nested graph of lists/tuples/other objects into a list of objects. Parameters ---------- l : list/tuple/other objects Might be nested. Returns