repository_name
stringlengths
7
55
func_path_in_repository
stringlengths
4
223
func_name
stringlengths
1
134
whole_func_string
stringlengths
75
104k
language
stringclasses
1 value
func_code_string
stringlengths
75
104k
func_code_tokens
listlengths
19
28.4k
func_documentation_string
stringlengths
1
46.9k
func_documentation_tokens
listlengths
1
1.97k
split_name
stringclasses
1 value
func_code_url
stringlengths
87
315
spacetelescope/stsci.tools
lib/stsci/tools/minmatch.py
MinMatchDict.get
def get(self, key, failobj=None, exact=0): """Raises exception if key is ambiguous""" if not exact: key = self.getfullkey(key,new=1) return self.data.get(key,failobj)
python
def get(self, key, failobj=None, exact=0): """Raises exception if key is ambiguous""" if not exact: key = self.getfullkey(key,new=1) return self.data.get(key,failobj)
[ "def", "get", "(", "self", ",", "key", ",", "failobj", "=", "None", ",", "exact", "=", "0", ")", ":", "if", "not", "exact", ":", "key", "=", "self", ".", "getfullkey", "(", "key", ",", "new", "=", "1", ")", "return", "self", ".", "data", ".", "get", "(", "key", ",", "failobj", ")" ]
Raises exception if key is ambiguous
[ "Raises", "exception", "if", "key", "is", "ambiguous" ]
train
https://github.com/spacetelescope/stsci.tools/blob/9a022503ad24ca54ce83331482dfa3ff6de9f403/lib/stsci/tools/minmatch.py#L136-L140
spacetelescope/stsci.tools
lib/stsci/tools/minmatch.py
MinMatchDict._has
def _has(self, key, exact=0): """Raises an exception if key is ambiguous""" if not exact: key = self.getfullkey(key,new=1) return key in self.data
python
def _has(self, key, exact=0): """Raises an exception if key is ambiguous""" if not exact: key = self.getfullkey(key,new=1) return key in self.data
[ "def", "_has", "(", "self", ",", "key", ",", "exact", "=", "0", ")", ":", "if", "not", "exact", ":", "key", "=", "self", ".", "getfullkey", "(", "key", ",", "new", "=", "1", ")", "return", "key", "in", "self", ".", "data" ]
Raises an exception if key is ambiguous
[ "Raises", "an", "exception", "if", "key", "is", "ambiguous" ]
train
https://github.com/spacetelescope/stsci.tools/blob/9a022503ad24ca54ce83331482dfa3ff6de9f403/lib/stsci/tools/minmatch.py#L169-L173
spacetelescope/stsci.tools
lib/stsci/tools/minmatch.py
MinMatchDict.getall
def getall(self, key, failobj=None): """Returns a list of all the matching values for key, containing a single entry for unambiguous matches and multiple entries for ambiguous matches.""" if self.mmkeys is None: self._mmInit() k = self.mmkeys.get(key) if not k: return failobj return list(map(self.data.get, k))
python
def getall(self, key, failobj=None): """Returns a list of all the matching values for key, containing a single entry for unambiguous matches and multiple entries for ambiguous matches.""" if self.mmkeys is None: self._mmInit() k = self.mmkeys.get(key) if not k: return failobj return list(map(self.data.get, k))
[ "def", "getall", "(", "self", ",", "key", ",", "failobj", "=", "None", ")", ":", "if", "self", ".", "mmkeys", "is", "None", ":", "self", ".", "_mmInit", "(", ")", "k", "=", "self", ".", "mmkeys", ".", "get", "(", "key", ")", "if", "not", "k", ":", "return", "failobj", "return", "list", "(", "map", "(", "self", ".", "data", ".", "get", ",", "k", ")", ")" ]
Returns a list of all the matching values for key, containing a single entry for unambiguous matches and multiple entries for ambiguous matches.
[ "Returns", "a", "list", "of", "all", "the", "matching", "values", "for", "key", "containing", "a", "single", "entry", "for", "unambiguous", "matches", "and", "multiple", "entries", "for", "ambiguous", "matches", "." ]
train
https://github.com/spacetelescope/stsci.tools/blob/9a022503ad24ca54ce83331482dfa3ff6de9f403/lib/stsci/tools/minmatch.py#L195-L202
spacetelescope/stsci.tools
lib/stsci/tools/minmatch.py
MinMatchDict.getallkeys
def getallkeys(self, key, failobj=None): """Returns a list of the full key names (not the items) for all the matching values for key. The list will contain a single entry for unambiguous matches and multiple entries for ambiguous matches.""" if self.mmkeys is None: self._mmInit() return self.mmkeys.get(key, failobj)
python
def getallkeys(self, key, failobj=None): """Returns a list of the full key names (not the items) for all the matching values for key. The list will contain a single entry for unambiguous matches and multiple entries for ambiguous matches.""" if self.mmkeys is None: self._mmInit() return self.mmkeys.get(key, failobj)
[ "def", "getallkeys", "(", "self", ",", "key", ",", "failobj", "=", "None", ")", ":", "if", "self", ".", "mmkeys", "is", "None", ":", "self", ".", "_mmInit", "(", ")", "return", "self", ".", "mmkeys", ".", "get", "(", "key", ",", "failobj", ")" ]
Returns a list of the full key names (not the items) for all the matching values for key. The list will contain a single entry for unambiguous matches and multiple entries for ambiguous matches.
[ "Returns", "a", "list", "of", "the", "full", "key", "names", "(", "not", "the", "items", ")", "for", "all", "the", "matching", "values", "for", "key", ".", "The", "list", "will", "contain", "a", "single", "entry", "for", "unambiguous", "matches", "and", "multiple", "entries", "for", "ambiguous", "matches", "." ]
train
https://github.com/spacetelescope/stsci.tools/blob/9a022503ad24ca54ce83331482dfa3ff6de9f403/lib/stsci/tools/minmatch.py#L204-L210
spacetelescope/stsci.tools
lib/stsci/tools/minmatch.py
QuietMinMatchDict.get
def get(self, key, failobj=None, exact=0): """Returns failobj if key is not found or is ambiguous""" if not exact: try: key = self.getfullkey(key) except KeyError: return failobj return self.data.get(key,failobj)
python
def get(self, key, failobj=None, exact=0): """Returns failobj if key is not found or is ambiguous""" if not exact: try: key = self.getfullkey(key) except KeyError: return failobj return self.data.get(key,failobj)
[ "def", "get", "(", "self", ",", "key", ",", "failobj", "=", "None", ",", "exact", "=", "0", ")", ":", "if", "not", "exact", ":", "try", ":", "key", "=", "self", ".", "getfullkey", "(", "key", ")", "except", "KeyError", ":", "return", "failobj", "return", "self", ".", "data", ".", "get", "(", "key", ",", "failobj", ")" ]
Returns failobj if key is not found or is ambiguous
[ "Returns", "failobj", "if", "key", "is", "not", "found", "or", "is", "ambiguous" ]
train
https://github.com/spacetelescope/stsci.tools/blob/9a022503ad24ca54ce83331482dfa3ff6de9f403/lib/stsci/tools/minmatch.py#L225-L234
spacetelescope/stsci.tools
lib/stsci/tools/minmatch.py
QuietMinMatchDict._has
def _has(self, key, exact=0): """Returns false if key is not found or is ambiguous""" if not exact: try: key = self.getfullkey(key) return 1 except KeyError: return 0 else: return key in self.data
python
def _has(self, key, exact=0): """Returns false if key is not found or is ambiguous""" if not exact: try: key = self.getfullkey(key) return 1 except KeyError: return 0 else: return key in self.data
[ "def", "_has", "(", "self", ",", "key", ",", "exact", "=", "0", ")", ":", "if", "not", "exact", ":", "try", ":", "key", "=", "self", ".", "getfullkey", "(", "key", ")", "return", "1", "except", "KeyError", ":", "return", "0", "else", ":", "return", "key", "in", "self", ".", "data" ]
Returns false if key is not found or is ambiguous
[ "Returns", "false", "if", "key", "is", "not", "found", "or", "is", "ambiguous" ]
train
https://github.com/spacetelescope/stsci.tools/blob/9a022503ad24ca54ce83331482dfa3ff6de9f403/lib/stsci/tools/minmatch.py#L237-L248
spacetelescope/stsci.tools
lib/stsci/tools/basicpar.py
parFactory
def parFactory(fields, strict=0): """parameter factory function fields is a list of the comma-separated fields (as in the .par file). Each entry is a string or None (indicating that field was omitted.) Set the strict parameter to a non-zero value to do stricter parsing (to find errors in the input)""" if len(fields) < 3 or None in fields[0:3]: raise SyntaxError("At least 3 fields must be given") type = fields[1] if type in _string_types: return IrafParS(fields,strict) elif type == 'R': return StrictParR(fields,1) elif type in _real_types: return IrafParR(fields,strict) elif type == "I": return StrictParI(fields,1) elif type == "i": return IrafParI(fields,strict) elif type == "b": return IrafParB(fields,strict) elif type == "ar": return IrafParAR(fields,strict) elif type == "ai": return IrafParAI(fields,strict) elif type == "as": return IrafParAS(fields,strict) elif type == "ab": return IrafParAB(fields,strict) elif type[:1] == "a": raise SyntaxError("Cannot handle arrays of type %s" % type) else: raise SyntaxError("Cannot handle parameter type %s" % type)
python
def parFactory(fields, strict=0): """parameter factory function fields is a list of the comma-separated fields (as in the .par file). Each entry is a string or None (indicating that field was omitted.) Set the strict parameter to a non-zero value to do stricter parsing (to find errors in the input)""" if len(fields) < 3 or None in fields[0:3]: raise SyntaxError("At least 3 fields must be given") type = fields[1] if type in _string_types: return IrafParS(fields,strict) elif type == 'R': return StrictParR(fields,1) elif type in _real_types: return IrafParR(fields,strict) elif type == "I": return StrictParI(fields,1) elif type == "i": return IrafParI(fields,strict) elif type == "b": return IrafParB(fields,strict) elif type == "ar": return IrafParAR(fields,strict) elif type == "ai": return IrafParAI(fields,strict) elif type == "as": return IrafParAS(fields,strict) elif type == "ab": return IrafParAB(fields,strict) elif type[:1] == "a": raise SyntaxError("Cannot handle arrays of type %s" % type) else: raise SyntaxError("Cannot handle parameter type %s" % type)
[ "def", "parFactory", "(", "fields", ",", "strict", "=", "0", ")", ":", "if", "len", "(", "fields", ")", "<", "3", "or", "None", "in", "fields", "[", "0", ":", "3", "]", ":", "raise", "SyntaxError", "(", "\"At least 3 fields must be given\"", ")", "type", "=", "fields", "[", "1", "]", "if", "type", "in", "_string_types", ":", "return", "IrafParS", "(", "fields", ",", "strict", ")", "elif", "type", "==", "'R'", ":", "return", "StrictParR", "(", "fields", ",", "1", ")", "elif", "type", "in", "_real_types", ":", "return", "IrafParR", "(", "fields", ",", "strict", ")", "elif", "type", "==", "\"I\"", ":", "return", "StrictParI", "(", "fields", ",", "1", ")", "elif", "type", "==", "\"i\"", ":", "return", "IrafParI", "(", "fields", ",", "strict", ")", "elif", "type", "==", "\"b\"", ":", "return", "IrafParB", "(", "fields", ",", "strict", ")", "elif", "type", "==", "\"ar\"", ":", "return", "IrafParAR", "(", "fields", ",", "strict", ")", "elif", "type", "==", "\"ai\"", ":", "return", "IrafParAI", "(", "fields", ",", "strict", ")", "elif", "type", "==", "\"as\"", ":", "return", "IrafParAS", "(", "fields", ",", "strict", ")", "elif", "type", "==", "\"ab\"", ":", "return", "IrafParAB", "(", "fields", ",", "strict", ")", "elif", "type", "[", ":", "1", "]", "==", "\"a\"", ":", "raise", "SyntaxError", "(", "\"Cannot handle arrays of type %s\"", "%", "type", ")", "else", ":", "raise", "SyntaxError", "(", "\"Cannot handle parameter type %s\"", "%", "type", ")" ]
parameter factory function fields is a list of the comma-separated fields (as in the .par file). Each entry is a string or None (indicating that field was omitted.) Set the strict parameter to a non-zero value to do stricter parsing (to find errors in the input)
[ "parameter", "factory", "function" ]
train
https://github.com/spacetelescope/stsci.tools/blob/9a022503ad24ca54ce83331482dfa3ff6de9f403/lib/stsci/tools/basicpar.py#L47-L83
spacetelescope/stsci.tools
lib/stsci/tools/basicpar.py
IrafPar.setCmdline
def setCmdline(self,value=1): """Set cmdline flag""" # set through dictionary to avoid extra calls to __setattr__ if value: self.__dict__['flags'] = self.flags | _cmdlineFlag else: self.__dict__['flags'] = self.flags & ~_cmdlineFlag
python
def setCmdline(self,value=1): """Set cmdline flag""" # set through dictionary to avoid extra calls to __setattr__ if value: self.__dict__['flags'] = self.flags | _cmdlineFlag else: self.__dict__['flags'] = self.flags & ~_cmdlineFlag
[ "def", "setCmdline", "(", "self", ",", "value", "=", "1", ")", ":", "# set through dictionary to avoid extra calls to __setattr__", "if", "value", ":", "self", ".", "__dict__", "[", "'flags'", "]", "=", "self", ".", "flags", "|", "_cmdlineFlag", "else", ":", "self", ".", "__dict__", "[", "'flags'", "]", "=", "self", ".", "flags", "&", "~", "_cmdlineFlag" ]
Set cmdline flag
[ "Set", "cmdline", "flag" ]
train
https://github.com/spacetelescope/stsci.tools/blob/9a022503ad24ca54ce83331482dfa3ff6de9f403/lib/stsci/tools/basicpar.py#L244-L250
spacetelescope/stsci.tools
lib/stsci/tools/basicpar.py
IrafPar.setChanged
def setChanged(self,value=1): """Set changed flag""" # set through dictionary to avoid another call to __setattr__ if value: self.__dict__['flags'] = self.flags | _changedFlag else: self.__dict__['flags'] = self.flags & ~_changedFlag
python
def setChanged(self,value=1): """Set changed flag""" # set through dictionary to avoid another call to __setattr__ if value: self.__dict__['flags'] = self.flags | _changedFlag else: self.__dict__['flags'] = self.flags & ~_changedFlag
[ "def", "setChanged", "(", "self", ",", "value", "=", "1", ")", ":", "# set through dictionary to avoid another call to __setattr__", "if", "value", ":", "self", ".", "__dict__", "[", "'flags'", "]", "=", "self", ".", "flags", "|", "_changedFlag", "else", ":", "self", ".", "__dict__", "[", "'flags'", "]", "=", "self", ".", "flags", "&", "~", "_changedFlag" ]
Set changed flag
[ "Set", "changed", "flag" ]
train
https://github.com/spacetelescope/stsci.tools/blob/9a022503ad24ca54ce83331482dfa3ff6de9f403/lib/stsci/tools/basicpar.py#L256-L262
spacetelescope/stsci.tools
lib/stsci/tools/basicpar.py
IrafPar.isLearned
def isLearned(self, mode=None): """Return true if this parameter is learned Hidden parameters are not learned; automatic parameters inherit behavior from package/cl; other parameters are learned. If mode is set, it determines how automatic parameters behave. If not set, cl.mode parameter determines behavior. """ if "l" in self.mode: return 1 if "h" in self.mode: return 0 if "a" in self.mode: if mode is None: mode = 'ql' # that is, iraf.cl.mode if "h" in mode and "l" not in mode: return 0 return 1
python
def isLearned(self, mode=None): """Return true if this parameter is learned Hidden parameters are not learned; automatic parameters inherit behavior from package/cl; other parameters are learned. If mode is set, it determines how automatic parameters behave. If not set, cl.mode parameter determines behavior. """ if "l" in self.mode: return 1 if "h" in self.mode: return 0 if "a" in self.mode: if mode is None: mode = 'ql' # that is, iraf.cl.mode if "h" in mode and "l" not in mode: return 0 return 1
[ "def", "isLearned", "(", "self", ",", "mode", "=", "None", ")", ":", "if", "\"l\"", "in", "self", ".", "mode", ":", "return", "1", "if", "\"h\"", "in", "self", ".", "mode", ":", "return", "0", "if", "\"a\"", "in", "self", ".", "mode", ":", "if", "mode", "is", "None", ":", "mode", "=", "'ql'", "# that is, iraf.cl.mode", "if", "\"h\"", "in", "mode", "and", "\"l\"", "not", "in", "mode", ":", "return", "0", "return", "1" ]
Return true if this parameter is learned Hidden parameters are not learned; automatic parameters inherit behavior from package/cl; other parameters are learned. If mode is set, it determines how automatic parameters behave. If not set, cl.mode parameter determines behavior.
[ "Return", "true", "if", "this", "parameter", "is", "learned" ]
train
https://github.com/spacetelescope/stsci.tools/blob/9a022503ad24ca54ce83331482dfa3ff6de9f403/lib/stsci/tools/basicpar.py#L272-L286
spacetelescope/stsci.tools
lib/stsci/tools/basicpar.py
IrafPar.getWithPrompt
def getWithPrompt(self): """Interactively prompt for parameter value""" if self.prompt: pstring = self.prompt.split("\n")[0].strip() else: pstring = self.name if self.choice: schoice = list(map(self.toString, self.choice)) pstring = pstring + " (" + "|".join(schoice) + ")" elif self.min not in [None, INDEF] or \ self.max not in [None, INDEF]: pstring = pstring + " (" if self.min not in [None, INDEF]: pstring = pstring + self.toString(self.min) pstring = pstring + ":" if self.max not in [None, INDEF]: pstring = pstring + self.toString(self.max) pstring = pstring + ")" # add current value as default if self.value is not None: pstring = pstring + " (" + self.toString(self.value,quoted=1) + ")" pstring = pstring + ": " # don't redirect stdin/out unless redirected filehandles are also ttys # or unless originals are NOT ttys stdout = sys.__stdout__ try: if sys.stdout.isatty() or not stdout.isatty(): stdout = sys.stdout except AttributeError: pass stdin = sys.__stdin__ try: if sys.stdin.isatty() or not stdin.isatty(): stdin = sys.stdin except AttributeError: pass # print prompt, suppressing both newline and following space stdout.write(pstring) stdout.flush() ovalue = irafutils.tkreadline(stdin) value = ovalue.strip() # loop until we get an acceptable value while (1): try: # null input usually means use current value as default # check it anyway since it might not be acceptable if value == "": value = self._nullPrompt() self.set(value) # None (no value) is not acceptable value after prompt if self.value is not None: return # if not EOF, keep looping if ovalue == "": stdout.flush() raise EOFError("EOF on parameter prompt") print("Error: specify a value for the parameter") except ValueError as e: print(str(e)) stdout.write(pstring) stdout.flush() ovalue = irafutils.tkreadline(stdin) value = ovalue.strip()
python
def getWithPrompt(self): """Interactively prompt for parameter value""" if self.prompt: pstring = self.prompt.split("\n")[0].strip() else: pstring = self.name if self.choice: schoice = list(map(self.toString, self.choice)) pstring = pstring + " (" + "|".join(schoice) + ")" elif self.min not in [None, INDEF] or \ self.max not in [None, INDEF]: pstring = pstring + " (" if self.min not in [None, INDEF]: pstring = pstring + self.toString(self.min) pstring = pstring + ":" if self.max not in [None, INDEF]: pstring = pstring + self.toString(self.max) pstring = pstring + ")" # add current value as default if self.value is not None: pstring = pstring + " (" + self.toString(self.value,quoted=1) + ")" pstring = pstring + ": " # don't redirect stdin/out unless redirected filehandles are also ttys # or unless originals are NOT ttys stdout = sys.__stdout__ try: if sys.stdout.isatty() or not stdout.isatty(): stdout = sys.stdout except AttributeError: pass stdin = sys.__stdin__ try: if sys.stdin.isatty() or not stdin.isatty(): stdin = sys.stdin except AttributeError: pass # print prompt, suppressing both newline and following space stdout.write(pstring) stdout.flush() ovalue = irafutils.tkreadline(stdin) value = ovalue.strip() # loop until we get an acceptable value while (1): try: # null input usually means use current value as default # check it anyway since it might not be acceptable if value == "": value = self._nullPrompt() self.set(value) # None (no value) is not acceptable value after prompt if self.value is not None: return # if not EOF, keep looping if ovalue == "": stdout.flush() raise EOFError("EOF on parameter prompt") print("Error: specify a value for the parameter") except ValueError as e: print(str(e)) stdout.write(pstring) stdout.flush() ovalue = irafutils.tkreadline(stdin) value = ovalue.strip()
[ "def", "getWithPrompt", "(", "self", ")", ":", "if", "self", ".", "prompt", ":", "pstring", "=", "self", ".", "prompt", ".", "split", "(", "\"\\n\"", ")", "[", "0", "]", ".", "strip", "(", ")", "else", ":", "pstring", "=", "self", ".", "name", "if", "self", ".", "choice", ":", "schoice", "=", "list", "(", "map", "(", "self", ".", "toString", ",", "self", ".", "choice", ")", ")", "pstring", "=", "pstring", "+", "\" (\"", "+", "\"|\"", ".", "join", "(", "schoice", ")", "+", "\")\"", "elif", "self", ".", "min", "not", "in", "[", "None", ",", "INDEF", "]", "or", "self", ".", "max", "not", "in", "[", "None", ",", "INDEF", "]", ":", "pstring", "=", "pstring", "+", "\" (\"", "if", "self", ".", "min", "not", "in", "[", "None", ",", "INDEF", "]", ":", "pstring", "=", "pstring", "+", "self", ".", "toString", "(", "self", ".", "min", ")", "pstring", "=", "pstring", "+", "\":\"", "if", "self", ".", "max", "not", "in", "[", "None", ",", "INDEF", "]", ":", "pstring", "=", "pstring", "+", "self", ".", "toString", "(", "self", ".", "max", ")", "pstring", "=", "pstring", "+", "\")\"", "# add current value as default", "if", "self", ".", "value", "is", "not", "None", ":", "pstring", "=", "pstring", "+", "\" (\"", "+", "self", ".", "toString", "(", "self", ".", "value", ",", "quoted", "=", "1", ")", "+", "\")\"", "pstring", "=", "pstring", "+", "\": \"", "# don't redirect stdin/out unless redirected filehandles are also ttys", "# or unless originals are NOT ttys", "stdout", "=", "sys", ".", "__stdout__", "try", ":", "if", "sys", ".", "stdout", ".", "isatty", "(", ")", "or", "not", "stdout", ".", "isatty", "(", ")", ":", "stdout", "=", "sys", ".", "stdout", "except", "AttributeError", ":", "pass", "stdin", "=", "sys", ".", "__stdin__", "try", ":", "if", "sys", ".", "stdin", ".", "isatty", "(", ")", "or", "not", "stdin", ".", "isatty", "(", ")", ":", "stdin", "=", "sys", ".", "stdin", "except", "AttributeError", ":", "pass", "# print prompt, suppressing both newline and following space", "stdout", ".", "write", "(", "pstring", ")", "stdout", ".", "flush", "(", ")", "ovalue", "=", "irafutils", ".", "tkreadline", "(", "stdin", ")", "value", "=", "ovalue", ".", "strip", "(", ")", "# loop until we get an acceptable value", "while", "(", "1", ")", ":", "try", ":", "# null input usually means use current value as default", "# check it anyway since it might not be acceptable", "if", "value", "==", "\"\"", ":", "value", "=", "self", ".", "_nullPrompt", "(", ")", "self", ".", "set", "(", "value", ")", "# None (no value) is not acceptable value after prompt", "if", "self", ".", "value", "is", "not", "None", ":", "return", "# if not EOF, keep looping", "if", "ovalue", "==", "\"\"", ":", "stdout", ".", "flush", "(", ")", "raise", "EOFError", "(", "\"EOF on parameter prompt\"", ")", "print", "(", "\"Error: specify a value for the parameter\"", ")", "except", "ValueError", "as", "e", ":", "print", "(", "str", "(", "e", ")", ")", "stdout", ".", "write", "(", "pstring", ")", "stdout", ".", "flush", "(", ")", "ovalue", "=", "irafutils", ".", "tkreadline", "(", "stdin", ")", "value", "=", "ovalue", ".", "strip", "(", ")" ]
Interactively prompt for parameter value
[ "Interactively", "prompt", "for", "parameter", "value" ]
train
https://github.com/spacetelescope/stsci.tools/blob/9a022503ad24ca54ce83331482dfa3ff6de9f403/lib/stsci/tools/basicpar.py#L296-L356
spacetelescope/stsci.tools
lib/stsci/tools/basicpar.py
IrafPar.get
def get(self, field=None, index=None, lpar=0, prompt=1, native=0, mode="h"): """Return value of this parameter as a string (or in native format if native is non-zero.)""" if field and field != "p_value": # note p_value comes back to this routine, so shortcut that case return self._getField(field,native=native,prompt=prompt) # may prompt for value if prompt flag is set if prompt: self._optionalPrompt(mode) if index is not None: raise SyntaxError("Parameter "+self.name+" is not an array") if native: rv = self.value else: rv = self.toString(self.value) return rv
python
def get(self, field=None, index=None, lpar=0, prompt=1, native=0, mode="h"): """Return value of this parameter as a string (or in native format if native is non-zero.)""" if field and field != "p_value": # note p_value comes back to this routine, so shortcut that case return self._getField(field,native=native,prompt=prompt) # may prompt for value if prompt flag is set if prompt: self._optionalPrompt(mode) if index is not None: raise SyntaxError("Parameter "+self.name+" is not an array") if native: rv = self.value else: rv = self.toString(self.value) return rv
[ "def", "get", "(", "self", ",", "field", "=", "None", ",", "index", "=", "None", ",", "lpar", "=", "0", ",", "prompt", "=", "1", ",", "native", "=", "0", ",", "mode", "=", "\"h\"", ")", ":", "if", "field", "and", "field", "!=", "\"p_value\"", ":", "# note p_value comes back to this routine, so shortcut that case", "return", "self", ".", "_getField", "(", "field", ",", "native", "=", "native", ",", "prompt", "=", "prompt", ")", "# may prompt for value if prompt flag is set", "if", "prompt", ":", "self", ".", "_optionalPrompt", "(", "mode", ")", "if", "index", "is", "not", "None", ":", "raise", "SyntaxError", "(", "\"Parameter \"", "+", "self", ".", "name", "+", "\" is not an array\"", ")", "if", "native", ":", "rv", "=", "self", ".", "value", "else", ":", "rv", "=", "self", ".", "toString", "(", "self", ".", "value", ")", "return", "rv" ]
Return value of this parameter as a string (or in native format if native is non-zero.)
[ "Return", "value", "of", "this", "parameter", "as", "a", "string", "(", "or", "in", "native", "format", "if", "native", "is", "non", "-", "zero", ".", ")" ]
train
https://github.com/spacetelescope/stsci.tools/blob/9a022503ad24ca54ce83331482dfa3ff6de9f403/lib/stsci/tools/basicpar.py#L358-L376
spacetelescope/stsci.tools
lib/stsci/tools/basicpar.py
IrafPar.set
def set(self, value, field=None, index=None, check=1): """Set value of this parameter from a string or other value. Field is optional parameter field (p_prompt, p_minimum, etc.) Index is optional array index (zero-based). Set check=0 to assign the value without checking to see if it is within the min-max range or in the choice list.""" if index is not None: raise SyntaxError("Parameter "+self.name+" is not an array") if field: self._setField(value,field,check=check) else: if check: self.value = self.checkValue(value) else: self.value = self._coerceValue(value) self.setChanged()
python
def set(self, value, field=None, index=None, check=1): """Set value of this parameter from a string or other value. Field is optional parameter field (p_prompt, p_minimum, etc.) Index is optional array index (zero-based). Set check=0 to assign the value without checking to see if it is within the min-max range or in the choice list.""" if index is not None: raise SyntaxError("Parameter "+self.name+" is not an array") if field: self._setField(value,field,check=check) else: if check: self.value = self.checkValue(value) else: self.value = self._coerceValue(value) self.setChanged()
[ "def", "set", "(", "self", ",", "value", ",", "field", "=", "None", ",", "index", "=", "None", ",", "check", "=", "1", ")", ":", "if", "index", "is", "not", "None", ":", "raise", "SyntaxError", "(", "\"Parameter \"", "+", "self", ".", "name", "+", "\" is not an array\"", ")", "if", "field", ":", "self", ".", "_setField", "(", "value", ",", "field", ",", "check", "=", "check", ")", "else", ":", "if", "check", ":", "self", ".", "value", "=", "self", ".", "checkValue", "(", "value", ")", "else", ":", "self", ".", "value", "=", "self", ".", "_coerceValue", "(", "value", ")", "self", ".", "setChanged", "(", ")" ]
Set value of this parameter from a string or other value. Field is optional parameter field (p_prompt, p_minimum, etc.) Index is optional array index (zero-based). Set check=0 to assign the value without checking to see if it is within the min-max range or in the choice list.
[ "Set", "value", "of", "this", "parameter", "from", "a", "string", "or", "other", "value", ".", "Field", "is", "optional", "parameter", "field", "(", "p_prompt", "p_minimum", "etc", ".", ")", "Index", "is", "optional", "array", "index", "(", "zero", "-", "based", ")", ".", "Set", "check", "=", "0", "to", "assign", "the", "value", "without", "checking", "to", "see", "if", "it", "is", "within", "the", "min", "-", "max", "range", "or", "in", "the", "choice", "list", "." ]
train
https://github.com/spacetelescope/stsci.tools/blob/9a022503ad24ca54ce83331482dfa3ff6de9f403/lib/stsci/tools/basicpar.py#L378-L395
spacetelescope/stsci.tools
lib/stsci/tools/basicpar.py
IrafPar.checkValue
def checkValue(self,value,strict=0): """Check and convert a parameter value. Raises an exception if the value is not permitted for this parameter. Otherwise returns the value (converted to the right type.) """ v = self._coerceValue(value,strict) return self.checkOneValue(v,strict)
python
def checkValue(self,value,strict=0): """Check and convert a parameter value. Raises an exception if the value is not permitted for this parameter. Otherwise returns the value (converted to the right type.) """ v = self._coerceValue(value,strict) return self.checkOneValue(v,strict)
[ "def", "checkValue", "(", "self", ",", "value", ",", "strict", "=", "0", ")", ":", "v", "=", "self", ".", "_coerceValue", "(", "value", ",", "strict", ")", "return", "self", ".", "checkOneValue", "(", "v", ",", "strict", ")" ]
Check and convert a parameter value. Raises an exception if the value is not permitted for this parameter. Otherwise returns the value (converted to the right type.)
[ "Check", "and", "convert", "a", "parameter", "value", "." ]
train
https://github.com/spacetelescope/stsci.tools/blob/9a022503ad24ca54ce83331482dfa3ff6de9f403/lib/stsci/tools/basicpar.py#L397-L405
spacetelescope/stsci.tools
lib/stsci/tools/basicpar.py
IrafPar.checkOneValue
def checkOneValue(self,v,strict=0): """Checks a single value to see if it is in range or choice list Allows indirection strings starting with ")". Assumes v has already been converted to right value by _coerceOneValue. Returns value if OK, or raises ValueError if not OK. """ if v in [None, INDEF] or (isinstance(v,str) and v[:1] == ")"): return v elif v == "": # most parameters treat null string as omitted value return None elif self.choice is not None and v not in self.choiceDict: schoice = list(map(self.toString, self.choice)) schoice = "|".join(schoice) raise ValueError("Parameter %s: " "value %s is not in choice list (%s)" % (self.name, str(v), schoice)) elif (self.min not in [None, INDEF] and v<self.min): raise ValueError("Parameter %s: " "value `%s' is less than minimum `%s'" % (self.name, str(v), str(self.min))) elif (self.max not in [None, INDEF] and v>self.max): raise ValueError("Parameter %s: " "value `%s' is greater than maximum `%s'" % (self.name, str(v), str(self.max))) return v
python
def checkOneValue(self,v,strict=0): """Checks a single value to see if it is in range or choice list Allows indirection strings starting with ")". Assumes v has already been converted to right value by _coerceOneValue. Returns value if OK, or raises ValueError if not OK. """ if v in [None, INDEF] or (isinstance(v,str) and v[:1] == ")"): return v elif v == "": # most parameters treat null string as omitted value return None elif self.choice is not None and v not in self.choiceDict: schoice = list(map(self.toString, self.choice)) schoice = "|".join(schoice) raise ValueError("Parameter %s: " "value %s is not in choice list (%s)" % (self.name, str(v), schoice)) elif (self.min not in [None, INDEF] and v<self.min): raise ValueError("Parameter %s: " "value `%s' is less than minimum `%s'" % (self.name, str(v), str(self.min))) elif (self.max not in [None, INDEF] and v>self.max): raise ValueError("Parameter %s: " "value `%s' is greater than maximum `%s'" % (self.name, str(v), str(self.max))) return v
[ "def", "checkOneValue", "(", "self", ",", "v", ",", "strict", "=", "0", ")", ":", "if", "v", "in", "[", "None", ",", "INDEF", "]", "or", "(", "isinstance", "(", "v", ",", "str", ")", "and", "v", "[", ":", "1", "]", "==", "\")\"", ")", ":", "return", "v", "elif", "v", "==", "\"\"", ":", "# most parameters treat null string as omitted value", "return", "None", "elif", "self", ".", "choice", "is", "not", "None", "and", "v", "not", "in", "self", ".", "choiceDict", ":", "schoice", "=", "list", "(", "map", "(", "self", ".", "toString", ",", "self", ".", "choice", ")", ")", "schoice", "=", "\"|\"", ".", "join", "(", "schoice", ")", "raise", "ValueError", "(", "\"Parameter %s: \"", "\"value %s is not in choice list (%s)\"", "%", "(", "self", ".", "name", ",", "str", "(", "v", ")", ",", "schoice", ")", ")", "elif", "(", "self", ".", "min", "not", "in", "[", "None", ",", "INDEF", "]", "and", "v", "<", "self", ".", "min", ")", ":", "raise", "ValueError", "(", "\"Parameter %s: \"", "\"value `%s' is less than minimum `%s'\"", "%", "(", "self", ".", "name", ",", "str", "(", "v", ")", ",", "str", "(", "self", ".", "min", ")", ")", ")", "elif", "(", "self", ".", "max", "not", "in", "[", "None", ",", "INDEF", "]", "and", "v", ">", "self", ".", "max", ")", ":", "raise", "ValueError", "(", "\"Parameter %s: \"", "\"value `%s' is greater than maximum `%s'\"", "%", "(", "self", ".", "name", ",", "str", "(", "v", ")", ",", "str", "(", "self", ".", "max", ")", ")", ")", "return", "v" ]
Checks a single value to see if it is in range or choice list Allows indirection strings starting with ")". Assumes v has already been converted to right value by _coerceOneValue. Returns value if OK, or raises ValueError if not OK.
[ "Checks", "a", "single", "value", "to", "see", "if", "it", "is", "in", "range", "or", "choice", "list" ]
train
https://github.com/spacetelescope/stsci.tools/blob/9a022503ad24ca54ce83331482dfa3ff6de9f403/lib/stsci/tools/basicpar.py#L407-L434
spacetelescope/stsci.tools
lib/stsci/tools/basicpar.py
IrafPar.dpar
def dpar(self, cl=1): """Return dpar-style executable assignment for parameter Default is to write CL version of code; if cl parameter is false, writes Python executable code instead. """ sval = self.toString(self.value, quoted=1) if not cl: if sval == "": sval = "None" s = "%s = %s" % (self.name, sval) return s
python
def dpar(self, cl=1): """Return dpar-style executable assignment for parameter Default is to write CL version of code; if cl parameter is false, writes Python executable code instead. """ sval = self.toString(self.value, quoted=1) if not cl: if sval == "": sval = "None" s = "%s = %s" % (self.name, sval) return s
[ "def", "dpar", "(", "self", ",", "cl", "=", "1", ")", ":", "sval", "=", "self", ".", "toString", "(", "self", ".", "value", ",", "quoted", "=", "1", ")", "if", "not", "cl", ":", "if", "sval", "==", "\"\"", ":", "sval", "=", "\"None\"", "s", "=", "\"%s = %s\"", "%", "(", "self", ".", "name", ",", "sval", ")", "return", "s" ]
Return dpar-style executable assignment for parameter Default is to write CL version of code; if cl parameter is false, writes Python executable code instead.
[ "Return", "dpar", "-", "style", "executable", "assignment", "for", "parameter" ]
train
https://github.com/spacetelescope/stsci.tools/blob/9a022503ad24ca54ce83331482dfa3ff6de9f403/lib/stsci/tools/basicpar.py#L436-L446
spacetelescope/stsci.tools
lib/stsci/tools/basicpar.py
IrafPar.pretty
def pretty(self,verbose=0): """Return pretty list description of parameter""" # split prompt lines and add blanks in later lines to align them plines = self.prompt.split('\n') for i in range(len(plines)-1): plines[i+1] = 32*' ' + plines[i+1] plines = '\n'.join(plines) namelen = min(len(self.name), 12) pvalue = self.get(prompt=0,lpar=1) alwaysquoted = ['s', 'f', '*gcur', '*imcur', '*ukey', 'pset'] if self.type in alwaysquoted and self.value is not None: pvalue = '"' + pvalue + '"' if self.mode == "h": s = "%13s = %-15s %s" % ("("+self.name[:namelen], pvalue+")", plines) else: s = "%13s = %-15s %s" % (self.name[:namelen], pvalue, plines) if not verbose: return s if self.choice is not None: s = s + "\n" + 32*" " + "|" nline = 33 for i in range(len(self.choice)): sch = str(self.choice[i]) + "|" s = s + sch nline = nline + len(sch) + 1 if nline > 80: s = s + "\n" + 32*" " + "|" nline = 33 elif self.min not in [None, INDEF] or self.max not in [None, INDEF]: s = s + "\n" + 32*" " if self.min not in [None, INDEF]: s = s + str(self.min) + " <= " s = s + self.name if self.max not in [None, INDEF]: s = s + " <= " + str(self.max) return s
python
def pretty(self,verbose=0): """Return pretty list description of parameter""" # split prompt lines and add blanks in later lines to align them plines = self.prompt.split('\n') for i in range(len(plines)-1): plines[i+1] = 32*' ' + plines[i+1] plines = '\n'.join(plines) namelen = min(len(self.name), 12) pvalue = self.get(prompt=0,lpar=1) alwaysquoted = ['s', 'f', '*gcur', '*imcur', '*ukey', 'pset'] if self.type in alwaysquoted and self.value is not None: pvalue = '"' + pvalue + '"' if self.mode == "h": s = "%13s = %-15s %s" % ("("+self.name[:namelen], pvalue+")", plines) else: s = "%13s = %-15s %s" % (self.name[:namelen], pvalue, plines) if not verbose: return s if self.choice is not None: s = s + "\n" + 32*" " + "|" nline = 33 for i in range(len(self.choice)): sch = str(self.choice[i]) + "|" s = s + sch nline = nline + len(sch) + 1 if nline > 80: s = s + "\n" + 32*" " + "|" nline = 33 elif self.min not in [None, INDEF] or self.max not in [None, INDEF]: s = s + "\n" + 32*" " if self.min not in [None, INDEF]: s = s + str(self.min) + " <= " s = s + self.name if self.max not in [None, INDEF]: s = s + " <= " + str(self.max) return s
[ "def", "pretty", "(", "self", ",", "verbose", "=", "0", ")", ":", "# split prompt lines and add blanks in later lines to align them", "plines", "=", "self", ".", "prompt", ".", "split", "(", "'\\n'", ")", "for", "i", "in", "range", "(", "len", "(", "plines", ")", "-", "1", ")", ":", "plines", "[", "i", "+", "1", "]", "=", "32", "*", "' '", "+", "plines", "[", "i", "+", "1", "]", "plines", "=", "'\\n'", ".", "join", "(", "plines", ")", "namelen", "=", "min", "(", "len", "(", "self", ".", "name", ")", ",", "12", ")", "pvalue", "=", "self", ".", "get", "(", "prompt", "=", "0", ",", "lpar", "=", "1", ")", "alwaysquoted", "=", "[", "'s'", ",", "'f'", ",", "'*gcur'", ",", "'*imcur'", ",", "'*ukey'", ",", "'pset'", "]", "if", "self", ".", "type", "in", "alwaysquoted", "and", "self", ".", "value", "is", "not", "None", ":", "pvalue", "=", "'\"'", "+", "pvalue", "+", "'\"'", "if", "self", ".", "mode", "==", "\"h\"", ":", "s", "=", "\"%13s = %-15s %s\"", "%", "(", "\"(\"", "+", "self", ".", "name", "[", ":", "namelen", "]", ",", "pvalue", "+", "\")\"", ",", "plines", ")", "else", ":", "s", "=", "\"%13s = %-15s %s\"", "%", "(", "self", ".", "name", "[", ":", "namelen", "]", ",", "pvalue", ",", "plines", ")", "if", "not", "verbose", ":", "return", "s", "if", "self", ".", "choice", "is", "not", "None", ":", "s", "=", "s", "+", "\"\\n\"", "+", "32", "*", "\" \"", "+", "\"|\"", "nline", "=", "33", "for", "i", "in", "range", "(", "len", "(", "self", ".", "choice", ")", ")", ":", "sch", "=", "str", "(", "self", ".", "choice", "[", "i", "]", ")", "+", "\"|\"", "s", "=", "s", "+", "sch", "nline", "=", "nline", "+", "len", "(", "sch", ")", "+", "1", "if", "nline", ">", "80", ":", "s", "=", "s", "+", "\"\\n\"", "+", "32", "*", "\" \"", "+", "\"|\"", "nline", "=", "33", "elif", "self", ".", "min", "not", "in", "[", "None", ",", "INDEF", "]", "or", "self", ".", "max", "not", "in", "[", "None", ",", "INDEF", "]", ":", "s", "=", "s", "+", "\"\\n\"", "+", "32", "*", "\" \"", "if", "self", ".", "min", "not", "in", "[", "None", ",", "INDEF", "]", ":", "s", "=", "s", "+", "str", "(", "self", ".", "min", ")", "+", "\" <= \"", "s", "=", "s", "+", "self", ".", "name", "if", "self", ".", "max", "not", "in", "[", "None", ",", "INDEF", "]", ":", "s", "=", "s", "+", "\" <= \"", "+", "str", "(", "self", ".", "max", ")", "return", "s" ]
Return pretty list description of parameter
[ "Return", "pretty", "list", "description", "of", "parameter" ]
train
https://github.com/spacetelescope/stsci.tools/blob/9a022503ad24ca54ce83331482dfa3ff6de9f403/lib/stsci/tools/basicpar.py#L452-L487
spacetelescope/stsci.tools
lib/stsci/tools/basicpar.py
IrafPar.save
def save(self, dolist=0): """Return .par format string for this parameter If dolist is set, returns fields as a list of strings. Default is to return a single string appropriate for writing to a file. """ quoted = not dolist fields = 7*[""] fields[0] = self.name fields[1] = self.type fields[2] = self.mode fields[3] = self.toString(self.value,quoted=quoted) if self.choice is not None: schoice = list(map(self.toString, self.choice)) schoice.insert(0,'') schoice.append('') fields[4] = repr('|'.join(schoice)) elif self.min not in [None,INDEF]: fields[4] = self.toString(self.min,quoted=quoted) if self.max not in [None,INDEF]: fields[5] = self.toString(self.max,quoted=quoted) if self.prompt: if quoted: sprompt = repr(self.prompt) else: sprompt = self.prompt # prompt can have embedded newlines (which are printed) sprompt = sprompt.replace(r'\012', '\n') sprompt = sprompt.replace(r'\n', '\n') fields[6] = sprompt # delete trailing null parameters for i in [6,5,4]: if fields[i] != "": break del fields[i] if dolist: return fields else: return ','.join(fields)
python
def save(self, dolist=0): """Return .par format string for this parameter If dolist is set, returns fields as a list of strings. Default is to return a single string appropriate for writing to a file. """ quoted = not dolist fields = 7*[""] fields[0] = self.name fields[1] = self.type fields[2] = self.mode fields[3] = self.toString(self.value,quoted=quoted) if self.choice is not None: schoice = list(map(self.toString, self.choice)) schoice.insert(0,'') schoice.append('') fields[4] = repr('|'.join(schoice)) elif self.min not in [None,INDEF]: fields[4] = self.toString(self.min,quoted=quoted) if self.max not in [None,INDEF]: fields[5] = self.toString(self.max,quoted=quoted) if self.prompt: if quoted: sprompt = repr(self.prompt) else: sprompt = self.prompt # prompt can have embedded newlines (which are printed) sprompt = sprompt.replace(r'\012', '\n') sprompt = sprompt.replace(r'\n', '\n') fields[6] = sprompt # delete trailing null parameters for i in [6,5,4]: if fields[i] != "": break del fields[i] if dolist: return fields else: return ','.join(fields)
[ "def", "save", "(", "self", ",", "dolist", "=", "0", ")", ":", "quoted", "=", "not", "dolist", "fields", "=", "7", "*", "[", "\"\"", "]", "fields", "[", "0", "]", "=", "self", ".", "name", "fields", "[", "1", "]", "=", "self", ".", "type", "fields", "[", "2", "]", "=", "self", ".", "mode", "fields", "[", "3", "]", "=", "self", ".", "toString", "(", "self", ".", "value", ",", "quoted", "=", "quoted", ")", "if", "self", ".", "choice", "is", "not", "None", ":", "schoice", "=", "list", "(", "map", "(", "self", ".", "toString", ",", "self", ".", "choice", ")", ")", "schoice", ".", "insert", "(", "0", ",", "''", ")", "schoice", ".", "append", "(", "''", ")", "fields", "[", "4", "]", "=", "repr", "(", "'|'", ".", "join", "(", "schoice", ")", ")", "elif", "self", ".", "min", "not", "in", "[", "None", ",", "INDEF", "]", ":", "fields", "[", "4", "]", "=", "self", ".", "toString", "(", "self", ".", "min", ",", "quoted", "=", "quoted", ")", "if", "self", ".", "max", "not", "in", "[", "None", ",", "INDEF", "]", ":", "fields", "[", "5", "]", "=", "self", ".", "toString", "(", "self", ".", "max", ",", "quoted", "=", "quoted", ")", "if", "self", ".", "prompt", ":", "if", "quoted", ":", "sprompt", "=", "repr", "(", "self", ".", "prompt", ")", "else", ":", "sprompt", "=", "self", ".", "prompt", "# prompt can have embedded newlines (which are printed)", "sprompt", "=", "sprompt", ".", "replace", "(", "r'\\012'", ",", "'\\n'", ")", "sprompt", "=", "sprompt", ".", "replace", "(", "r'\\n'", ",", "'\\n'", ")", "fields", "[", "6", "]", "=", "sprompt", "# delete trailing null parameters", "for", "i", "in", "[", "6", ",", "5", ",", "4", "]", ":", "if", "fields", "[", "i", "]", "!=", "\"\"", ":", "break", "del", "fields", "[", "i", "]", "if", "dolist", ":", "return", "fields", "else", ":", "return", "','", ".", "join", "(", "fields", ")" ]
Return .par format string for this parameter If dolist is set, returns fields as a list of strings. Default is to return a single string appropriate for writing to a file.
[ "Return", ".", "par", "format", "string", "for", "this", "parameter" ]
train
https://github.com/spacetelescope/stsci.tools/blob/9a022503ad24ca54ce83331482dfa3ff6de9f403/lib/stsci/tools/basicpar.py#L489-L526
spacetelescope/stsci.tools
lib/stsci/tools/basicpar.py
IrafPar._setChoice
def _setChoice(self,s,strict=0): """Set choice parameter from string s""" clist = _getChoice(s,strict) self.choice = list(map(self._coerceValue, clist)) self._setChoiceDict()
python
def _setChoice(self,s,strict=0): """Set choice parameter from string s""" clist = _getChoice(s,strict) self.choice = list(map(self._coerceValue, clist)) self._setChoiceDict()
[ "def", "_setChoice", "(", "self", ",", "s", ",", "strict", "=", "0", ")", ":", "clist", "=", "_getChoice", "(", "s", ",", "strict", ")", "self", ".", "choice", "=", "list", "(", "map", "(", "self", ".", "_coerceValue", ",", "clist", ")", ")", "self", ".", "_setChoiceDict", "(", ")" ]
Set choice parameter from string s
[ "Set", "choice", "parameter", "from", "string", "s" ]
train
https://github.com/spacetelescope/stsci.tools/blob/9a022503ad24ca54ce83331482dfa3ff6de9f403/lib/stsci/tools/basicpar.py#L615-L619
spacetelescope/stsci.tools
lib/stsci/tools/basicpar.py
IrafPar._setChoiceDict
def _setChoiceDict(self): """Create dictionary for choice list""" # value is name of choice parameter (same as key) self.choiceDict = {} for c in self.choice: self.choiceDict[c] = c
python
def _setChoiceDict(self): """Create dictionary for choice list""" # value is name of choice parameter (same as key) self.choiceDict = {} for c in self.choice: self.choiceDict[c] = c
[ "def", "_setChoiceDict", "(", "self", ")", ":", "# value is name of choice parameter (same as key)", "self", ".", "choiceDict", "=", "{", "}", "for", "c", "in", "self", ".", "choice", ":", "self", ".", "choiceDict", "[", "c", "]", "=", "c" ]
Create dictionary for choice list
[ "Create", "dictionary", "for", "choice", "list" ]
train
https://github.com/spacetelescope/stsci.tools/blob/9a022503ad24ca54ce83331482dfa3ff6de9f403/lib/stsci/tools/basicpar.py#L621-L625
spacetelescope/stsci.tools
lib/stsci/tools/basicpar.py
IrafPar._optionalPrompt
def _optionalPrompt(self, mode): """Interactively prompt for parameter if necessary Prompt for value if (1) mode is hidden but value is undefined or bad, or (2) mode is query and value was not set on command line Never prompt for "u" mode parameters, which are local variables. """ if (self.mode == "h") or (self.mode == "a" and mode == "h"): # hidden parameter if not self.isLegal(): self.getWithPrompt() elif self.mode == "u": # "u" is a special mode used for local variables in CL scripts # They should never prompt under any circumstances if not self.isLegal(): raise ValueError( "Attempt to access undefined local variable `%s'" % self.name) else: # query parameter if self.isCmdline()==0: self.getWithPrompt()
python
def _optionalPrompt(self, mode): """Interactively prompt for parameter if necessary Prompt for value if (1) mode is hidden but value is undefined or bad, or (2) mode is query and value was not set on command line Never prompt for "u" mode parameters, which are local variables. """ if (self.mode == "h") or (self.mode == "a" and mode == "h"): # hidden parameter if not self.isLegal(): self.getWithPrompt() elif self.mode == "u": # "u" is a special mode used for local variables in CL scripts # They should never prompt under any circumstances if not self.isLegal(): raise ValueError( "Attempt to access undefined local variable `%s'" % self.name) else: # query parameter if self.isCmdline()==0: self.getWithPrompt()
[ "def", "_optionalPrompt", "(", "self", ",", "mode", ")", ":", "if", "(", "self", ".", "mode", "==", "\"h\"", ")", "or", "(", "self", ".", "mode", "==", "\"a\"", "and", "mode", "==", "\"h\"", ")", ":", "# hidden parameter", "if", "not", "self", ".", "isLegal", "(", ")", ":", "self", ".", "getWithPrompt", "(", ")", "elif", "self", ".", "mode", "==", "\"u\"", ":", "# \"u\" is a special mode used for local variables in CL scripts", "# They should never prompt under any circumstances", "if", "not", "self", ".", "isLegal", "(", ")", ":", "raise", "ValueError", "(", "\"Attempt to access undefined local variable `%s'\"", "%", "self", ".", "name", ")", "else", ":", "# query parameter", "if", "self", ".", "isCmdline", "(", ")", "==", "0", ":", "self", ".", "getWithPrompt", "(", ")" ]
Interactively prompt for parameter if necessary Prompt for value if (1) mode is hidden but value is undefined or bad, or (2) mode is query and value was not set on command line Never prompt for "u" mode parameters, which are local variables.
[ "Interactively", "prompt", "for", "parameter", "if", "necessary" ]
train
https://github.com/spacetelescope/stsci.tools/blob/9a022503ad24ca54ce83331482dfa3ff6de9f403/lib/stsci/tools/basicpar.py#L632-L654
spacetelescope/stsci.tools
lib/stsci/tools/basicpar.py
IrafPar._getPFilename
def _getPFilename(self,native,prompt): """Get p_filename field for this parameter Same as get for non-list params """ return self.get(native=native,prompt=prompt)
python
def _getPFilename(self,native,prompt): """Get p_filename field for this parameter Same as get for non-list params """ return self.get(native=native,prompt=prompt)
[ "def", "_getPFilename", "(", "self", ",", "native", ",", "prompt", ")", ":", "return", "self", ".", "get", "(", "native", "=", "native", ",", "prompt", "=", "prompt", ")" ]
Get p_filename field for this parameter Same as get for non-list params
[ "Get", "p_filename", "field", "for", "this", "parameter" ]
train
https://github.com/spacetelescope/stsci.tools/blob/9a022503ad24ca54ce83331482dfa3ff6de9f403/lib/stsci/tools/basicpar.py#L656-L661
spacetelescope/stsci.tools
lib/stsci/tools/basicpar.py
IrafPar._getField
def _getField(self, field, native=0, prompt=1): """Get a parameter field value""" try: # expand field name using minimum match field = _getFieldDict[field] except KeyError as e: # re-raise the exception with a bit more info raise SyntaxError("Cannot get field " + field + " for parameter " + self.name + "\n" + str(e)) if field == "p_value": # return value of parameter # Note that IRAF returns the filename for list parameters # when p_value is used. I consider this a bug, and it does # not appear to be used by any cl scripts or SPP programs # in either IRAF or STSDAS. It is also in conflict with # the IRAF help documentation. I am making p_value exactly # the same as just a simple CL parameter reference. return self.get(native=native,prompt=prompt) elif field == "p_name": return self.name elif field == "p_xtype": return self.type elif field == "p_type": return self._getPType() elif field == "p_mode": return self.mode elif field == "p_prompt": return self.prompt elif field == "p_scope": return self.scope elif field == "p_default" or field == "p_filename": # these all appear to be equivalent -- they just return the # current PFilename of the parameter (which is the same as the value # for non-list parameters, and is the filename for list parameters) return self._getPFilename(native,prompt) elif field == "p_maximum": if native: return self.max else: return self.toString(self.max) elif field == "p_minimum": if self.choice is not None: if native: return self.choice else: schoice = list(map(self.toString, self.choice)) return "|" + "|".join(schoice) + "|" else: if native: return self.min else: return self.toString(self.min) else: # XXX unimplemented fields: # p_length: maximum string length in bytes -- what to do with it? raise RuntimeError("Program bug in IrafPar._getField()\n" + "Requested field " + field + " for parameter " + self.name)
python
def _getField(self, field, native=0, prompt=1): """Get a parameter field value""" try: # expand field name using minimum match field = _getFieldDict[field] except KeyError as e: # re-raise the exception with a bit more info raise SyntaxError("Cannot get field " + field + " for parameter " + self.name + "\n" + str(e)) if field == "p_value": # return value of parameter # Note that IRAF returns the filename for list parameters # when p_value is used. I consider this a bug, and it does # not appear to be used by any cl scripts or SPP programs # in either IRAF or STSDAS. It is also in conflict with # the IRAF help documentation. I am making p_value exactly # the same as just a simple CL parameter reference. return self.get(native=native,prompt=prompt) elif field == "p_name": return self.name elif field == "p_xtype": return self.type elif field == "p_type": return self._getPType() elif field == "p_mode": return self.mode elif field == "p_prompt": return self.prompt elif field == "p_scope": return self.scope elif field == "p_default" or field == "p_filename": # these all appear to be equivalent -- they just return the # current PFilename of the parameter (which is the same as the value # for non-list parameters, and is the filename for list parameters) return self._getPFilename(native,prompt) elif field == "p_maximum": if native: return self.max else: return self.toString(self.max) elif field == "p_minimum": if self.choice is not None: if native: return self.choice else: schoice = list(map(self.toString, self.choice)) return "|" + "|".join(schoice) + "|" else: if native: return self.min else: return self.toString(self.min) else: # XXX unimplemented fields: # p_length: maximum string length in bytes -- what to do with it? raise RuntimeError("Program bug in IrafPar._getField()\n" + "Requested field " + field + " for parameter " + self.name)
[ "def", "_getField", "(", "self", ",", "field", ",", "native", "=", "0", ",", "prompt", "=", "1", ")", ":", "try", ":", "# expand field name using minimum match", "field", "=", "_getFieldDict", "[", "field", "]", "except", "KeyError", "as", "e", ":", "# re-raise the exception with a bit more info", "raise", "SyntaxError", "(", "\"Cannot get field \"", "+", "field", "+", "\" for parameter \"", "+", "self", ".", "name", "+", "\"\\n\"", "+", "str", "(", "e", ")", ")", "if", "field", "==", "\"p_value\"", ":", "# return value of parameter", "# Note that IRAF returns the filename for list parameters", "# when p_value is used. I consider this a bug, and it does", "# not appear to be used by any cl scripts or SPP programs", "# in either IRAF or STSDAS. It is also in conflict with", "# the IRAF help documentation. I am making p_value exactly", "# the same as just a simple CL parameter reference.", "return", "self", ".", "get", "(", "native", "=", "native", ",", "prompt", "=", "prompt", ")", "elif", "field", "==", "\"p_name\"", ":", "return", "self", ".", "name", "elif", "field", "==", "\"p_xtype\"", ":", "return", "self", ".", "type", "elif", "field", "==", "\"p_type\"", ":", "return", "self", ".", "_getPType", "(", ")", "elif", "field", "==", "\"p_mode\"", ":", "return", "self", ".", "mode", "elif", "field", "==", "\"p_prompt\"", ":", "return", "self", ".", "prompt", "elif", "field", "==", "\"p_scope\"", ":", "return", "self", ".", "scope", "elif", "field", "==", "\"p_default\"", "or", "field", "==", "\"p_filename\"", ":", "# these all appear to be equivalent -- they just return the", "# current PFilename of the parameter (which is the same as the value", "# for non-list parameters, and is the filename for list parameters)", "return", "self", ".", "_getPFilename", "(", "native", ",", "prompt", ")", "elif", "field", "==", "\"p_maximum\"", ":", "if", "native", ":", "return", "self", ".", "max", "else", ":", "return", "self", ".", "toString", "(", "self", ".", "max", ")", "elif", "field", "==", "\"p_minimum\"", ":", "if", "self", ".", "choice", "is", "not", "None", ":", "if", "native", ":", "return", "self", ".", "choice", "else", ":", "schoice", "=", "list", "(", "map", "(", "self", ".", "toString", ",", "self", ".", "choice", ")", ")", "return", "\"|\"", "+", "\"|\"", ".", "join", "(", "schoice", ")", "+", "\"|\"", "else", ":", "if", "native", ":", "return", "self", ".", "min", "else", ":", "return", "self", ".", "toString", "(", "self", ".", "min", ")", "else", ":", "# XXX unimplemented fields:", "# p_length: maximum string length in bytes -- what to do with it?", "raise", "RuntimeError", "(", "\"Program bug in IrafPar._getField()\\n\"", "+", "\"Requested field \"", "+", "field", "+", "\" for parameter \"", "+", "self", ".", "name", ")" ]
Get a parameter field value
[ "Get", "a", "parameter", "field", "value" ]
train
https://github.com/spacetelescope/stsci.tools/blob/9a022503ad24ca54ce83331482dfa3ff6de9f403/lib/stsci/tools/basicpar.py#L670-L720
spacetelescope/stsci.tools
lib/stsci/tools/basicpar.py
IrafPar._setField
def _setField(self, value, field, check=1): """Set a parameter field value""" try: # expand field name using minimum match field = _setFieldDict[field] except KeyError as e: raise SyntaxError("Cannot set field " + field + " for parameter " + self.name + "\n" + str(e)) if field == "p_prompt": self.prompt = irafutils.removeEscapes(irafutils.stripQuotes(value)) elif field == "p_value": self.set(value,check=check) elif field == "p_filename": # this is only relevant for list parameters (*imcur, *gcur, etc.) self.set(value,check=check) elif field == "p_scope": self.scope = value elif field == "p_maximum": self.max = self._coerceOneValue(value) elif field == "p_minimum": if isinstance(value,str) and '|' in value: self._setChoice(irafutils.stripQuotes(value)) else: self.min = self._coerceOneValue(value) elif field == "p_mode": # not doing any type or value checking here -- setting mode is # rare, so assume that it is being done correctly self.mode = irafutils.stripQuotes(value) else: raise RuntimeError("Program bug in IrafPar._setField()" + "Requested field " + field + " for parameter " + self.name)
python
def _setField(self, value, field, check=1): """Set a parameter field value""" try: # expand field name using minimum match field = _setFieldDict[field] except KeyError as e: raise SyntaxError("Cannot set field " + field + " for parameter " + self.name + "\n" + str(e)) if field == "p_prompt": self.prompt = irafutils.removeEscapes(irafutils.stripQuotes(value)) elif field == "p_value": self.set(value,check=check) elif field == "p_filename": # this is only relevant for list parameters (*imcur, *gcur, etc.) self.set(value,check=check) elif field == "p_scope": self.scope = value elif field == "p_maximum": self.max = self._coerceOneValue(value) elif field == "p_minimum": if isinstance(value,str) and '|' in value: self._setChoice(irafutils.stripQuotes(value)) else: self.min = self._coerceOneValue(value) elif field == "p_mode": # not doing any type or value checking here -- setting mode is # rare, so assume that it is being done correctly self.mode = irafutils.stripQuotes(value) else: raise RuntimeError("Program bug in IrafPar._setField()" + "Requested field " + field + " for parameter " + self.name)
[ "def", "_setField", "(", "self", ",", "value", ",", "field", ",", "check", "=", "1", ")", ":", "try", ":", "# expand field name using minimum match", "field", "=", "_setFieldDict", "[", "field", "]", "except", "KeyError", "as", "e", ":", "raise", "SyntaxError", "(", "\"Cannot set field \"", "+", "field", "+", "\" for parameter \"", "+", "self", ".", "name", "+", "\"\\n\"", "+", "str", "(", "e", ")", ")", "if", "field", "==", "\"p_prompt\"", ":", "self", ".", "prompt", "=", "irafutils", ".", "removeEscapes", "(", "irafutils", ".", "stripQuotes", "(", "value", ")", ")", "elif", "field", "==", "\"p_value\"", ":", "self", ".", "set", "(", "value", ",", "check", "=", "check", ")", "elif", "field", "==", "\"p_filename\"", ":", "# this is only relevant for list parameters (*imcur, *gcur, etc.)", "self", ".", "set", "(", "value", ",", "check", "=", "check", ")", "elif", "field", "==", "\"p_scope\"", ":", "self", ".", "scope", "=", "value", "elif", "field", "==", "\"p_maximum\"", ":", "self", ".", "max", "=", "self", ".", "_coerceOneValue", "(", "value", ")", "elif", "field", "==", "\"p_minimum\"", ":", "if", "isinstance", "(", "value", ",", "str", ")", "and", "'|'", "in", "value", ":", "self", ".", "_setChoice", "(", "irafutils", ".", "stripQuotes", "(", "value", ")", ")", "else", ":", "self", ".", "min", "=", "self", ".", "_coerceOneValue", "(", "value", ")", "elif", "field", "==", "\"p_mode\"", ":", "# not doing any type or value checking here -- setting mode is", "# rare, so assume that it is being done correctly", "self", ".", "mode", "=", "irafutils", ".", "stripQuotes", "(", "value", ")", "else", ":", "raise", "RuntimeError", "(", "\"Program bug in IrafPar._setField()\"", "+", "\"Requested field \"", "+", "field", "+", "\" for parameter \"", "+", "self", ".", "name", ")" ]
Set a parameter field value
[ "Set", "a", "parameter", "field", "value" ]
train
https://github.com/spacetelescope/stsci.tools/blob/9a022503ad24ca54ce83331482dfa3ff6de9f403/lib/stsci/tools/basicpar.py#L722-L752
spacetelescope/stsci.tools
lib/stsci/tools/basicpar.py
IrafArrayPar.save
def save(self, dolist=0): """Return .par format string for this parameter If dolist is set, returns fields as a list of strings. Default is to return a single string appropriate for writing to a file. """ quoted = not dolist array_size = 1 for d in self.shape: array_size = d*array_size ndim = len(self.shape) fields = (7+2*ndim+len(self.value))*[""] fields[0] = self.name fields[1] = self.type fields[2] = self.mode fields[3] = str(ndim) next = 4 for d in self.shape: fields[next] = str(d); next += 1 fields[next] = '1'; next += 1 nvstart = 7+2*ndim if self.choice is not None: schoice = list(map(self.toString, self.choice)) schoice.insert(0,'') schoice.append('') fields[nvstart-3] = repr('|'.join(schoice)) elif self.min not in [None,INDEF]: fields[nvstart-3] = self.toString(self.min,quoted=quoted) # insert an escaped line break before min field if quoted: fields[nvstart-3] = '\\\n' + fields[nvstart-3] if self.max not in [None,INDEF]: fields[nvstart-2] = self.toString(self.max,quoted=quoted) if self.prompt: if quoted: sprompt = repr(self.prompt) else: sprompt = self.prompt # prompt can have embedded newlines (which are printed) sprompt = sprompt.replace(r'\012', '\n') sprompt = sprompt.replace(r'\n', '\n') fields[nvstart-1] = sprompt for i in range(len(self.value)): fields[nvstart+i] = self.toString(self.value[i],quoted=quoted) # insert an escaped line break before value fields if dolist: return fields else: fields[nvstart] = '\\\n' + fields[nvstart] return ','.join(fields)
python
def save(self, dolist=0): """Return .par format string for this parameter If dolist is set, returns fields as a list of strings. Default is to return a single string appropriate for writing to a file. """ quoted = not dolist array_size = 1 for d in self.shape: array_size = d*array_size ndim = len(self.shape) fields = (7+2*ndim+len(self.value))*[""] fields[0] = self.name fields[1] = self.type fields[2] = self.mode fields[3] = str(ndim) next = 4 for d in self.shape: fields[next] = str(d); next += 1 fields[next] = '1'; next += 1 nvstart = 7+2*ndim if self.choice is not None: schoice = list(map(self.toString, self.choice)) schoice.insert(0,'') schoice.append('') fields[nvstart-3] = repr('|'.join(schoice)) elif self.min not in [None,INDEF]: fields[nvstart-3] = self.toString(self.min,quoted=quoted) # insert an escaped line break before min field if quoted: fields[nvstart-3] = '\\\n' + fields[nvstart-3] if self.max not in [None,INDEF]: fields[nvstart-2] = self.toString(self.max,quoted=quoted) if self.prompt: if quoted: sprompt = repr(self.prompt) else: sprompt = self.prompt # prompt can have embedded newlines (which are printed) sprompt = sprompt.replace(r'\012', '\n') sprompt = sprompt.replace(r'\n', '\n') fields[nvstart-1] = sprompt for i in range(len(self.value)): fields[nvstart+i] = self.toString(self.value[i],quoted=quoted) # insert an escaped line break before value fields if dolist: return fields else: fields[nvstart] = '\\\n' + fields[nvstart] return ','.join(fields)
[ "def", "save", "(", "self", ",", "dolist", "=", "0", ")", ":", "quoted", "=", "not", "dolist", "array_size", "=", "1", "for", "d", "in", "self", ".", "shape", ":", "array_size", "=", "d", "*", "array_size", "ndim", "=", "len", "(", "self", ".", "shape", ")", "fields", "=", "(", "7", "+", "2", "*", "ndim", "+", "len", "(", "self", ".", "value", ")", ")", "*", "[", "\"\"", "]", "fields", "[", "0", "]", "=", "self", ".", "name", "fields", "[", "1", "]", "=", "self", ".", "type", "fields", "[", "2", "]", "=", "self", ".", "mode", "fields", "[", "3", "]", "=", "str", "(", "ndim", ")", "next", "=", "4", "for", "d", "in", "self", ".", "shape", ":", "fields", "[", "next", "]", "=", "str", "(", "d", ")", "next", "+=", "1", "fields", "[", "next", "]", "=", "'1'", "next", "+=", "1", "nvstart", "=", "7", "+", "2", "*", "ndim", "if", "self", ".", "choice", "is", "not", "None", ":", "schoice", "=", "list", "(", "map", "(", "self", ".", "toString", ",", "self", ".", "choice", ")", ")", "schoice", ".", "insert", "(", "0", ",", "''", ")", "schoice", ".", "append", "(", "''", ")", "fields", "[", "nvstart", "-", "3", "]", "=", "repr", "(", "'|'", ".", "join", "(", "schoice", ")", ")", "elif", "self", ".", "min", "not", "in", "[", "None", ",", "INDEF", "]", ":", "fields", "[", "nvstart", "-", "3", "]", "=", "self", ".", "toString", "(", "self", ".", "min", ",", "quoted", "=", "quoted", ")", "# insert an escaped line break before min field", "if", "quoted", ":", "fields", "[", "nvstart", "-", "3", "]", "=", "'\\\\\\n'", "+", "fields", "[", "nvstart", "-", "3", "]", "if", "self", ".", "max", "not", "in", "[", "None", ",", "INDEF", "]", ":", "fields", "[", "nvstart", "-", "2", "]", "=", "self", ".", "toString", "(", "self", ".", "max", ",", "quoted", "=", "quoted", ")", "if", "self", ".", "prompt", ":", "if", "quoted", ":", "sprompt", "=", "repr", "(", "self", ".", "prompt", ")", "else", ":", "sprompt", "=", "self", ".", "prompt", "# prompt can have embedded newlines (which are printed)", "sprompt", "=", "sprompt", ".", "replace", "(", "r'\\012'", ",", "'\\n'", ")", "sprompt", "=", "sprompt", ".", "replace", "(", "r'\\n'", ",", "'\\n'", ")", "fields", "[", "nvstart", "-", "1", "]", "=", "sprompt", "for", "i", "in", "range", "(", "len", "(", "self", ".", "value", ")", ")", ":", "fields", "[", "nvstart", "+", "i", "]", "=", "self", ".", "toString", "(", "self", ".", "value", "[", "i", "]", ",", "quoted", "=", "quoted", ")", "# insert an escaped line break before value fields", "if", "dolist", ":", "return", "fields", "else", ":", "fields", "[", "nvstart", "]", "=", "'\\\\\\n'", "+", "fields", "[", "nvstart", "]", "return", "','", ".", "join", "(", "fields", ")" ]
Return .par format string for this parameter If dolist is set, returns fields as a list of strings. Default is to return a single string appropriate for writing to a file.
[ "Return", ".", "par", "format", "string", "for", "this", "parameter" ]
train
https://github.com/spacetelescope/stsci.tools/blob/9a022503ad24ca54ce83331482dfa3ff6de9f403/lib/stsci/tools/basicpar.py#L864-L913
spacetelescope/stsci.tools
lib/stsci/tools/basicpar.py
IrafArrayPar.dpar
def dpar(self, cl=1): """Return dpar-style executable assignment for parameter Default is to write CL version of code; if cl parameter is false, writes Python executable code instead. Note that dpar doesn't even work for arrays in the CL, so we just use Python syntax here. """ sval = list(map(self.toString, self.value, len(self.value)*[1])) for i in range(len(sval)): if sval[i] == "": sval[i] = "None" s = "%s = [%s]" % (self.name, ', '.join(sval)) return s
python
def dpar(self, cl=1): """Return dpar-style executable assignment for parameter Default is to write CL version of code; if cl parameter is false, writes Python executable code instead. Note that dpar doesn't even work for arrays in the CL, so we just use Python syntax here. """ sval = list(map(self.toString, self.value, len(self.value)*[1])) for i in range(len(sval)): if sval[i] == "": sval[i] = "None" s = "%s = [%s]" % (self.name, ', '.join(sval)) return s
[ "def", "dpar", "(", "self", ",", "cl", "=", "1", ")", ":", "sval", "=", "list", "(", "map", "(", "self", ".", "toString", ",", "self", ".", "value", ",", "len", "(", "self", ".", "value", ")", "*", "[", "1", "]", ")", ")", "for", "i", "in", "range", "(", "len", "(", "sval", ")", ")", ":", "if", "sval", "[", "i", "]", "==", "\"\"", ":", "sval", "[", "i", "]", "=", "\"None\"", "s", "=", "\"%s = [%s]\"", "%", "(", "self", ".", "name", ",", "', '", ".", "join", "(", "sval", ")", ")", "return", "s" ]
Return dpar-style executable assignment for parameter Default is to write CL version of code; if cl parameter is false, writes Python executable code instead. Note that dpar doesn't even work for arrays in the CL, so we just use Python syntax here.
[ "Return", "dpar", "-", "style", "executable", "assignment", "for", "parameter" ]
train
https://github.com/spacetelescope/stsci.tools/blob/9a022503ad24ca54ce83331482dfa3ff6de9f403/lib/stsci/tools/basicpar.py#L915-L928
spacetelescope/stsci.tools
lib/stsci/tools/basicpar.py
IrafArrayPar.get
def get(self, field=None, index=None, lpar=0, prompt=1, native=0, mode="h"): """Return value of this parameter as a string (or in native format if native is non-zero.)""" if field: return self._getField(field,native=native,prompt=prompt) # may prompt for value if prompt flag is set #XXX should change _optionalPrompt so we prompt for each element of #XXX the array separately? I think array parameters are #XXX not useful as non-hidden params. if prompt: self._optionalPrompt(mode) if index is not None: sumindex = self._sumindex(index) try: if native: return self.value[sumindex] else: return self.toString(self.value[sumindex]) except IndexError: # should never happen raise SyntaxError("Illegal index [" + repr(sumindex) + "] for array parameter " + self.name) elif native: # return object itself for an array because it is # indexable, can have values assigned, etc. return self else: # return blank-separated string of values for array return str(self)
python
def get(self, field=None, index=None, lpar=0, prompt=1, native=0, mode="h"): """Return value of this parameter as a string (or in native format if native is non-zero.)""" if field: return self._getField(field,native=native,prompt=prompt) # may prompt for value if prompt flag is set #XXX should change _optionalPrompt so we prompt for each element of #XXX the array separately? I think array parameters are #XXX not useful as non-hidden params. if prompt: self._optionalPrompt(mode) if index is not None: sumindex = self._sumindex(index) try: if native: return self.value[sumindex] else: return self.toString(self.value[sumindex]) except IndexError: # should never happen raise SyntaxError("Illegal index [" + repr(sumindex) + "] for array parameter " + self.name) elif native: # return object itself for an array because it is # indexable, can have values assigned, etc. return self else: # return blank-separated string of values for array return str(self)
[ "def", "get", "(", "self", ",", "field", "=", "None", ",", "index", "=", "None", ",", "lpar", "=", "0", ",", "prompt", "=", "1", ",", "native", "=", "0", ",", "mode", "=", "\"h\"", ")", ":", "if", "field", ":", "return", "self", ".", "_getField", "(", "field", ",", "native", "=", "native", ",", "prompt", "=", "prompt", ")", "# may prompt for value if prompt flag is set", "#XXX should change _optionalPrompt so we prompt for each element of", "#XXX the array separately? I think array parameters are", "#XXX not useful as non-hidden params.", "if", "prompt", ":", "self", ".", "_optionalPrompt", "(", "mode", ")", "if", "index", "is", "not", "None", ":", "sumindex", "=", "self", ".", "_sumindex", "(", "index", ")", "try", ":", "if", "native", ":", "return", "self", ".", "value", "[", "sumindex", "]", "else", ":", "return", "self", ".", "toString", "(", "self", ".", "value", "[", "sumindex", "]", ")", "except", "IndexError", ":", "# should never happen", "raise", "SyntaxError", "(", "\"Illegal index [\"", "+", "repr", "(", "sumindex", ")", "+", "\"] for array parameter \"", "+", "self", ".", "name", ")", "elif", "native", ":", "# return object itself for an array because it is", "# indexable, can have values assigned, etc.", "return", "self", "else", ":", "# return blank-separated string of values for array", "return", "str", "(", "self", ")" ]
Return value of this parameter as a string (or in native format if native is non-zero.)
[ "Return", "value", "of", "this", "parameter", "as", "a", "string", "(", "or", "in", "native", "format", "if", "native", "is", "non", "-", "zero", ".", ")" ]
train
https://github.com/spacetelescope/stsci.tools/blob/9a022503ad24ca54ce83331482dfa3ff6de9f403/lib/stsci/tools/basicpar.py#L930-L960
spacetelescope/stsci.tools
lib/stsci/tools/basicpar.py
IrafArrayPar.set
def set(self, value, field=None, index=None, check=1): """Set value of this parameter from a string or other value. Field is optional parameter field (p_prompt, p_minimum, etc.) Index is optional array index (zero-based). Set check=0 to assign the value without checking to see if it is within the min-max range or in the choice list.""" if index is not None: sumindex = self._sumindex(index) try: value = self._coerceOneValue(value) if check: self.value[sumindex] = self.checkOneValue(value) else: self.value[sumindex] = value return except IndexError: # should never happen raise SyntaxError("Illegal index [" + repr(sumindex) + "] for array parameter " + self.name) if field: self._setField(value,field,check=check) else: if check: self.value = self.checkValue(value) else: self.value = self._coerceValue(value) self.setChanged()
python
def set(self, value, field=None, index=None, check=1): """Set value of this parameter from a string or other value. Field is optional parameter field (p_prompt, p_minimum, etc.) Index is optional array index (zero-based). Set check=0 to assign the value without checking to see if it is within the min-max range or in the choice list.""" if index is not None: sumindex = self._sumindex(index) try: value = self._coerceOneValue(value) if check: self.value[sumindex] = self.checkOneValue(value) else: self.value[sumindex] = value return except IndexError: # should never happen raise SyntaxError("Illegal index [" + repr(sumindex) + "] for array parameter " + self.name) if field: self._setField(value,field,check=check) else: if check: self.value = self.checkValue(value) else: self.value = self._coerceValue(value) self.setChanged()
[ "def", "set", "(", "self", ",", "value", ",", "field", "=", "None", ",", "index", "=", "None", ",", "check", "=", "1", ")", ":", "if", "index", "is", "not", "None", ":", "sumindex", "=", "self", ".", "_sumindex", "(", "index", ")", "try", ":", "value", "=", "self", ".", "_coerceOneValue", "(", "value", ")", "if", "check", ":", "self", ".", "value", "[", "sumindex", "]", "=", "self", ".", "checkOneValue", "(", "value", ")", "else", ":", "self", ".", "value", "[", "sumindex", "]", "=", "value", "return", "except", "IndexError", ":", "# should never happen", "raise", "SyntaxError", "(", "\"Illegal index [\"", "+", "repr", "(", "sumindex", ")", "+", "\"] for array parameter \"", "+", "self", ".", "name", ")", "if", "field", ":", "self", ".", "_setField", "(", "value", ",", "field", ",", "check", "=", "check", ")", "else", ":", "if", "check", ":", "self", ".", "value", "=", "self", ".", "checkValue", "(", "value", ")", "else", ":", "self", ".", "value", "=", "self", ".", "_coerceValue", "(", "value", ")", "self", ".", "setChanged", "(", ")" ]
Set value of this parameter from a string or other value. Field is optional parameter field (p_prompt, p_minimum, etc.) Index is optional array index (zero-based). Set check=0 to assign the value without checking to see if it is within the min-max range or in the choice list.
[ "Set", "value", "of", "this", "parameter", "from", "a", "string", "or", "other", "value", ".", "Field", "is", "optional", "parameter", "field", "(", "p_prompt", "p_minimum", "etc", ".", ")", "Index", "is", "optional", "array", "index", "(", "zero", "-", "based", ")", ".", "Set", "check", "=", "0", "to", "assign", "the", "value", "without", "checking", "to", "see", "if", "it", "is", "within", "the", "min", "-", "max", "range", "or", "in", "the", "choice", "list", "." ]
train
https://github.com/spacetelescope/stsci.tools/blob/9a022503ad24ca54ce83331482dfa3ff6de9f403/lib/stsci/tools/basicpar.py#L962-L988
spacetelescope/stsci.tools
lib/stsci/tools/basicpar.py
IrafArrayPar.checkValue
def checkValue(self,value,strict=0): """Check and convert a parameter value. Raises an exception if the value is not permitted for this parameter. Otherwise returns the value (converted to the right type.) """ v = self._coerceValue(value,strict) for i in range(len(v)): self.checkOneValue(v[i],strict=strict) return v
python
def checkValue(self,value,strict=0): """Check and convert a parameter value. Raises an exception if the value is not permitted for this parameter. Otherwise returns the value (converted to the right type.) """ v = self._coerceValue(value,strict) for i in range(len(v)): self.checkOneValue(v[i],strict=strict) return v
[ "def", "checkValue", "(", "self", ",", "value", ",", "strict", "=", "0", ")", ":", "v", "=", "self", ".", "_coerceValue", "(", "value", ",", "strict", ")", "for", "i", "in", "range", "(", "len", "(", "v", ")", ")", ":", "self", ".", "checkOneValue", "(", "v", "[", "i", "]", ",", "strict", "=", "strict", ")", "return", "v" ]
Check and convert a parameter value. Raises an exception if the value is not permitted for this parameter. Otherwise returns the value (converted to the right type.)
[ "Check", "and", "convert", "a", "parameter", "value", "." ]
train
https://github.com/spacetelescope/stsci.tools/blob/9a022503ad24ca54ce83331482dfa3ff6de9f403/lib/stsci/tools/basicpar.py#L990-L1000
spacetelescope/stsci.tools
lib/stsci/tools/basicpar.py
IrafArrayPar._sumindex
def _sumindex(self, index=None): """Convert tuple index to 1-D index into value""" try: ndim = len(index) except TypeError: # turn index into a 1-tuple index = (index,) ndim = 1 if len(self.shape) != ndim: raise ValueError("Index to %d-dimensional array %s has too %s dimensions" % (len(self.shape), self.name, ["many","few"][len(self.shape) > ndim])) sumindex = 0 for i in range(ndim-1,-1,-1): index1 = index[i] if index1 < 0 or index1 >= self.shape[i]: raise ValueError("Dimension %d index for array %s is out of bounds (value=%d)" % (i+1, self.name, index1)) sumindex = index1 + sumindex*self.shape[i] return sumindex
python
def _sumindex(self, index=None): """Convert tuple index to 1-D index into value""" try: ndim = len(index) except TypeError: # turn index into a 1-tuple index = (index,) ndim = 1 if len(self.shape) != ndim: raise ValueError("Index to %d-dimensional array %s has too %s dimensions" % (len(self.shape), self.name, ["many","few"][len(self.shape) > ndim])) sumindex = 0 for i in range(ndim-1,-1,-1): index1 = index[i] if index1 < 0 or index1 >= self.shape[i]: raise ValueError("Dimension %d index for array %s is out of bounds (value=%d)" % (i+1, self.name, index1)) sumindex = index1 + sumindex*self.shape[i] return sumindex
[ "def", "_sumindex", "(", "self", ",", "index", "=", "None", ")", ":", "try", ":", "ndim", "=", "len", "(", "index", ")", "except", "TypeError", ":", "# turn index into a 1-tuple", "index", "=", "(", "index", ",", ")", "ndim", "=", "1", "if", "len", "(", "self", ".", "shape", ")", "!=", "ndim", ":", "raise", "ValueError", "(", "\"Index to %d-dimensional array %s has too %s dimensions\"", "%", "(", "len", "(", "self", ".", "shape", ")", ",", "self", ".", "name", ",", "[", "\"many\"", ",", "\"few\"", "]", "[", "len", "(", "self", ".", "shape", ")", ">", "ndim", "]", ")", ")", "sumindex", "=", "0", "for", "i", "in", "range", "(", "ndim", "-", "1", ",", "-", "1", ",", "-", "1", ")", ":", "index1", "=", "index", "[", "i", "]", "if", "index1", "<", "0", "or", "index1", ">=", "self", ".", "shape", "[", "i", "]", ":", "raise", "ValueError", "(", "\"Dimension %d index for array %s is out of bounds (value=%d)\"", "%", "(", "i", "+", "1", ",", "self", ".", "name", ",", "index1", ")", ")", "sumindex", "=", "index1", "+", "sumindex", "*", "self", ".", "shape", "[", "i", "]", "return", "sumindex" ]
Convert tuple index to 1-D index into value
[ "Convert", "tuple", "index", "to", "1", "-", "D", "index", "into", "value" ]
train
https://github.com/spacetelescope/stsci.tools/blob/9a022503ad24ca54ce83331482dfa3ff6de9f403/lib/stsci/tools/basicpar.py#L1034-L1052
spacetelescope/stsci.tools
lib/stsci/tools/basicpar.py
IrafArrayPar._coerceValue
def _coerceValue(self,value,strict=0): """Coerce parameter to appropriate type Should accept None or null string. Must be an array. """ try: if isinstance(value,str): # allow single blank-separated string as input value = value.split() if len(value) != len(self.value): raise IndexError v = len(self.value)*[0] for i in range(len(v)): v[i] = self._coerceOneValue(value[i],strict) return v except (IndexError, TypeError): raise ValueError("Value must be a " + repr(len(self.value)) + "-element array for " + self.name)
python
def _coerceValue(self,value,strict=0): """Coerce parameter to appropriate type Should accept None or null string. Must be an array. """ try: if isinstance(value,str): # allow single blank-separated string as input value = value.split() if len(value) != len(self.value): raise IndexError v = len(self.value)*[0] for i in range(len(v)): v[i] = self._coerceOneValue(value[i],strict) return v except (IndexError, TypeError): raise ValueError("Value must be a " + repr(len(self.value)) + "-element array for " + self.name)
[ "def", "_coerceValue", "(", "self", ",", "value", ",", "strict", "=", "0", ")", ":", "try", ":", "if", "isinstance", "(", "value", ",", "str", ")", ":", "# allow single blank-separated string as input", "value", "=", "value", ".", "split", "(", ")", "if", "len", "(", "value", ")", "!=", "len", "(", "self", ".", "value", ")", ":", "raise", "IndexError", "v", "=", "len", "(", "self", ".", "value", ")", "*", "[", "0", "]", "for", "i", "in", "range", "(", "len", "(", "v", ")", ")", ":", "v", "[", "i", "]", "=", "self", ".", "_coerceOneValue", "(", "value", "[", "i", "]", ",", "strict", ")", "return", "v", "except", "(", "IndexError", ",", "TypeError", ")", ":", "raise", "ValueError", "(", "\"Value must be a \"", "+", "repr", "(", "len", "(", "self", ".", "value", ")", ")", "+", "\"-element array for \"", "+", "self", ".", "name", ")" ]
Coerce parameter to appropriate type Should accept None or null string. Must be an array.
[ "Coerce", "parameter", "to", "appropriate", "type" ]
train
https://github.com/spacetelescope/stsci.tools/blob/9a022503ad24ca54ce83331482dfa3ff6de9f403/lib/stsci/tools/basicpar.py#L1058-L1075
spacetelescope/stsci.tools
lib/stsci/tools/basicpar.py
_StringMixin._setChoiceDict
def _setChoiceDict(self): """Create min-match dictionary for choice list""" # value is full name of choice parameter self.choiceDict = minmatch.MinMatchDict() for c in self.choice: self.choiceDict.add(c, c)
python
def _setChoiceDict(self): """Create min-match dictionary for choice list""" # value is full name of choice parameter self.choiceDict = minmatch.MinMatchDict() for c in self.choice: self.choiceDict.add(c, c)
[ "def", "_setChoiceDict", "(", "self", ")", ":", "# value is full name of choice parameter", "self", ".", "choiceDict", "=", "minmatch", ".", "MinMatchDict", "(", ")", "for", "c", "in", "self", ".", "choice", ":", "self", ".", "choiceDict", ".", "add", "(", "c", ",", "c", ")" ]
Create min-match dictionary for choice list
[ "Create", "min", "-", "match", "dictionary", "for", "choice", "list" ]
train
https://github.com/spacetelescope/stsci.tools/blob/9a022503ad24ca54ce83331482dfa3ff6de9f403/lib/stsci/tools/basicpar.py#L1163-L1167
spacetelescope/stsci.tools
lib/stsci/tools/basicpar.py
_BooleanMixin._checkAttribs
def _checkAttribs(self, strict): """Check initial attributes to make sure they are legal""" if self.min: warning("Minimum value not allowed for boolean-type parameter " + self.name, strict) self.min = None if self.max: if not self.prompt: warning("Maximum value not allowed for boolean-type parameter " + self.name + " (probably missing comma)", strict) # try to recover by assuming max string is prompt self.prompt = self.max else: warning("Maximum value not allowed for boolean-type parameter " + self.name, strict) self.max = None if self.choice: warning("Choice values not allowed for boolean-type parameter " + self.name, strict) self.choice = None
python
def _checkAttribs(self, strict): """Check initial attributes to make sure they are legal""" if self.min: warning("Minimum value not allowed for boolean-type parameter " + self.name, strict) self.min = None if self.max: if not self.prompt: warning("Maximum value not allowed for boolean-type parameter " + self.name + " (probably missing comma)", strict) # try to recover by assuming max string is prompt self.prompt = self.max else: warning("Maximum value not allowed for boolean-type parameter " + self.name, strict) self.max = None if self.choice: warning("Choice values not allowed for boolean-type parameter " + self.name, strict) self.choice = None
[ "def", "_checkAttribs", "(", "self", ",", "strict", ")", ":", "if", "self", ".", "min", ":", "warning", "(", "\"Minimum value not allowed for boolean-type parameter \"", "+", "self", ".", "name", ",", "strict", ")", "self", ".", "min", "=", "None", "if", "self", ".", "max", ":", "if", "not", "self", ".", "prompt", ":", "warning", "(", "\"Maximum value not allowed for boolean-type parameter \"", "+", "self", ".", "name", "+", "\" (probably missing comma)\"", ",", "strict", ")", "# try to recover by assuming max string is prompt", "self", ".", "prompt", "=", "self", ".", "max", "else", ":", "warning", "(", "\"Maximum value not allowed for boolean-type parameter \"", "+", "self", ".", "name", ",", "strict", ")", "self", ".", "max", "=", "None", "if", "self", ".", "choice", ":", "warning", "(", "\"Choice values not allowed for boolean-type parameter \"", "+", "self", ".", "name", ",", "strict", ")", "self", ".", "choice", "=", "None" ]
Check initial attributes to make sure they are legal
[ "Check", "initial", "attributes", "to", "make", "sure", "they", "are", "legal" ]
train
https://github.com/spacetelescope/stsci.tools/blob/9a022503ad24ca54ce83331482dfa3ff6de9f403/lib/stsci/tools/basicpar.py#L1234-L1254
spacetelescope/stsci.tools
lib/stsci/tools/basicpar.py
_RealMixin._checkAttribs
def _checkAttribs(self, strict): """Check initial attributes to make sure they are legal""" if self.choice: warning("Choice values not allowed for real-type parameter " + self.name, strict) self.choice = None
python
def _checkAttribs(self, strict): """Check initial attributes to make sure they are legal""" if self.choice: warning("Choice values not allowed for real-type parameter " + self.name, strict) self.choice = None
[ "def", "_checkAttribs", "(", "self", ",", "strict", ")", ":", "if", "self", ".", "choice", ":", "warning", "(", "\"Choice values not allowed for real-type parameter \"", "+", "self", ".", "name", ",", "strict", ")", "self", ".", "choice", "=", "None" ]
Check initial attributes to make sure they are legal
[ "Check", "initial", "attributes", "to", "make", "sure", "they", "are", "legal" ]
train
https://github.com/spacetelescope/stsci.tools/blob/9a022503ad24ca54ce83331482dfa3ff6de9f403/lib/stsci/tools/basicpar.py#L1458-L1463
3ll3d00d/vibe
backend/src/recorder/resources/recordingdevices.py
RecordingDevice.patch
def patch(self, deviceId): """ Updates the device with the given data. Supports a json payload like { fs: newFs samplesPerBatch: samplesPerBatch gyroEnabled: true gyroSensitivity: 500 accelerometerEnabled: true accelerometerSensitivity: 2 } A heartbeat is sent on completion of the request to ensure the analyser gets a rapid update. :return: the device and 200 if the update was ok, 400 if not. """ try: device = self.recordingDevices.get(deviceId) if device.status == RecordingDeviceStatus.INITIALISED: errors = self._handlePatch(device) if len(errors) == 0: return device, 200 else: return device, 500 else: return device, 400 finally: logger.info("Sending adhoc heartbeat on device state update") self.heartbeater.sendHeartbeat()
python
def patch(self, deviceId): """ Updates the device with the given data. Supports a json payload like { fs: newFs samplesPerBatch: samplesPerBatch gyroEnabled: true gyroSensitivity: 500 accelerometerEnabled: true accelerometerSensitivity: 2 } A heartbeat is sent on completion of the request to ensure the analyser gets a rapid update. :return: the device and 200 if the update was ok, 400 if not. """ try: device = self.recordingDevices.get(deviceId) if device.status == RecordingDeviceStatus.INITIALISED: errors = self._handlePatch(device) if len(errors) == 0: return device, 200 else: return device, 500 else: return device, 400 finally: logger.info("Sending adhoc heartbeat on device state update") self.heartbeater.sendHeartbeat()
[ "def", "patch", "(", "self", ",", "deviceId", ")", ":", "try", ":", "device", "=", "self", ".", "recordingDevices", ".", "get", "(", "deviceId", ")", "if", "device", ".", "status", "==", "RecordingDeviceStatus", ".", "INITIALISED", ":", "errors", "=", "self", ".", "_handlePatch", "(", "device", ")", "if", "len", "(", "errors", ")", "==", "0", ":", "return", "device", ",", "200", "else", ":", "return", "device", ",", "500", "else", ":", "return", "device", ",", "400", "finally", ":", "logger", ".", "info", "(", "\"Sending adhoc heartbeat on device state update\"", ")", "self", ".", "heartbeater", ".", "sendHeartbeat", "(", ")" ]
Updates the device with the given data. Supports a json payload like { fs: newFs samplesPerBatch: samplesPerBatch gyroEnabled: true gyroSensitivity: 500 accelerometerEnabled: true accelerometerSensitivity: 2 } A heartbeat is sent on completion of the request to ensure the analyser gets a rapid update. :return: the device and 200 if the update was ok, 400 if not.
[ "Updates", "the", "device", "with", "the", "given", "data", ".", "Supports", "a", "json", "payload", "like", "{", "fs", ":", "newFs", "samplesPerBatch", ":", "samplesPerBatch", "gyroEnabled", ":", "true", "gyroSensitivity", ":", "500", "accelerometerEnabled", ":", "true", "accelerometerSensitivity", ":", "2", "}", "A", "heartbeat", "is", "sent", "on", "completion", "of", "the", "request", "to", "ensure", "the", "analyser", "gets", "a", "rapid", "update", ".", ":", "return", ":", "the", "device", "and", "200", "if", "the", "update", "was", "ok", "400", "if", "not", "." ]
train
https://github.com/3ll3d00d/vibe/blob/124b029f13ac746723e92cb47e9cb56edd2e54b5/backend/src/recorder/resources/recordingdevices.py#L36-L62
fitnr/convertdate
convertdate/gregorian.py
legal_date
def legal_date(year, month, day): '''Check if this is a legal date in the Gregorian calendar''' if month == 2: daysinmonth = 29 if isleap(year) else 28 else: daysinmonth = 30 if month in HAVE_30_DAYS else 31 if not (0 < day <= daysinmonth): raise ValueError("Month {} doesn't have a day {}".format(month, day)) return True
python
def legal_date(year, month, day): '''Check if this is a legal date in the Gregorian calendar''' if month == 2: daysinmonth = 29 if isleap(year) else 28 else: daysinmonth = 30 if month in HAVE_30_DAYS else 31 if not (0 < day <= daysinmonth): raise ValueError("Month {} doesn't have a day {}".format(month, day)) return True
[ "def", "legal_date", "(", "year", ",", "month", ",", "day", ")", ":", "if", "month", "==", "2", ":", "daysinmonth", "=", "29", "if", "isleap", "(", "year", ")", "else", "28", "else", ":", "daysinmonth", "=", "30", "if", "month", "in", "HAVE_30_DAYS", "else", "31", "if", "not", "(", "0", "<", "day", "<=", "daysinmonth", ")", ":", "raise", "ValueError", "(", "\"Month {} doesn't have a day {}\"", ".", "format", "(", "month", ",", "day", ")", ")", "return", "True" ]
Check if this is a legal date in the Gregorian calendar
[ "Check", "if", "this", "is", "a", "legal", "date", "in", "the", "Gregorian", "calendar" ]
train
https://github.com/fitnr/convertdate/blob/e920f168a87f99183b0aa7290d6c3af222582d43/convertdate/gregorian.py#L30-L40
fitnr/convertdate
convertdate/gregorian.py
to_jd2
def to_jd2(year, month, day): '''Gregorian to Julian Day Count for years between 1801-2099''' # http://quasar.as.utexas.edu/BillInfo/JulianDatesG.html legal_date(year, month, day) if month <= 2: year = year - 1 month = month + 12 a = floor(year / 100) b = floor(a / 4) c = 2 - a + b e = floor(365.25 * (year + 4716)) f = floor(30.6001 * (month + 1)) return c + day + e + f - 1524.5
python
def to_jd2(year, month, day): '''Gregorian to Julian Day Count for years between 1801-2099''' # http://quasar.as.utexas.edu/BillInfo/JulianDatesG.html legal_date(year, month, day) if month <= 2: year = year - 1 month = month + 12 a = floor(year / 100) b = floor(a / 4) c = 2 - a + b e = floor(365.25 * (year + 4716)) f = floor(30.6001 * (month + 1)) return c + day + e + f - 1524.5
[ "def", "to_jd2", "(", "year", ",", "month", ",", "day", ")", ":", "# http://quasar.as.utexas.edu/BillInfo/JulianDatesG.html", "legal_date", "(", "year", ",", "month", ",", "day", ")", "if", "month", "<=", "2", ":", "year", "=", "year", "-", "1", "month", "=", "month", "+", "12", "a", "=", "floor", "(", "year", "/", "100", ")", "b", "=", "floor", "(", "a", "/", "4", ")", "c", "=", "2", "-", "a", "+", "b", "e", "=", "floor", "(", "365.25", "*", "(", "year", "+", "4716", ")", ")", "f", "=", "floor", "(", "30.6001", "*", "(", "month", "+", "1", ")", ")", "return", "c", "+", "day", "+", "e", "+", "f", "-", "1524.5" ]
Gregorian to Julian Day Count for years between 1801-2099
[ "Gregorian", "to", "Julian", "Day", "Count", "for", "years", "between", "1801", "-", "2099" ]
train
https://github.com/fitnr/convertdate/blob/e920f168a87f99183b0aa7290d6c3af222582d43/convertdate/gregorian.py#L43-L58
fitnr/convertdate
convertdate/gregorian.py
from_jd
def from_jd(jd): '''Return Gregorian date in a (Y, M, D) tuple''' wjd = floor(jd - 0.5) + 0.5 depoch = wjd - EPOCH quadricent = floor(depoch / INTERCALATION_CYCLE_DAYS) dqc = depoch % INTERCALATION_CYCLE_DAYS cent = floor(dqc / LEAP_SUPPRESSION_DAYS) dcent = dqc % LEAP_SUPPRESSION_DAYS quad = floor(dcent / LEAP_CYCLE_DAYS) dquad = dcent % LEAP_CYCLE_DAYS yindex = floor(dquad / YEAR_DAYS) year = ( quadricent * INTERCALATION_CYCLE_YEARS + cent * LEAP_SUPPRESSION_YEARS + quad * LEAP_CYCLE_YEARS + yindex ) if not (cent == 4 or yindex == 4): year += 1 yearday = wjd - to_jd(year, 1, 1) leap = isleap(year) if yearday < 58 + leap: leap_adj = 0 elif leap: leap_adj = 1 else: leap_adj = 2 month = floor((((yearday + leap_adj) * 12) + 373) / 367) day = int(wjd - to_jd(year, month, 1)) + 1 return (year, month, day)
python
def from_jd(jd): '''Return Gregorian date in a (Y, M, D) tuple''' wjd = floor(jd - 0.5) + 0.5 depoch = wjd - EPOCH quadricent = floor(depoch / INTERCALATION_CYCLE_DAYS) dqc = depoch % INTERCALATION_CYCLE_DAYS cent = floor(dqc / LEAP_SUPPRESSION_DAYS) dcent = dqc % LEAP_SUPPRESSION_DAYS quad = floor(dcent / LEAP_CYCLE_DAYS) dquad = dcent % LEAP_CYCLE_DAYS yindex = floor(dquad / YEAR_DAYS) year = ( quadricent * INTERCALATION_CYCLE_YEARS + cent * LEAP_SUPPRESSION_YEARS + quad * LEAP_CYCLE_YEARS + yindex ) if not (cent == 4 or yindex == 4): year += 1 yearday = wjd - to_jd(year, 1, 1) leap = isleap(year) if yearday < 58 + leap: leap_adj = 0 elif leap: leap_adj = 1 else: leap_adj = 2 month = floor((((yearday + leap_adj) * 12) + 373) / 367) day = int(wjd - to_jd(year, month, 1)) + 1 return (year, month, day)
[ "def", "from_jd", "(", "jd", ")", ":", "wjd", "=", "floor", "(", "jd", "-", "0.5", ")", "+", "0.5", "depoch", "=", "wjd", "-", "EPOCH", "quadricent", "=", "floor", "(", "depoch", "/", "INTERCALATION_CYCLE_DAYS", ")", "dqc", "=", "depoch", "%", "INTERCALATION_CYCLE_DAYS", "cent", "=", "floor", "(", "dqc", "/", "LEAP_SUPPRESSION_DAYS", ")", "dcent", "=", "dqc", "%", "LEAP_SUPPRESSION_DAYS", "quad", "=", "floor", "(", "dcent", "/", "LEAP_CYCLE_DAYS", ")", "dquad", "=", "dcent", "%", "LEAP_CYCLE_DAYS", "yindex", "=", "floor", "(", "dquad", "/", "YEAR_DAYS", ")", "year", "=", "(", "quadricent", "*", "INTERCALATION_CYCLE_YEARS", "+", "cent", "*", "LEAP_SUPPRESSION_YEARS", "+", "quad", "*", "LEAP_CYCLE_YEARS", "+", "yindex", ")", "if", "not", "(", "cent", "==", "4", "or", "yindex", "==", "4", ")", ":", "year", "+=", "1", "yearday", "=", "wjd", "-", "to_jd", "(", "year", ",", "1", ",", "1", ")", "leap", "=", "isleap", "(", "year", ")", "if", "yearday", "<", "58", "+", "leap", ":", "leap_adj", "=", "0", "elif", "leap", ":", "leap_adj", "=", "1", "else", ":", "leap_adj", "=", "2", "month", "=", "floor", "(", "(", "(", "(", "yearday", "+", "leap_adj", ")", "*", "12", ")", "+", "373", ")", "/", "367", ")", "day", "=", "int", "(", "wjd", "-", "to_jd", "(", "year", ",", "month", ",", "1", ")", ")", "+", "1", "return", "(", "year", ",", "month", ",", "day", ")" ]
Return Gregorian date in a (Y, M, D) tuple
[ "Return", "Gregorian", "date", "in", "a", "(", "Y", "M", "D", ")", "tuple" ]
train
https://github.com/fitnr/convertdate/blob/e920f168a87f99183b0aa7290d6c3af222582d43/convertdate/gregorian.py#L80-L118
3ll3d00d/vibe
backend/src/analyser/common/targetcontroller.py
TargetController.storeFromHinge
def storeFromHinge(self, name, hingePoints): """ Stores a new item in the cache if it is allowed in. :param name: the name. :param hingePoints: the hinge points. :return: true if it is stored. """ if name not in self._cache: if self._valid(hingePoints): self._cache[name] = {'name': name, 'type': 'hinge', 'hinge': hingePoints} self.writeCache() return True return False
python
def storeFromHinge(self, name, hingePoints): """ Stores a new item in the cache if it is allowed in. :param name: the name. :param hingePoints: the hinge points. :return: true if it is stored. """ if name not in self._cache: if self._valid(hingePoints): self._cache[name] = {'name': name, 'type': 'hinge', 'hinge': hingePoints} self.writeCache() return True return False
[ "def", "storeFromHinge", "(", "self", ",", "name", ",", "hingePoints", ")", ":", "if", "name", "not", "in", "self", ".", "_cache", ":", "if", "self", ".", "_valid", "(", "hingePoints", ")", ":", "self", ".", "_cache", "[", "name", "]", "=", "{", "'name'", ":", "name", ",", "'type'", ":", "'hinge'", ",", "'hinge'", ":", "hingePoints", "}", "self", ".", "writeCache", "(", ")", "return", "True", "return", "False" ]
Stores a new item in the cache if it is allowed in. :param name: the name. :param hingePoints: the hinge points. :return: true if it is stored.
[ "Stores", "a", "new", "item", "in", "the", "cache", "if", "it", "is", "allowed", "in", ".", ":", "param", "name", ":", "the", "name", ".", ":", "param", "hingePoints", ":", "the", "hinge", "points", ".", ":", "return", ":", "true", "if", "it", "is", "stored", "." ]
train
https://github.com/3ll3d00d/vibe/blob/124b029f13ac746723e92cb47e9cb56edd2e54b5/backend/src/analyser/common/targetcontroller.py#L23-L35
3ll3d00d/vibe
backend/src/analyser/common/targetcontroller.py
TargetController.storeFromWav
def storeFromWav(self, uploadCacheEntry, start, end): """ Stores a new item in the cache. :param name: file name. :param start: start time. :param end: end time. :return: true if stored. """ prefix = uploadCacheEntry['name'] + '_' + start + '_' + end match = next((x for x in self._cache.values() if x['type'] == 'wav' and x['name'].startswith(prefix)), None) if match is None: cached = [ { 'name': prefix + '_' + n, 'analysis': n, 'start': start, 'end': end, 'type': 'wav', 'filename': uploadCacheEntry['name'] } for n in ['spectrum', 'peakSpectrum'] ] for cache in cached: self._cache[cache['name']] = cache self.writeCache() return True return False
python
def storeFromWav(self, uploadCacheEntry, start, end): """ Stores a new item in the cache. :param name: file name. :param start: start time. :param end: end time. :return: true if stored. """ prefix = uploadCacheEntry['name'] + '_' + start + '_' + end match = next((x for x in self._cache.values() if x['type'] == 'wav' and x['name'].startswith(prefix)), None) if match is None: cached = [ { 'name': prefix + '_' + n, 'analysis': n, 'start': start, 'end': end, 'type': 'wav', 'filename': uploadCacheEntry['name'] } for n in ['spectrum', 'peakSpectrum'] ] for cache in cached: self._cache[cache['name']] = cache self.writeCache() return True return False
[ "def", "storeFromWav", "(", "self", ",", "uploadCacheEntry", ",", "start", ",", "end", ")", ":", "prefix", "=", "uploadCacheEntry", "[", "'name'", "]", "+", "'_'", "+", "start", "+", "'_'", "+", "end", "match", "=", "next", "(", "(", "x", "for", "x", "in", "self", ".", "_cache", ".", "values", "(", ")", "if", "x", "[", "'type'", "]", "==", "'wav'", "and", "x", "[", "'name'", "]", ".", "startswith", "(", "prefix", ")", ")", ",", "None", ")", "if", "match", "is", "None", ":", "cached", "=", "[", "{", "'name'", ":", "prefix", "+", "'_'", "+", "n", ",", "'analysis'", ":", "n", ",", "'start'", ":", "start", ",", "'end'", ":", "end", ",", "'type'", ":", "'wav'", ",", "'filename'", ":", "uploadCacheEntry", "[", "'name'", "]", "}", "for", "n", "in", "[", "'spectrum'", ",", "'peakSpectrum'", "]", "]", "for", "cache", "in", "cached", ":", "self", ".", "_cache", "[", "cache", "[", "'name'", "]", "]", "=", "cache", "self", ".", "writeCache", "(", ")", "return", "True", "return", "False" ]
Stores a new item in the cache. :param name: file name. :param start: start time. :param end: end time. :return: true if stored.
[ "Stores", "a", "new", "item", "in", "the", "cache", ".", ":", "param", "name", ":", "file", "name", ".", ":", "param", "start", ":", "start", "time", ".", ":", "param", "end", ":", "end", "time", ".", ":", "return", ":", "true", "if", "stored", "." ]
train
https://github.com/3ll3d00d/vibe/blob/124b029f13ac746723e92cb47e9cb56edd2e54b5/backend/src/analyser/common/targetcontroller.py#L37-L62
3ll3d00d/vibe
backend/src/analyser/common/targetcontroller.py
TargetController.delete
def delete(self, name): """ Deletes the named entry in the cache. :param name: the name. :return: true if it is deleted. """ if name in self._cache: del self._cache[name] self.writeCache() # TODO clean files return True return False
python
def delete(self, name): """ Deletes the named entry in the cache. :param name: the name. :return: true if it is deleted. """ if name in self._cache: del self._cache[name] self.writeCache() # TODO clean files return True return False
[ "def", "delete", "(", "self", ",", "name", ")", ":", "if", "name", "in", "self", ".", "_cache", ":", "del", "self", ".", "_cache", "[", "name", "]", "self", ".", "writeCache", "(", ")", "# TODO clean files", "return", "True", "return", "False" ]
Deletes the named entry in the cache. :param name: the name. :return: true if it is deleted.
[ "Deletes", "the", "named", "entry", "in", "the", "cache", ".", ":", "param", "name", ":", "the", "name", ".", ":", "return", ":", "true", "if", "it", "is", "deleted", "." ]
train
https://github.com/3ll3d00d/vibe/blob/124b029f13ac746723e92cb47e9cb56edd2e54b5/backend/src/analyser/common/targetcontroller.py#L64-L75
3ll3d00d/vibe
backend/src/analyser/common/targetcontroller.py
TargetController.analyse
def analyse(self, name): """ reads the specified file. :param name: the name. :return: the analysis as frequency/Pxx. """ if name in self._cache: target = self._cache[name] if target['type'] == 'wav': signal = self._uploadController.loadSignal(target['filename'], start=target['start'] if target['start'] != 'start' else None, end=target['end'] if target['end'] != 'end' else None) if signal is not None: # TODO allow user defined window return getattr(signal, target['analysis'])(ref=1.0) else: return None, 404 pass elif target['type'] == 'hinge': hingePoints = np.array(target['hinge']).astype(np.float64) x = hingePoints[:, 1] y = hingePoints[:, 0] # extend as straight line from 0 to 500 if x[0] != 0: x = np.insert(x, 0, 0.0000001) y = np.insert(y, 0, y[0]) if x[-1] != 500: x = np.insert(x, len(x), 500.0) y = np.insert(y, len(y), y[-1]) # convert the y axis dB values into a linear value y = 10 ** (y / 10) # perform a logspace interpolation f = self.log_interp1d(x, y) # remap to 0-500 xnew = np.linspace(x[0], x[-1], num=500, endpoint=False) # and convert back to dB return xnew, 10 * np.log10(f(xnew)) else: logger.error('Unknown target type with name ' + name) return None
python
def analyse(self, name): """ reads the specified file. :param name: the name. :return: the analysis as frequency/Pxx. """ if name in self._cache: target = self._cache[name] if target['type'] == 'wav': signal = self._uploadController.loadSignal(target['filename'], start=target['start'] if target['start'] != 'start' else None, end=target['end'] if target['end'] != 'end' else None) if signal is not None: # TODO allow user defined window return getattr(signal, target['analysis'])(ref=1.0) else: return None, 404 pass elif target['type'] == 'hinge': hingePoints = np.array(target['hinge']).astype(np.float64) x = hingePoints[:, 1] y = hingePoints[:, 0] # extend as straight line from 0 to 500 if x[0] != 0: x = np.insert(x, 0, 0.0000001) y = np.insert(y, 0, y[0]) if x[-1] != 500: x = np.insert(x, len(x), 500.0) y = np.insert(y, len(y), y[-1]) # convert the y axis dB values into a linear value y = 10 ** (y / 10) # perform a logspace interpolation f = self.log_interp1d(x, y) # remap to 0-500 xnew = np.linspace(x[0], x[-1], num=500, endpoint=False) # and convert back to dB return xnew, 10 * np.log10(f(xnew)) else: logger.error('Unknown target type with name ' + name) return None
[ "def", "analyse", "(", "self", ",", "name", ")", ":", "if", "name", "in", "self", ".", "_cache", ":", "target", "=", "self", ".", "_cache", "[", "name", "]", "if", "target", "[", "'type'", "]", "==", "'wav'", ":", "signal", "=", "self", ".", "_uploadController", ".", "loadSignal", "(", "target", "[", "'filename'", "]", ",", "start", "=", "target", "[", "'start'", "]", "if", "target", "[", "'start'", "]", "!=", "'start'", "else", "None", ",", "end", "=", "target", "[", "'end'", "]", "if", "target", "[", "'end'", "]", "!=", "'end'", "else", "None", ")", "if", "signal", "is", "not", "None", ":", "# TODO allow user defined window", "return", "getattr", "(", "signal", ",", "target", "[", "'analysis'", "]", ")", "(", "ref", "=", "1.0", ")", "else", ":", "return", "None", ",", "404", "pass", "elif", "target", "[", "'type'", "]", "==", "'hinge'", ":", "hingePoints", "=", "np", ".", "array", "(", "target", "[", "'hinge'", "]", ")", ".", "astype", "(", "np", ".", "float64", ")", "x", "=", "hingePoints", "[", ":", ",", "1", "]", "y", "=", "hingePoints", "[", ":", ",", "0", "]", "# extend as straight line from 0 to 500", "if", "x", "[", "0", "]", "!=", "0", ":", "x", "=", "np", ".", "insert", "(", "x", ",", "0", ",", "0.0000001", ")", "y", "=", "np", ".", "insert", "(", "y", ",", "0", ",", "y", "[", "0", "]", ")", "if", "x", "[", "-", "1", "]", "!=", "500", ":", "x", "=", "np", ".", "insert", "(", "x", ",", "len", "(", "x", ")", ",", "500.0", ")", "y", "=", "np", ".", "insert", "(", "y", ",", "len", "(", "y", ")", ",", "y", "[", "-", "1", "]", ")", "# convert the y axis dB values into a linear value", "y", "=", "10", "**", "(", "y", "/", "10", ")", "# perform a logspace interpolation", "f", "=", "self", ".", "log_interp1d", "(", "x", ",", "y", ")", "# remap to 0-500", "xnew", "=", "np", ".", "linspace", "(", "x", "[", "0", "]", ",", "x", "[", "-", "1", "]", ",", "num", "=", "500", ",", "endpoint", "=", "False", ")", "# and convert back to dB", "return", "xnew", ",", "10", "*", "np", ".", "log10", "(", "f", "(", "xnew", ")", ")", "else", ":", "logger", ".", "error", "(", "'Unknown target type with name '", "+", "name", ")", "return", "None" ]
reads the specified file. :param name: the name. :return: the analysis as frequency/Pxx.
[ "reads", "the", "specified", "file", ".", ":", "param", "name", ":", "the", "name", ".", ":", "return", ":", "the", "analysis", "as", "frequency", "/", "Pxx", "." ]
train
https://github.com/3ll3d00d/vibe/blob/124b029f13ac746723e92cb47e9cb56edd2e54b5/backend/src/analyser/common/targetcontroller.py#L91-L130
3ll3d00d/vibe
backend/src/analyser/common/targetcontroller.py
TargetController.log_interp1d
def log_interp1d(self, xx, yy, kind='linear'): """ Performs a log space 1d interpolation. :param xx: the x values. :param yy: the y values. :param kind: the type of interpolation to apply (as per scipy interp1d) :return: the interpolation function. """ logx = np.log10(xx) logy = np.log10(yy) lin_interp = interp1d(logx, logy, kind=kind) log_interp = lambda zz: np.power(10.0, lin_interp(np.log10(zz))) return log_interp
python
def log_interp1d(self, xx, yy, kind='linear'): """ Performs a log space 1d interpolation. :param xx: the x values. :param yy: the y values. :param kind: the type of interpolation to apply (as per scipy interp1d) :return: the interpolation function. """ logx = np.log10(xx) logy = np.log10(yy) lin_interp = interp1d(logx, logy, kind=kind) log_interp = lambda zz: np.power(10.0, lin_interp(np.log10(zz))) return log_interp
[ "def", "log_interp1d", "(", "self", ",", "xx", ",", "yy", ",", "kind", "=", "'linear'", ")", ":", "logx", "=", "np", ".", "log10", "(", "xx", ")", "logy", "=", "np", ".", "log10", "(", "yy", ")", "lin_interp", "=", "interp1d", "(", "logx", ",", "logy", ",", "kind", "=", "kind", ")", "log_interp", "=", "lambda", "zz", ":", "np", ".", "power", "(", "10.0", ",", "lin_interp", "(", "np", ".", "log10", "(", "zz", ")", ")", ")", "return", "log_interp" ]
Performs a log space 1d interpolation. :param xx: the x values. :param yy: the y values. :param kind: the type of interpolation to apply (as per scipy interp1d) :return: the interpolation function.
[ "Performs", "a", "log", "space", "1d", "interpolation", ".", ":", "param", "xx", ":", "the", "x", "values", ".", ":", "param", "yy", ":", "the", "y", "values", ".", ":", "param", "kind", ":", "the", "type", "of", "interpolation", "to", "apply", "(", "as", "per", "scipy", "interp1d", ")", ":", "return", ":", "the", "interpolation", "function", "." ]
train
https://github.com/3ll3d00d/vibe/blob/124b029f13ac746723e92cb47e9cb56edd2e54b5/backend/src/analyser/common/targetcontroller.py#L140-L152
fitnr/convertdate
convertdate/coptic.py
to_jd
def to_jd(year, month, day): "Retrieve the Julian date equivalent for this date" return day + (month - 1) * 30 + (year - 1) * 365 + floor(year / 4) + EPOCH - 1
python
def to_jd(year, month, day): "Retrieve the Julian date equivalent for this date" return day + (month - 1) * 30 + (year - 1) * 365 + floor(year / 4) + EPOCH - 1
[ "def", "to_jd", "(", "year", ",", "month", ",", "day", ")", ":", "return", "day", "+", "(", "month", "-", "1", ")", "*", "30", "+", "(", "year", "-", "1", ")", "*", "365", "+", "floor", "(", "year", "/", "4", ")", "+", "EPOCH", "-", "1" ]
Retrieve the Julian date equivalent for this date
[ "Retrieve", "the", "Julian", "date", "equivalent", "for", "this", "date" ]
train
https://github.com/fitnr/convertdate/blob/e920f168a87f99183b0aa7290d6c3af222582d43/convertdate/coptic.py#L24-L26
fitnr/convertdate
convertdate/coptic.py
from_jd
def from_jd(jdc): "Create a new date from a Julian date." cdc = floor(jdc) + 0.5 - EPOCH year = floor((cdc - floor((cdc + 366) / 1461)) / 365) + 1 yday = jdc - to_jd(year, 1, 1) month = floor(yday / 30) + 1 day = yday - (month - 1) * 30 + 1 return year, month, day
python
def from_jd(jdc): "Create a new date from a Julian date." cdc = floor(jdc) + 0.5 - EPOCH year = floor((cdc - floor((cdc + 366) / 1461)) / 365) + 1 yday = jdc - to_jd(year, 1, 1) month = floor(yday / 30) + 1 day = yday - (month - 1) * 30 + 1 return year, month, day
[ "def", "from_jd", "(", "jdc", ")", ":", "cdc", "=", "floor", "(", "jdc", ")", "+", "0.5", "-", "EPOCH", "year", "=", "floor", "(", "(", "cdc", "-", "floor", "(", "(", "cdc", "+", "366", ")", "/", "1461", ")", ")", "/", "365", ")", "+", "1", "yday", "=", "jdc", "-", "to_jd", "(", "year", ",", "1", ",", "1", ")", "month", "=", "floor", "(", "yday", "/", "30", ")", "+", "1", "day", "=", "yday", "-", "(", "month", "-", "1", ")", "*", "30", "+", "1", "return", "year", ",", "month", ",", "day" ]
Create a new date from a Julian date.
[ "Create", "a", "new", "date", "from", "a", "Julian", "date", "." ]
train
https://github.com/fitnr/convertdate/blob/e920f168a87f99183b0aa7290d6c3af222582d43/convertdate/coptic.py#L29-L38
spacetelescope/stsci.tools
lib/stsci/tools/wcsutil.py
troll
def troll(roll, dec, v2, v3): """ Computes the roll angle at the target position based on:: the roll angle at the V1 axis(roll), the dec of the target(dec), and the V2/V3 position of the aperture (v2,v3) in arcseconds. Based on the algorithm provided by Colin Cox that is used in Generic Conversion at STScI. """ # Convert all angles to radians _roll = DEGTORAD(roll) _dec = DEGTORAD(dec) _v2 = DEGTORAD(v2 / 3600.) _v3 = DEGTORAD(v3 / 3600.) # compute components sin_rho = sqrt((pow(sin(_v2),2)+pow(sin(_v3),2)) - (pow(sin(_v2),2)*pow(sin(_v3),2))) rho = asin(sin_rho) beta = asin(sin(_v3)/sin_rho) if _v2 < 0: beta = pi - beta gamma = asin(sin(_v2)/sin_rho) if _v3 < 0: gamma = pi - gamma A = pi/2. + _roll - beta B = atan2( sin(A)*cos(_dec), (sin(_dec)*sin_rho - cos(_dec)*cos(rho)*cos(A))) # compute final value troll = RADTODEG(pi - (gamma+B)) return troll
python
def troll(roll, dec, v2, v3): """ Computes the roll angle at the target position based on:: the roll angle at the V1 axis(roll), the dec of the target(dec), and the V2/V3 position of the aperture (v2,v3) in arcseconds. Based on the algorithm provided by Colin Cox that is used in Generic Conversion at STScI. """ # Convert all angles to radians _roll = DEGTORAD(roll) _dec = DEGTORAD(dec) _v2 = DEGTORAD(v2 / 3600.) _v3 = DEGTORAD(v3 / 3600.) # compute components sin_rho = sqrt((pow(sin(_v2),2)+pow(sin(_v3),2)) - (pow(sin(_v2),2)*pow(sin(_v3),2))) rho = asin(sin_rho) beta = asin(sin(_v3)/sin_rho) if _v2 < 0: beta = pi - beta gamma = asin(sin(_v2)/sin_rho) if _v3 < 0: gamma = pi - gamma A = pi/2. + _roll - beta B = atan2( sin(A)*cos(_dec), (sin(_dec)*sin_rho - cos(_dec)*cos(rho)*cos(A))) # compute final value troll = RADTODEG(pi - (gamma+B)) return troll
[ "def", "troll", "(", "roll", ",", "dec", ",", "v2", ",", "v3", ")", ":", "# Convert all angles to radians", "_roll", "=", "DEGTORAD", "(", "roll", ")", "_dec", "=", "DEGTORAD", "(", "dec", ")", "_v2", "=", "DEGTORAD", "(", "v2", "/", "3600.", ")", "_v3", "=", "DEGTORAD", "(", "v3", "/", "3600.", ")", "# compute components", "sin_rho", "=", "sqrt", "(", "(", "pow", "(", "sin", "(", "_v2", ")", ",", "2", ")", "+", "pow", "(", "sin", "(", "_v3", ")", ",", "2", ")", ")", "-", "(", "pow", "(", "sin", "(", "_v2", ")", ",", "2", ")", "*", "pow", "(", "sin", "(", "_v3", ")", ",", "2", ")", ")", ")", "rho", "=", "asin", "(", "sin_rho", ")", "beta", "=", "asin", "(", "sin", "(", "_v3", ")", "/", "sin_rho", ")", "if", "_v2", "<", "0", ":", "beta", "=", "pi", "-", "beta", "gamma", "=", "asin", "(", "sin", "(", "_v2", ")", "/", "sin_rho", ")", "if", "_v3", "<", "0", ":", "gamma", "=", "pi", "-", "gamma", "A", "=", "pi", "/", "2.", "+", "_roll", "-", "beta", "B", "=", "atan2", "(", "sin", "(", "A", ")", "*", "cos", "(", "_dec", ")", ",", "(", "sin", "(", "_dec", ")", "*", "sin_rho", "-", "cos", "(", "_dec", ")", "*", "cos", "(", "rho", ")", "*", "cos", "(", "A", ")", ")", ")", "# compute final value", "troll", "=", "RADTODEG", "(", "pi", "-", "(", "gamma", "+", "B", ")", ")", "return", "troll" ]
Computes the roll angle at the target position based on:: the roll angle at the V1 axis(roll), the dec of the target(dec), and the V2/V3 position of the aperture (v2,v3) in arcseconds. Based on the algorithm provided by Colin Cox that is used in Generic Conversion at STScI.
[ "Computes", "the", "roll", "angle", "at", "the", "target", "position", "based", "on", "::" ]
train
https://github.com/spacetelescope/stsci.tools/blob/9a022503ad24ca54ce83331482dfa3ff6de9f403/lib/stsci/tools/wcsutil.py#L106-L135
spacetelescope/stsci.tools
lib/stsci/tools/wcsutil.py
WCSObject.print_archive
def print_archive(self,format=True): """ Prints out archived WCS keywords.""" if len(list(self.orig_wcs.keys())) > 0: block = 'Original WCS keywords for ' + self.rootname+ '\n' block += ' backed up on '+repr(self.orig_wcs['WCSCDATE'])+'\n' if not format: for key in self.wcstrans.keys(): block += key.upper() + " = " + repr(self.get_archivekw(key)) + '\n' block = 'PA_V3: '+repr(self.pa_obs)+'\n' else: block += 'CD_11 CD_12: '+repr(self.get_archivekw('CD1_1'))+' '+repr(self.get_archivekw('CD1_2')) +'\n' block += 'CD_21 CD_22: '+repr(self.get_archivekw('CD2_1'))+' '+repr(self.get_archivekw('CD2_2')) +'\n' block += 'CRVAL : '+repr(self.get_archivekw('CRVAL1'))+' '+repr(self.get_archivekw('CRVAL2')) + '\n' block += 'CRPIX : '+repr(self.get_archivekw('CRPIX1'))+' '+repr(self.get_archivekw('CRPIX2')) + '\n' block += 'NAXIS : '+repr(int(self.get_archivekw('NAXIS1')))+' '+repr(int(self.get_archivekw('NAXIS2'))) + '\n' block += 'Plate Scale : '+repr(self.get_archivekw('pixel scale'))+'\n' block += 'ORIENTAT : '+repr(self.get_archivekw('ORIENTAT'))+'\n' print(block)
python
def print_archive(self,format=True): """ Prints out archived WCS keywords.""" if len(list(self.orig_wcs.keys())) > 0: block = 'Original WCS keywords for ' + self.rootname+ '\n' block += ' backed up on '+repr(self.orig_wcs['WCSCDATE'])+'\n' if not format: for key in self.wcstrans.keys(): block += key.upper() + " = " + repr(self.get_archivekw(key)) + '\n' block = 'PA_V3: '+repr(self.pa_obs)+'\n' else: block += 'CD_11 CD_12: '+repr(self.get_archivekw('CD1_1'))+' '+repr(self.get_archivekw('CD1_2')) +'\n' block += 'CD_21 CD_22: '+repr(self.get_archivekw('CD2_1'))+' '+repr(self.get_archivekw('CD2_2')) +'\n' block += 'CRVAL : '+repr(self.get_archivekw('CRVAL1'))+' '+repr(self.get_archivekw('CRVAL2')) + '\n' block += 'CRPIX : '+repr(self.get_archivekw('CRPIX1'))+' '+repr(self.get_archivekw('CRPIX2')) + '\n' block += 'NAXIS : '+repr(int(self.get_archivekw('NAXIS1')))+' '+repr(int(self.get_archivekw('NAXIS2'))) + '\n' block += 'Plate Scale : '+repr(self.get_archivekw('pixel scale'))+'\n' block += 'ORIENTAT : '+repr(self.get_archivekw('ORIENTAT'))+'\n' print(block)
[ "def", "print_archive", "(", "self", ",", "format", "=", "True", ")", ":", "if", "len", "(", "list", "(", "self", ".", "orig_wcs", ".", "keys", "(", ")", ")", ")", ">", "0", ":", "block", "=", "'Original WCS keywords for '", "+", "self", ".", "rootname", "+", "'\\n'", "block", "+=", "' backed up on '", "+", "repr", "(", "self", ".", "orig_wcs", "[", "'WCSCDATE'", "]", ")", "+", "'\\n'", "if", "not", "format", ":", "for", "key", "in", "self", ".", "wcstrans", ".", "keys", "(", ")", ":", "block", "+=", "key", ".", "upper", "(", ")", "+", "\" = \"", "+", "repr", "(", "self", ".", "get_archivekw", "(", "key", ")", ")", "+", "'\\n'", "block", "=", "'PA_V3: '", "+", "repr", "(", "self", ".", "pa_obs", ")", "+", "'\\n'", "else", ":", "block", "+=", "'CD_11 CD_12: '", "+", "repr", "(", "self", ".", "get_archivekw", "(", "'CD1_1'", ")", ")", "+", "' '", "+", "repr", "(", "self", ".", "get_archivekw", "(", "'CD1_2'", ")", ")", "+", "'\\n'", "block", "+=", "'CD_21 CD_22: '", "+", "repr", "(", "self", ".", "get_archivekw", "(", "'CD2_1'", ")", ")", "+", "' '", "+", "repr", "(", "self", ".", "get_archivekw", "(", "'CD2_2'", ")", ")", "+", "'\\n'", "block", "+=", "'CRVAL : '", "+", "repr", "(", "self", ".", "get_archivekw", "(", "'CRVAL1'", ")", ")", "+", "' '", "+", "repr", "(", "self", ".", "get_archivekw", "(", "'CRVAL2'", ")", ")", "+", "'\\n'", "block", "+=", "'CRPIX : '", "+", "repr", "(", "self", ".", "get_archivekw", "(", "'CRPIX1'", ")", ")", "+", "' '", "+", "repr", "(", "self", ".", "get_archivekw", "(", "'CRPIX2'", ")", ")", "+", "'\\n'", "block", "+=", "'NAXIS : '", "+", "repr", "(", "int", "(", "self", ".", "get_archivekw", "(", "'NAXIS1'", ")", ")", ")", "+", "' '", "+", "repr", "(", "int", "(", "self", ".", "get_archivekw", "(", "'NAXIS2'", ")", ")", ")", "+", "'\\n'", "block", "+=", "'Plate Scale : '", "+", "repr", "(", "self", ".", "get_archivekw", "(", "'pixel scale'", ")", ")", "+", "'\\n'", "block", "+=", "'ORIENTAT : '", "+", "repr", "(", "self", ".", "get_archivekw", "(", "'ORIENTAT'", ")", ")", "+", "'\\n'", "print", "(", "block", ")" ]
Prints out archived WCS keywords.
[ "Prints", "out", "archived", "WCS", "keywords", "." ]
train
https://github.com/spacetelescope/stsci.tools/blob/9a022503ad24ca54ce83331482dfa3ff6de9f403/lib/stsci/tools/wcsutil.py#L420-L439
spacetelescope/stsci.tools
lib/stsci/tools/wcsutil.py
WCSObject.set_pscale
def set_pscale(self): """ Compute the pixel scale based on active WCS values. """ if self.new: self.pscale = 1.0 else: self.pscale = self.compute_pscale(self.cd11,self.cd21)
python
def set_pscale(self): """ Compute the pixel scale based on active WCS values. """ if self.new: self.pscale = 1.0 else: self.pscale = self.compute_pscale(self.cd11,self.cd21)
[ "def", "set_pscale", "(", "self", ")", ":", "if", "self", ".", "new", ":", "self", ".", "pscale", "=", "1.0", "else", ":", "self", ".", "pscale", "=", "self", ".", "compute_pscale", "(", "self", ".", "cd11", ",", "self", ".", "cd21", ")" ]
Compute the pixel scale based on active WCS values.
[ "Compute", "the", "pixel", "scale", "based", "on", "active", "WCS", "values", "." ]
train
https://github.com/spacetelescope/stsci.tools/blob/9a022503ad24ca54ce83331482dfa3ff6de9f403/lib/stsci/tools/wcsutil.py#L445-L450
spacetelescope/stsci.tools
lib/stsci/tools/wcsutil.py
WCSObject.compute_pscale
def compute_pscale(self,cd11,cd21): """ Compute the pixel scale based on active WCS values. """ return N.sqrt(N.power(cd11,2)+N.power(cd21,2)) * 3600.
python
def compute_pscale(self,cd11,cd21): """ Compute the pixel scale based on active WCS values. """ return N.sqrt(N.power(cd11,2)+N.power(cd21,2)) * 3600.
[ "def", "compute_pscale", "(", "self", ",", "cd11", ",", "cd21", ")", ":", "return", "N", ".", "sqrt", "(", "N", ".", "power", "(", "cd11", ",", "2", ")", "+", "N", ".", "power", "(", "cd21", ",", "2", ")", ")", "*", "3600." ]
Compute the pixel scale based on active WCS values.
[ "Compute", "the", "pixel", "scale", "based", "on", "active", "WCS", "values", "." ]
train
https://github.com/spacetelescope/stsci.tools/blob/9a022503ad24ca54ce83331482dfa3ff6de9f403/lib/stsci/tools/wcsutil.py#L452-L454
spacetelescope/stsci.tools
lib/stsci/tools/wcsutil.py
WCSObject.set_orient
def set_orient(self): """ Return the computed orientation based on CD matrix. """ self.orient = RADTODEG(N.arctan2(self.cd12,self.cd22))
python
def set_orient(self): """ Return the computed orientation based on CD matrix. """ self.orient = RADTODEG(N.arctan2(self.cd12,self.cd22))
[ "def", "set_orient", "(", "self", ")", ":", "self", ".", "orient", "=", "RADTODEG", "(", "N", ".", "arctan2", "(", "self", ".", "cd12", ",", "self", ".", "cd22", ")", ")" ]
Return the computed orientation based on CD matrix.
[ "Return", "the", "computed", "orientation", "based", "on", "CD", "matrix", "." ]
train
https://github.com/spacetelescope/stsci.tools/blob/9a022503ad24ca54ce83331482dfa3ff6de9f403/lib/stsci/tools/wcsutil.py#L460-L462
spacetelescope/stsci.tools
lib/stsci/tools/wcsutil.py
WCSObject.updateWCS
def updateWCS(self, pixel_scale=None, orient=None,refpos=None,refval=None,size=None): """ Create a new CD Matrix from the absolute pixel scale and reference image orientation. """ # Set up parameters necessary for updating WCS # Check to see if new value is provided, # If not, fall back on old value as the default _updateCD = no if orient is not None and orient != self.orient: pa = DEGTORAD(orient) self.orient = orient self._orient_lin = orient _updateCD = yes else: # In case only pixel_scale was specified pa = DEGTORAD(self.orient) if pixel_scale is not None and pixel_scale != self.pscale: _ratio = pixel_scale / self.pscale self.pscale = pixel_scale _updateCD = yes else: # In case, only orient was specified pixel_scale = self.pscale _ratio = None # If a new plate scale was given, # the default size should be revised accordingly # along with the default reference pixel position. # Added 31 Mar 03, WJH. if _ratio is not None: self.naxis1 /= _ratio self.naxis2 /= _ratio self.crpix1 = self.naxis1/2. self.crpix2 = self.naxis2/2. # However, if the user provides a given size, # set it to use that no matter what. if size is not None: self.naxis1 = size[0] self.naxis2 = size[1] # Insure that naxis1,2 always return as integer values. self.naxis1 = int(self.naxis1) self.naxis2 = int(self.naxis2) if refpos is not None: self.crpix1 = refpos[0] self.crpix2 = refpos[1] if self.crpix1 is None: self.crpix1 = self.naxis1/2. self.crpix2 = self.naxis2/2. if refval is not None: self.crval1 = refval[0] self.crval2 = refval[1] # Reset WCS info now... if _updateCD: # Only update this should the pscale or orientation change... pscale = pixel_scale / 3600. self.cd11 = -pscale * N.cos(pa) self.cd12 = pscale * N.sin(pa) self.cd21 = self.cd12 self.cd22 = -self.cd11 # Now make sure that all derived values are really up-to-date based # on these changes self.update()
python
def updateWCS(self, pixel_scale=None, orient=None,refpos=None,refval=None,size=None): """ Create a new CD Matrix from the absolute pixel scale and reference image orientation. """ # Set up parameters necessary for updating WCS # Check to see if new value is provided, # If not, fall back on old value as the default _updateCD = no if orient is not None and orient != self.orient: pa = DEGTORAD(orient) self.orient = orient self._orient_lin = orient _updateCD = yes else: # In case only pixel_scale was specified pa = DEGTORAD(self.orient) if pixel_scale is not None and pixel_scale != self.pscale: _ratio = pixel_scale / self.pscale self.pscale = pixel_scale _updateCD = yes else: # In case, only orient was specified pixel_scale = self.pscale _ratio = None # If a new plate scale was given, # the default size should be revised accordingly # along with the default reference pixel position. # Added 31 Mar 03, WJH. if _ratio is not None: self.naxis1 /= _ratio self.naxis2 /= _ratio self.crpix1 = self.naxis1/2. self.crpix2 = self.naxis2/2. # However, if the user provides a given size, # set it to use that no matter what. if size is not None: self.naxis1 = size[0] self.naxis2 = size[1] # Insure that naxis1,2 always return as integer values. self.naxis1 = int(self.naxis1) self.naxis2 = int(self.naxis2) if refpos is not None: self.crpix1 = refpos[0] self.crpix2 = refpos[1] if self.crpix1 is None: self.crpix1 = self.naxis1/2. self.crpix2 = self.naxis2/2. if refval is not None: self.crval1 = refval[0] self.crval2 = refval[1] # Reset WCS info now... if _updateCD: # Only update this should the pscale or orientation change... pscale = pixel_scale / 3600. self.cd11 = -pscale * N.cos(pa) self.cd12 = pscale * N.sin(pa) self.cd21 = self.cd12 self.cd22 = -self.cd11 # Now make sure that all derived values are really up-to-date based # on these changes self.update()
[ "def", "updateWCS", "(", "self", ",", "pixel_scale", "=", "None", ",", "orient", "=", "None", ",", "refpos", "=", "None", ",", "refval", "=", "None", ",", "size", "=", "None", ")", ":", "# Set up parameters necessary for updating WCS", "# Check to see if new value is provided,", "# If not, fall back on old value as the default", "_updateCD", "=", "no", "if", "orient", "is", "not", "None", "and", "orient", "!=", "self", ".", "orient", ":", "pa", "=", "DEGTORAD", "(", "orient", ")", "self", ".", "orient", "=", "orient", "self", ".", "_orient_lin", "=", "orient", "_updateCD", "=", "yes", "else", ":", "# In case only pixel_scale was specified", "pa", "=", "DEGTORAD", "(", "self", ".", "orient", ")", "if", "pixel_scale", "is", "not", "None", "and", "pixel_scale", "!=", "self", ".", "pscale", ":", "_ratio", "=", "pixel_scale", "/", "self", ".", "pscale", "self", ".", "pscale", "=", "pixel_scale", "_updateCD", "=", "yes", "else", ":", "# In case, only orient was specified", "pixel_scale", "=", "self", ".", "pscale", "_ratio", "=", "None", "# If a new plate scale was given,", "# the default size should be revised accordingly", "# along with the default reference pixel position.", "# Added 31 Mar 03, WJH.", "if", "_ratio", "is", "not", "None", ":", "self", ".", "naxis1", "/=", "_ratio", "self", ".", "naxis2", "/=", "_ratio", "self", ".", "crpix1", "=", "self", ".", "naxis1", "/", "2.", "self", ".", "crpix2", "=", "self", ".", "naxis2", "/", "2.", "# However, if the user provides a given size,", "# set it to use that no matter what.", "if", "size", "is", "not", "None", ":", "self", ".", "naxis1", "=", "size", "[", "0", "]", "self", ".", "naxis2", "=", "size", "[", "1", "]", "# Insure that naxis1,2 always return as integer values.", "self", ".", "naxis1", "=", "int", "(", "self", ".", "naxis1", ")", "self", ".", "naxis2", "=", "int", "(", "self", ".", "naxis2", ")", "if", "refpos", "is", "not", "None", ":", "self", ".", "crpix1", "=", "refpos", "[", "0", "]", "self", ".", "crpix2", "=", "refpos", "[", "1", "]", "if", "self", ".", "crpix1", "is", "None", ":", "self", ".", "crpix1", "=", "self", ".", "naxis1", "/", "2.", "self", ".", "crpix2", "=", "self", ".", "naxis2", "/", "2.", "if", "refval", "is", "not", "None", ":", "self", ".", "crval1", "=", "refval", "[", "0", "]", "self", ".", "crval2", "=", "refval", "[", "1", "]", "# Reset WCS info now...", "if", "_updateCD", ":", "# Only update this should the pscale or orientation change...", "pscale", "=", "pixel_scale", "/", "3600.", "self", ".", "cd11", "=", "-", "pscale", "*", "N", ".", "cos", "(", "pa", ")", "self", ".", "cd12", "=", "pscale", "*", "N", ".", "sin", "(", "pa", ")", "self", ".", "cd21", "=", "self", ".", "cd12", "self", ".", "cd22", "=", "-", "self", ".", "cd11", "# Now make sure that all derived values are really up-to-date based", "# on these changes", "self", ".", "update", "(", ")" ]
Create a new CD Matrix from the absolute pixel scale and reference image orientation.
[ "Create", "a", "new", "CD", "Matrix", "from", "the", "absolute", "pixel", "scale", "and", "reference", "image", "orientation", "." ]
train
https://github.com/spacetelescope/stsci.tools/blob/9a022503ad24ca54ce83331482dfa3ff6de9f403/lib/stsci/tools/wcsutil.py#L469-L540
spacetelescope/stsci.tools
lib/stsci/tools/wcsutil.py
WCSObject.scale_WCS
def scale_WCS(self,pixel_scale,retain=True): ''' Scale the WCS to a new pixel_scale. The 'retain' parameter [default value: True] controls whether or not to retain the original distortion solution in the CD matrix. ''' _ratio = pixel_scale / self.pscale # Correct the size of the image and CRPIX values for scaled WCS self.naxis1 /= _ratio self.naxis2 /= _ratio self.crpix1 = self.naxis1/2. self.crpix2 = self.naxis2/2. if retain: # Correct the WCS while retaining original distortion information self.cd11 *= _ratio self.cd12 *= _ratio self.cd21 *= _ratio self.cd22 *= _ratio else: pscale = pixel_scale / 3600. self.cd11 = -pscale * N.cos(pa) self.cd12 = pscale * N.sin(pa) self.cd21 = self.cd12 self.cd22 = -self.cd11 # Now make sure that all derived values are really up-to-date based # on these changes self.update()
python
def scale_WCS(self,pixel_scale,retain=True): ''' Scale the WCS to a new pixel_scale. The 'retain' parameter [default value: True] controls whether or not to retain the original distortion solution in the CD matrix. ''' _ratio = pixel_scale / self.pscale # Correct the size of the image and CRPIX values for scaled WCS self.naxis1 /= _ratio self.naxis2 /= _ratio self.crpix1 = self.naxis1/2. self.crpix2 = self.naxis2/2. if retain: # Correct the WCS while retaining original distortion information self.cd11 *= _ratio self.cd12 *= _ratio self.cd21 *= _ratio self.cd22 *= _ratio else: pscale = pixel_scale / 3600. self.cd11 = -pscale * N.cos(pa) self.cd12 = pscale * N.sin(pa) self.cd21 = self.cd12 self.cd22 = -self.cd11 # Now make sure that all derived values are really up-to-date based # on these changes self.update()
[ "def", "scale_WCS", "(", "self", ",", "pixel_scale", ",", "retain", "=", "True", ")", ":", "_ratio", "=", "pixel_scale", "/", "self", ".", "pscale", "# Correct the size of the image and CRPIX values for scaled WCS", "self", ".", "naxis1", "/=", "_ratio", "self", ".", "naxis2", "/=", "_ratio", "self", ".", "crpix1", "=", "self", ".", "naxis1", "/", "2.", "self", ".", "crpix2", "=", "self", ".", "naxis2", "/", "2.", "if", "retain", ":", "# Correct the WCS while retaining original distortion information", "self", ".", "cd11", "*=", "_ratio", "self", ".", "cd12", "*=", "_ratio", "self", ".", "cd21", "*=", "_ratio", "self", ".", "cd22", "*=", "_ratio", "else", ":", "pscale", "=", "pixel_scale", "/", "3600.", "self", ".", "cd11", "=", "-", "pscale", "*", "N", ".", "cos", "(", "pa", ")", "self", ".", "cd12", "=", "pscale", "*", "N", ".", "sin", "(", "pa", ")", "self", ".", "cd21", "=", "self", ".", "cd12", "self", ".", "cd22", "=", "-", "self", ".", "cd11", "# Now make sure that all derived values are really up-to-date based", "# on these changes", "self", ".", "update", "(", ")" ]
Scale the WCS to a new pixel_scale. The 'retain' parameter [default value: True] controls whether or not to retain the original distortion solution in the CD matrix.
[ "Scale", "the", "WCS", "to", "a", "new", "pixel_scale", ".", "The", "retain", "parameter", "[", "default", "value", ":", "True", "]", "controls", "whether", "or", "not", "to", "retain", "the", "original", "distortion", "solution", "in", "the", "CD", "matrix", "." ]
train
https://github.com/spacetelescope/stsci.tools/blob/9a022503ad24ca54ce83331482dfa3ff6de9f403/lib/stsci/tools/wcsutil.py#L542-L570
spacetelescope/stsci.tools
lib/stsci/tools/wcsutil.py
WCSObject.xy2rd
def xy2rd(self,pos): """ This method would apply the WCS keywords to a position to generate a new sky position. The algorithm comes directly from 'imgtools.xy2rd' translate (x,y) to (ra, dec) """ if self.ctype1.find('TAN') < 0 or self.ctype2.find('TAN') < 0: print('XY2RD only supported for TAN projections.') raise TypeError if isinstance(pos,N.ndarray): # If we are working with an array of positions, # point to just X and Y values posx = pos[:,0] posy = pos[:,1] else: # Otherwise, we are working with a single X,Y tuple posx = pos[0] posy = pos[1] xi = self.cd11 * (posx - self.crpix1) + self.cd12 * (posy - self.crpix2) eta = self.cd21 * (posx - self.crpix1) + self.cd22 * (posy - self.crpix2) xi = DEGTORAD(xi) eta = DEGTORAD(eta) ra0 = DEGTORAD(self.crval1) dec0 = DEGTORAD(self.crval2) ra = N.arctan((xi / (N.cos(dec0)-eta*N.sin(dec0)))) + ra0 dec = N.arctan( ((eta*N.cos(dec0)+N.sin(dec0)) / (N.sqrt((N.cos(dec0)-eta*N.sin(dec0))**2 + xi**2))) ) ra = RADTODEG(ra) dec = RADTODEG(dec) ra = DIVMOD(ra, 360.) # Otherwise, just return the RA,Dec tuple. return ra,dec
python
def xy2rd(self,pos): """ This method would apply the WCS keywords to a position to generate a new sky position. The algorithm comes directly from 'imgtools.xy2rd' translate (x,y) to (ra, dec) """ if self.ctype1.find('TAN') < 0 or self.ctype2.find('TAN') < 0: print('XY2RD only supported for TAN projections.') raise TypeError if isinstance(pos,N.ndarray): # If we are working with an array of positions, # point to just X and Y values posx = pos[:,0] posy = pos[:,1] else: # Otherwise, we are working with a single X,Y tuple posx = pos[0] posy = pos[1] xi = self.cd11 * (posx - self.crpix1) + self.cd12 * (posy - self.crpix2) eta = self.cd21 * (posx - self.crpix1) + self.cd22 * (posy - self.crpix2) xi = DEGTORAD(xi) eta = DEGTORAD(eta) ra0 = DEGTORAD(self.crval1) dec0 = DEGTORAD(self.crval2) ra = N.arctan((xi / (N.cos(dec0)-eta*N.sin(dec0)))) + ra0 dec = N.arctan( ((eta*N.cos(dec0)+N.sin(dec0)) / (N.sqrt((N.cos(dec0)-eta*N.sin(dec0))**2 + xi**2))) ) ra = RADTODEG(ra) dec = RADTODEG(dec) ra = DIVMOD(ra, 360.) # Otherwise, just return the RA,Dec tuple. return ra,dec
[ "def", "xy2rd", "(", "self", ",", "pos", ")", ":", "if", "self", ".", "ctype1", ".", "find", "(", "'TAN'", ")", "<", "0", "or", "self", ".", "ctype2", ".", "find", "(", "'TAN'", ")", "<", "0", ":", "print", "(", "'XY2RD only supported for TAN projections.'", ")", "raise", "TypeError", "if", "isinstance", "(", "pos", ",", "N", ".", "ndarray", ")", ":", "# If we are working with an array of positions,", "# point to just X and Y values", "posx", "=", "pos", "[", ":", ",", "0", "]", "posy", "=", "pos", "[", ":", ",", "1", "]", "else", ":", "# Otherwise, we are working with a single X,Y tuple", "posx", "=", "pos", "[", "0", "]", "posy", "=", "pos", "[", "1", "]", "xi", "=", "self", ".", "cd11", "*", "(", "posx", "-", "self", ".", "crpix1", ")", "+", "self", ".", "cd12", "*", "(", "posy", "-", "self", ".", "crpix2", ")", "eta", "=", "self", ".", "cd21", "*", "(", "posx", "-", "self", ".", "crpix1", ")", "+", "self", ".", "cd22", "*", "(", "posy", "-", "self", ".", "crpix2", ")", "xi", "=", "DEGTORAD", "(", "xi", ")", "eta", "=", "DEGTORAD", "(", "eta", ")", "ra0", "=", "DEGTORAD", "(", "self", ".", "crval1", ")", "dec0", "=", "DEGTORAD", "(", "self", ".", "crval2", ")", "ra", "=", "N", ".", "arctan", "(", "(", "xi", "/", "(", "N", ".", "cos", "(", "dec0", ")", "-", "eta", "*", "N", ".", "sin", "(", "dec0", ")", ")", ")", ")", "+", "ra0", "dec", "=", "N", ".", "arctan", "(", "(", "(", "eta", "*", "N", ".", "cos", "(", "dec0", ")", "+", "N", ".", "sin", "(", "dec0", ")", ")", "/", "(", "N", ".", "sqrt", "(", "(", "N", ".", "cos", "(", "dec0", ")", "-", "eta", "*", "N", ".", "sin", "(", "dec0", ")", ")", "**", "2", "+", "xi", "**", "2", ")", ")", ")", ")", "ra", "=", "RADTODEG", "(", "ra", ")", "dec", "=", "RADTODEG", "(", "dec", ")", "ra", "=", "DIVMOD", "(", "ra", ",", "360.", ")", "# Otherwise, just return the RA,Dec tuple.", "return", "ra", ",", "dec" ]
This method would apply the WCS keywords to a position to generate a new sky position. The algorithm comes directly from 'imgtools.xy2rd' translate (x,y) to (ra, dec)
[ "This", "method", "would", "apply", "the", "WCS", "keywords", "to", "a", "position", "to", "generate", "a", "new", "sky", "position", "." ]
train
https://github.com/spacetelescope/stsci.tools/blob/9a022503ad24ca54ce83331482dfa3ff6de9f403/lib/stsci/tools/wcsutil.py#L572-L612
spacetelescope/stsci.tools
lib/stsci/tools/wcsutil.py
WCSObject.rd2xy
def rd2xy(self,skypos,hour=no): """ This method would use the WCS keywords to compute the XY position from a given RA/Dec tuple (in deg). NOTE: Investigate how to let this function accept arrays as well as single positions. WJH 27Mar03 """ if self.ctype1.find('TAN') < 0 or self.ctype2.find('TAN') < 0: print('RD2XY only supported for TAN projections.') raise TypeError det = self.cd11*self.cd22 - self.cd12*self.cd21 if det == 0.0: raise ArithmeticError("singular CD matrix!") cdinv11 = self.cd22 / det cdinv12 = -self.cd12 / det cdinv21 = -self.cd21 / det cdinv22 = self.cd11 / det # translate (ra, dec) to (x, y) ra0 = DEGTORAD(self.crval1) dec0 = DEGTORAD(self.crval2) if hour: skypos[0] = skypos[0] * 15. ra = DEGTORAD(skypos[0]) dec = DEGTORAD(skypos[1]) bottom = float(N.sin(dec)*N.sin(dec0) + N.cos(dec)*N.cos(dec0)*N.cos(ra-ra0)) if bottom == 0.0: raise ArithmeticError("Unreasonable RA/Dec range!") xi = RADTODEG((N.cos(dec) * N.sin(ra-ra0) / bottom)) eta = RADTODEG((N.sin(dec)*N.cos(dec0) - N.cos(dec)*N.sin(dec0)*N.cos(ra-ra0)) / bottom) x = cdinv11 * xi + cdinv12 * eta + self.crpix1 y = cdinv21 * xi + cdinv22 * eta + self.crpix2 return x,y
python
def rd2xy(self,skypos,hour=no): """ This method would use the WCS keywords to compute the XY position from a given RA/Dec tuple (in deg). NOTE: Investigate how to let this function accept arrays as well as single positions. WJH 27Mar03 """ if self.ctype1.find('TAN') < 0 or self.ctype2.find('TAN') < 0: print('RD2XY only supported for TAN projections.') raise TypeError det = self.cd11*self.cd22 - self.cd12*self.cd21 if det == 0.0: raise ArithmeticError("singular CD matrix!") cdinv11 = self.cd22 / det cdinv12 = -self.cd12 / det cdinv21 = -self.cd21 / det cdinv22 = self.cd11 / det # translate (ra, dec) to (x, y) ra0 = DEGTORAD(self.crval1) dec0 = DEGTORAD(self.crval2) if hour: skypos[0] = skypos[0] * 15. ra = DEGTORAD(skypos[0]) dec = DEGTORAD(skypos[1]) bottom = float(N.sin(dec)*N.sin(dec0) + N.cos(dec)*N.cos(dec0)*N.cos(ra-ra0)) if bottom == 0.0: raise ArithmeticError("Unreasonable RA/Dec range!") xi = RADTODEG((N.cos(dec) * N.sin(ra-ra0) / bottom)) eta = RADTODEG((N.sin(dec)*N.cos(dec0) - N.cos(dec)*N.sin(dec0)*N.cos(ra-ra0)) / bottom) x = cdinv11 * xi + cdinv12 * eta + self.crpix1 y = cdinv21 * xi + cdinv22 * eta + self.crpix2 return x,y
[ "def", "rd2xy", "(", "self", ",", "skypos", ",", "hour", "=", "no", ")", ":", "if", "self", ".", "ctype1", ".", "find", "(", "'TAN'", ")", "<", "0", "or", "self", ".", "ctype2", ".", "find", "(", "'TAN'", ")", "<", "0", ":", "print", "(", "'RD2XY only supported for TAN projections.'", ")", "raise", "TypeError", "det", "=", "self", ".", "cd11", "*", "self", ".", "cd22", "-", "self", ".", "cd12", "*", "self", ".", "cd21", "if", "det", "==", "0.0", ":", "raise", "ArithmeticError", "(", "\"singular CD matrix!\"", ")", "cdinv11", "=", "self", ".", "cd22", "/", "det", "cdinv12", "=", "-", "self", ".", "cd12", "/", "det", "cdinv21", "=", "-", "self", ".", "cd21", "/", "det", "cdinv22", "=", "self", ".", "cd11", "/", "det", "# translate (ra, dec) to (x, y)", "ra0", "=", "DEGTORAD", "(", "self", ".", "crval1", ")", "dec0", "=", "DEGTORAD", "(", "self", ".", "crval2", ")", "if", "hour", ":", "skypos", "[", "0", "]", "=", "skypos", "[", "0", "]", "*", "15.", "ra", "=", "DEGTORAD", "(", "skypos", "[", "0", "]", ")", "dec", "=", "DEGTORAD", "(", "skypos", "[", "1", "]", ")", "bottom", "=", "float", "(", "N", ".", "sin", "(", "dec", ")", "*", "N", ".", "sin", "(", "dec0", ")", "+", "N", ".", "cos", "(", "dec", ")", "*", "N", ".", "cos", "(", "dec0", ")", "*", "N", ".", "cos", "(", "ra", "-", "ra0", ")", ")", "if", "bottom", "==", "0.0", ":", "raise", "ArithmeticError", "(", "\"Unreasonable RA/Dec range!\"", ")", "xi", "=", "RADTODEG", "(", "(", "N", ".", "cos", "(", "dec", ")", "*", "N", ".", "sin", "(", "ra", "-", "ra0", ")", "/", "bottom", ")", ")", "eta", "=", "RADTODEG", "(", "(", "N", ".", "sin", "(", "dec", ")", "*", "N", ".", "cos", "(", "dec0", ")", "-", "N", ".", "cos", "(", "dec", ")", "*", "N", ".", "sin", "(", "dec0", ")", "*", "N", ".", "cos", "(", "ra", "-", "ra0", ")", ")", "/", "bottom", ")", "x", "=", "cdinv11", "*", "xi", "+", "cdinv12", "*", "eta", "+", "self", ".", "crpix1", "y", "=", "cdinv21", "*", "xi", "+", "cdinv22", "*", "eta", "+", "self", ".", "crpix2", "return", "x", ",", "y" ]
This method would use the WCS keywords to compute the XY position from a given RA/Dec tuple (in deg). NOTE: Investigate how to let this function accept arrays as well as single positions. WJH 27Mar03
[ "This", "method", "would", "use", "the", "WCS", "keywords", "to", "compute", "the", "XY", "position", "from", "a", "given", "RA", "/", "Dec", "tuple", "(", "in", "deg", ")", "." ]
train
https://github.com/spacetelescope/stsci.tools/blob/9a022503ad24ca54ce83331482dfa3ff6de9f403/lib/stsci/tools/wcsutil.py#L615-L657
spacetelescope/stsci.tools
lib/stsci/tools/wcsutil.py
WCSObject.rotateCD
def rotateCD(self,orient): """ Rotates WCS CD matrix to new orientation given by 'orient' """ # Determine where member CRVAL position falls in ref frame # Find out whether this needs to be rotated to align with # reference frame. _delta = self.get_orient() - orient if _delta == 0.: return # Start by building the rotation matrix... _rot = fileutil.buildRotMatrix(_delta) # ...then, rotate the CD matrix and update the values... _cd = N.array([[self.cd11,self.cd12],[self.cd21,self.cd22]],dtype=N.float64) _cdrot = N.dot(_cd,_rot) self.cd11 = _cdrot[0][0] self.cd12 = _cdrot[0][1] self.cd21 = _cdrot[1][0] self.cd22 = _cdrot[1][1] self.orient = orient
python
def rotateCD(self,orient): """ Rotates WCS CD matrix to new orientation given by 'orient' """ # Determine where member CRVAL position falls in ref frame # Find out whether this needs to be rotated to align with # reference frame. _delta = self.get_orient() - orient if _delta == 0.: return # Start by building the rotation matrix... _rot = fileutil.buildRotMatrix(_delta) # ...then, rotate the CD matrix and update the values... _cd = N.array([[self.cd11,self.cd12],[self.cd21,self.cd22]],dtype=N.float64) _cdrot = N.dot(_cd,_rot) self.cd11 = _cdrot[0][0] self.cd12 = _cdrot[0][1] self.cd21 = _cdrot[1][0] self.cd22 = _cdrot[1][1] self.orient = orient
[ "def", "rotateCD", "(", "self", ",", "orient", ")", ":", "# Determine where member CRVAL position falls in ref frame", "# Find out whether this needs to be rotated to align with", "# reference frame.", "_delta", "=", "self", ".", "get_orient", "(", ")", "-", "orient", "if", "_delta", "==", "0.", ":", "return", "# Start by building the rotation matrix...", "_rot", "=", "fileutil", ".", "buildRotMatrix", "(", "_delta", ")", "# ...then, rotate the CD matrix and update the values...", "_cd", "=", "N", ".", "array", "(", "[", "[", "self", ".", "cd11", ",", "self", ".", "cd12", "]", ",", "[", "self", ".", "cd21", ",", "self", ".", "cd22", "]", "]", ",", "dtype", "=", "N", ".", "float64", ")", "_cdrot", "=", "N", ".", "dot", "(", "_cd", ",", "_rot", ")", "self", ".", "cd11", "=", "_cdrot", "[", "0", "]", "[", "0", "]", "self", ".", "cd12", "=", "_cdrot", "[", "0", "]", "[", "1", "]", "self", ".", "cd21", "=", "_cdrot", "[", "1", "]", "[", "0", "]", "self", ".", "cd22", "=", "_cdrot", "[", "1", "]", "[", "1", "]", "self", ".", "orient", "=", "orient" ]
Rotates WCS CD matrix to new orientation given by 'orient'
[ "Rotates", "WCS", "CD", "matrix", "to", "new", "orientation", "given", "by", "orient" ]
train
https://github.com/spacetelescope/stsci.tools/blob/9a022503ad24ca54ce83331482dfa3ff6de9f403/lib/stsci/tools/wcsutil.py#L659-L679
spacetelescope/stsci.tools
lib/stsci/tools/wcsutil.py
WCSObject.write
def write(self,fitsname=None,wcs=None,archive=True,overwrite=False,quiet=True): """ Write out the values of the WCS keywords to the specified image. If it is a GEIS image and 'fitsname' has been provided, it will automatically make a multi-extension FITS copy of the GEIS and update that file. Otherwise, it throw an Exception if the user attempts to directly update a GEIS image header. If archive=True, also write out archived WCS keyword values to file. If overwrite=True, replace archived WCS values in file with new values. If a WCSObject is passed through the 'wcs' keyword, then the WCS keywords of this object are copied to the header of the image to be updated. A use case fo rthis is updating the WCS of a WFPC2 data quality (_c1h.fits) file in order to be in sync with the science (_c0h.fits) file. """ ## Start by making sure all derived values are in sync with CD matrix self.update() image = self.rootname _fitsname = fitsname if image.find('.fits') < 0 and _fitsname is not None: # A non-FITS image was provided, and openImage made a copy # Update attributes to point to new copy instead self.geisname = image image = self.rootname = _fitsname # Open image as writable FITS object fimg = fileutil.openImage(image, mode='update', fitsname=_fitsname) _root,_iextn = fileutil.parseFilename(image) _extn = fileutil.getExtn(fimg,_iextn) # Write out values to header... if wcs: _wcsobj = wcs else: _wcsobj = self for key in _wcsobj.wcstrans.keys(): _dkey = _wcsobj.wcstrans[key] if _dkey != 'pscale': _extn.header[key] = _wcsobj.__dict__[_dkey] # Close the file fimg.close() del fimg if archive: self.write_archive(fitsname=fitsname,overwrite=overwrite,quiet=quiet)
python
def write(self,fitsname=None,wcs=None,archive=True,overwrite=False,quiet=True): """ Write out the values of the WCS keywords to the specified image. If it is a GEIS image and 'fitsname' has been provided, it will automatically make a multi-extension FITS copy of the GEIS and update that file. Otherwise, it throw an Exception if the user attempts to directly update a GEIS image header. If archive=True, also write out archived WCS keyword values to file. If overwrite=True, replace archived WCS values in file with new values. If a WCSObject is passed through the 'wcs' keyword, then the WCS keywords of this object are copied to the header of the image to be updated. A use case fo rthis is updating the WCS of a WFPC2 data quality (_c1h.fits) file in order to be in sync with the science (_c0h.fits) file. """ ## Start by making sure all derived values are in sync with CD matrix self.update() image = self.rootname _fitsname = fitsname if image.find('.fits') < 0 and _fitsname is not None: # A non-FITS image was provided, and openImage made a copy # Update attributes to point to new copy instead self.geisname = image image = self.rootname = _fitsname # Open image as writable FITS object fimg = fileutil.openImage(image, mode='update', fitsname=_fitsname) _root,_iextn = fileutil.parseFilename(image) _extn = fileutil.getExtn(fimg,_iextn) # Write out values to header... if wcs: _wcsobj = wcs else: _wcsobj = self for key in _wcsobj.wcstrans.keys(): _dkey = _wcsobj.wcstrans[key] if _dkey != 'pscale': _extn.header[key] = _wcsobj.__dict__[_dkey] # Close the file fimg.close() del fimg if archive: self.write_archive(fitsname=fitsname,overwrite=overwrite,quiet=quiet)
[ "def", "write", "(", "self", ",", "fitsname", "=", "None", ",", "wcs", "=", "None", ",", "archive", "=", "True", ",", "overwrite", "=", "False", ",", "quiet", "=", "True", ")", ":", "## Start by making sure all derived values are in sync with CD matrix", "self", ".", "update", "(", ")", "image", "=", "self", ".", "rootname", "_fitsname", "=", "fitsname", "if", "image", ".", "find", "(", "'.fits'", ")", "<", "0", "and", "_fitsname", "is", "not", "None", ":", "# A non-FITS image was provided, and openImage made a copy", "# Update attributes to point to new copy instead", "self", ".", "geisname", "=", "image", "image", "=", "self", ".", "rootname", "=", "_fitsname", "# Open image as writable FITS object", "fimg", "=", "fileutil", ".", "openImage", "(", "image", ",", "mode", "=", "'update'", ",", "fitsname", "=", "_fitsname", ")", "_root", ",", "_iextn", "=", "fileutil", ".", "parseFilename", "(", "image", ")", "_extn", "=", "fileutil", ".", "getExtn", "(", "fimg", ",", "_iextn", ")", "# Write out values to header...", "if", "wcs", ":", "_wcsobj", "=", "wcs", "else", ":", "_wcsobj", "=", "self", "for", "key", "in", "_wcsobj", ".", "wcstrans", ".", "keys", "(", ")", ":", "_dkey", "=", "_wcsobj", ".", "wcstrans", "[", "key", "]", "if", "_dkey", "!=", "'pscale'", ":", "_extn", ".", "header", "[", "key", "]", "=", "_wcsobj", ".", "__dict__", "[", "_dkey", "]", "# Close the file", "fimg", ".", "close", "(", ")", "del", "fimg", "if", "archive", ":", "self", ".", "write_archive", "(", "fitsname", "=", "fitsname", ",", "overwrite", "=", "overwrite", ",", "quiet", "=", "quiet", ")" ]
Write out the values of the WCS keywords to the specified image. If it is a GEIS image and 'fitsname' has been provided, it will automatically make a multi-extension FITS copy of the GEIS and update that file. Otherwise, it throw an Exception if the user attempts to directly update a GEIS image header. If archive=True, also write out archived WCS keyword values to file. If overwrite=True, replace archived WCS values in file with new values. If a WCSObject is passed through the 'wcs' keyword, then the WCS keywords of this object are copied to the header of the image to be updated. A use case fo rthis is updating the WCS of a WFPC2 data quality (_c1h.fits) file in order to be in sync with the science (_c0h.fits) file.
[ "Write", "out", "the", "values", "of", "the", "WCS", "keywords", "to", "the", "specified", "image", "." ]
train
https://github.com/spacetelescope/stsci.tools/blob/9a022503ad24ca54ce83331482dfa3ff6de9f403/lib/stsci/tools/wcsutil.py#L756-L809
spacetelescope/stsci.tools
lib/stsci/tools/wcsutil.py
WCSObject.restore
def restore(self): """ Reset the active WCS keywords to values stored in the backup keywords. """ # If there are no backup keys, do nothing... if len(list(self.backup.keys())) == 0: return for key in self.backup.keys(): if key != 'WCSCDATE': self.__dict__[self.wcstrans[key]] = self.orig_wcs[self.backup[key]] self.update()
python
def restore(self): """ Reset the active WCS keywords to values stored in the backup keywords. """ # If there are no backup keys, do nothing... if len(list(self.backup.keys())) == 0: return for key in self.backup.keys(): if key != 'WCSCDATE': self.__dict__[self.wcstrans[key]] = self.orig_wcs[self.backup[key]] self.update()
[ "def", "restore", "(", "self", ")", ":", "# If there are no backup keys, do nothing...", "if", "len", "(", "list", "(", "self", ".", "backup", ".", "keys", "(", ")", ")", ")", "==", "0", ":", "return", "for", "key", "in", "self", ".", "backup", ".", "keys", "(", ")", ":", "if", "key", "!=", "'WCSCDATE'", ":", "self", ".", "__dict__", "[", "self", ".", "wcstrans", "[", "key", "]", "]", "=", "self", ".", "orig_wcs", "[", "self", ".", "backup", "[", "key", "]", "]", "self", ".", "update", "(", ")" ]
Reset the active WCS keywords to values stored in the backup keywords.
[ "Reset", "the", "active", "WCS", "keywords", "to", "values", "stored", "in", "the", "backup", "keywords", "." ]
train
https://github.com/spacetelescope/stsci.tools/blob/9a022503ad24ca54ce83331482dfa3ff6de9f403/lib/stsci/tools/wcsutil.py#L811-L822
spacetelescope/stsci.tools
lib/stsci/tools/wcsutil.py
WCSObject.archive
def archive(self,prepend=None,overwrite=no,quiet=yes): """ Create backup copies of the WCS keywords with the given prepended string. If backup keywords are already present, only update them if 'overwrite' is set to 'yes', otherwise, do warn the user and do nothing. Set the WCSDATE at this time as well. """ # Verify that existing backup values are not overwritten accidentally. if len(list(self.backup.keys())) > 0 and overwrite == no: if not quiet: print('WARNING: Backup WCS keywords already exist! No backup made.') print(' The values can only be overridden if overwrite=yes.') return # Establish what prepend string to use... if prepend is None: if self.prepend is not None: _prefix = self.prepend else: _prefix = DEFAULT_PREFIX else: _prefix = prepend # Update backup and orig_wcs dictionaries # We have archive keywords and a defined prefix # Go through and append them to self.backup self.prepend = _prefix for key in self.wcstrans.keys(): if key != 'pixel scale': _archive_key = self._buildNewKeyname(key,_prefix) else: _archive_key = self.prepend.lower()+'pscale' # if key != 'pixel scale': self.orig_wcs[_archive_key] = self.__dict__[self.wcstrans[key]] self.backup[key] = _archive_key self.revert[_archive_key] = key # Setup keyword to record when these keywords were backed up. self.orig_wcs['WCSCDATE']= fileutil.getLTime() self.backup['WCSCDATE'] = 'WCSCDATE' self.revert['WCSCDATE'] = 'WCSCDATE'
python
def archive(self,prepend=None,overwrite=no,quiet=yes): """ Create backup copies of the WCS keywords with the given prepended string. If backup keywords are already present, only update them if 'overwrite' is set to 'yes', otherwise, do warn the user and do nothing. Set the WCSDATE at this time as well. """ # Verify that existing backup values are not overwritten accidentally. if len(list(self.backup.keys())) > 0 and overwrite == no: if not quiet: print('WARNING: Backup WCS keywords already exist! No backup made.') print(' The values can only be overridden if overwrite=yes.') return # Establish what prepend string to use... if prepend is None: if self.prepend is not None: _prefix = self.prepend else: _prefix = DEFAULT_PREFIX else: _prefix = prepend # Update backup and orig_wcs dictionaries # We have archive keywords and a defined prefix # Go through and append them to self.backup self.prepend = _prefix for key in self.wcstrans.keys(): if key != 'pixel scale': _archive_key = self._buildNewKeyname(key,_prefix) else: _archive_key = self.prepend.lower()+'pscale' # if key != 'pixel scale': self.orig_wcs[_archive_key] = self.__dict__[self.wcstrans[key]] self.backup[key] = _archive_key self.revert[_archive_key] = key # Setup keyword to record when these keywords were backed up. self.orig_wcs['WCSCDATE']= fileutil.getLTime() self.backup['WCSCDATE'] = 'WCSCDATE' self.revert['WCSCDATE'] = 'WCSCDATE'
[ "def", "archive", "(", "self", ",", "prepend", "=", "None", ",", "overwrite", "=", "no", ",", "quiet", "=", "yes", ")", ":", "# Verify that existing backup values are not overwritten accidentally.", "if", "len", "(", "list", "(", "self", ".", "backup", ".", "keys", "(", ")", ")", ")", ">", "0", "and", "overwrite", "==", "no", ":", "if", "not", "quiet", ":", "print", "(", "'WARNING: Backup WCS keywords already exist! No backup made.'", ")", "print", "(", "' The values can only be overridden if overwrite=yes.'", ")", "return", "# Establish what prepend string to use...", "if", "prepend", "is", "None", ":", "if", "self", ".", "prepend", "is", "not", "None", ":", "_prefix", "=", "self", ".", "prepend", "else", ":", "_prefix", "=", "DEFAULT_PREFIX", "else", ":", "_prefix", "=", "prepend", "# Update backup and orig_wcs dictionaries", "# We have archive keywords and a defined prefix", "# Go through and append them to self.backup", "self", ".", "prepend", "=", "_prefix", "for", "key", "in", "self", ".", "wcstrans", ".", "keys", "(", ")", ":", "if", "key", "!=", "'pixel scale'", ":", "_archive_key", "=", "self", ".", "_buildNewKeyname", "(", "key", ",", "_prefix", ")", "else", ":", "_archive_key", "=", "self", ".", "prepend", ".", "lower", "(", ")", "+", "'pscale'", "# if key != 'pixel scale':", "self", ".", "orig_wcs", "[", "_archive_key", "]", "=", "self", ".", "__dict__", "[", "self", ".", "wcstrans", "[", "key", "]", "]", "self", ".", "backup", "[", "key", "]", "=", "_archive_key", "self", ".", "revert", "[", "_archive_key", "]", "=", "key", "# Setup keyword to record when these keywords were backed up.", "self", ".", "orig_wcs", "[", "'WCSCDATE'", "]", "=", "fileutil", ".", "getLTime", "(", ")", "self", ".", "backup", "[", "'WCSCDATE'", "]", "=", "'WCSCDATE'", "self", ".", "revert", "[", "'WCSCDATE'", "]", "=", "'WCSCDATE'" ]
Create backup copies of the WCS keywords with the given prepended string. If backup keywords are already present, only update them if 'overwrite' is set to 'yes', otherwise, do warn the user and do nothing. Set the WCSDATE at this time as well.
[ "Create", "backup", "copies", "of", "the", "WCS", "keywords", "with", "the", "given", "prepended", "string", ".", "If", "backup", "keywords", "are", "already", "present", "only", "update", "them", "if", "overwrite", "is", "set", "to", "yes", "otherwise", "do", "warn", "the", "user", "and", "do", "nothing", ".", "Set", "the", "WCSDATE", "at", "this", "time", "as", "well", "." ]
train
https://github.com/spacetelescope/stsci.tools/blob/9a022503ad24ca54ce83331482dfa3ff6de9f403/lib/stsci/tools/wcsutil.py#L824-L864
spacetelescope/stsci.tools
lib/stsci/tools/wcsutil.py
WCSObject.read_archive
def read_archive(self,header,prepend=None): """ Extract a copy of WCS keywords from an open file header, if they have already been created and remember the prefix used for those keywords. Otherwise, setup the current WCS keywords as the archive values. """ # Start by looking for the any backup WCS keywords to # determine whether archived values are present and to set # the prefix used. _prefix = None _archive = False if header is not None: for kw in header.items(): if kw[0][1:] in self.wcstrans.keys(): _prefix = kw[0][0] _archive = True break if not _archive: self.archive(prepend=prepend) return # We have archive keywords and a defined prefix # Go through and append them to self.backup if _prefix is not None: self.prepend = _prefix else: self.prepend = DEFAULT_PREFIX for key in self.wcstrans.keys(): _archive_key = self._buildNewKeyname(key,_prefix) if key!= 'pixel scale': if _archive_key in header: self.orig_wcs[_archive_key] = header[_archive_key] else: self.orig_wcs[_archive_key] = header[key] self.backup[key] = _archive_key self.revert[_archive_key] = key # Establish plate scale value _cd11str = self.prepend+'CD1_1' _cd21str = self.prepend+'CD2_1' pscale = self.compute_pscale(self.orig_wcs[_cd11str],self.orig_wcs[_cd21str]) _archive_key = self.prepend.lower()+'pscale' self.orig_wcs[_archive_key] = pscale self.backup['pixel scale'] = _archive_key self.revert[_archive_key] = 'pixel scale' # Setup keyword to record when these keywords were backed up. if 'WCSCDATE' in header: self.orig_wcs['WCSCDATE'] = header['WCSCDATE'] else: self.orig_wcs['WCSCDATE'] = fileutil.getLTime() self.backup['WCSCDATE'] = 'WCSCDATE' self.revert['WCSCDATE'] = 'WCSCDATE'
python
def read_archive(self,header,prepend=None): """ Extract a copy of WCS keywords from an open file header, if they have already been created and remember the prefix used for those keywords. Otherwise, setup the current WCS keywords as the archive values. """ # Start by looking for the any backup WCS keywords to # determine whether archived values are present and to set # the prefix used. _prefix = None _archive = False if header is not None: for kw in header.items(): if kw[0][1:] in self.wcstrans.keys(): _prefix = kw[0][0] _archive = True break if not _archive: self.archive(prepend=prepend) return # We have archive keywords and a defined prefix # Go through and append them to self.backup if _prefix is not None: self.prepend = _prefix else: self.prepend = DEFAULT_PREFIX for key in self.wcstrans.keys(): _archive_key = self._buildNewKeyname(key,_prefix) if key!= 'pixel scale': if _archive_key in header: self.orig_wcs[_archive_key] = header[_archive_key] else: self.orig_wcs[_archive_key] = header[key] self.backup[key] = _archive_key self.revert[_archive_key] = key # Establish plate scale value _cd11str = self.prepend+'CD1_1' _cd21str = self.prepend+'CD2_1' pscale = self.compute_pscale(self.orig_wcs[_cd11str],self.orig_wcs[_cd21str]) _archive_key = self.prepend.lower()+'pscale' self.orig_wcs[_archive_key] = pscale self.backup['pixel scale'] = _archive_key self.revert[_archive_key] = 'pixel scale' # Setup keyword to record when these keywords were backed up. if 'WCSCDATE' in header: self.orig_wcs['WCSCDATE'] = header['WCSCDATE'] else: self.orig_wcs['WCSCDATE'] = fileutil.getLTime() self.backup['WCSCDATE'] = 'WCSCDATE' self.revert['WCSCDATE'] = 'WCSCDATE'
[ "def", "read_archive", "(", "self", ",", "header", ",", "prepend", "=", "None", ")", ":", "# Start by looking for the any backup WCS keywords to", "# determine whether archived values are present and to set", "# the prefix used.", "_prefix", "=", "None", "_archive", "=", "False", "if", "header", "is", "not", "None", ":", "for", "kw", "in", "header", ".", "items", "(", ")", ":", "if", "kw", "[", "0", "]", "[", "1", ":", "]", "in", "self", ".", "wcstrans", ".", "keys", "(", ")", ":", "_prefix", "=", "kw", "[", "0", "]", "[", "0", "]", "_archive", "=", "True", "break", "if", "not", "_archive", ":", "self", ".", "archive", "(", "prepend", "=", "prepend", ")", "return", "# We have archive keywords and a defined prefix", "# Go through and append them to self.backup", "if", "_prefix", "is", "not", "None", ":", "self", ".", "prepend", "=", "_prefix", "else", ":", "self", ".", "prepend", "=", "DEFAULT_PREFIX", "for", "key", "in", "self", ".", "wcstrans", ".", "keys", "(", ")", ":", "_archive_key", "=", "self", ".", "_buildNewKeyname", "(", "key", ",", "_prefix", ")", "if", "key", "!=", "'pixel scale'", ":", "if", "_archive_key", "in", "header", ":", "self", ".", "orig_wcs", "[", "_archive_key", "]", "=", "header", "[", "_archive_key", "]", "else", ":", "self", ".", "orig_wcs", "[", "_archive_key", "]", "=", "header", "[", "key", "]", "self", ".", "backup", "[", "key", "]", "=", "_archive_key", "self", ".", "revert", "[", "_archive_key", "]", "=", "key", "# Establish plate scale value", "_cd11str", "=", "self", ".", "prepend", "+", "'CD1_1'", "_cd21str", "=", "self", ".", "prepend", "+", "'CD2_1'", "pscale", "=", "self", ".", "compute_pscale", "(", "self", ".", "orig_wcs", "[", "_cd11str", "]", ",", "self", ".", "orig_wcs", "[", "_cd21str", "]", ")", "_archive_key", "=", "self", ".", "prepend", ".", "lower", "(", ")", "+", "'pscale'", "self", ".", "orig_wcs", "[", "_archive_key", "]", "=", "pscale", "self", ".", "backup", "[", "'pixel scale'", "]", "=", "_archive_key", "self", ".", "revert", "[", "_archive_key", "]", "=", "'pixel scale'", "# Setup keyword to record when these keywords were backed up.", "if", "'WCSCDATE'", "in", "header", ":", "self", ".", "orig_wcs", "[", "'WCSCDATE'", "]", "=", "header", "[", "'WCSCDATE'", "]", "else", ":", "self", ".", "orig_wcs", "[", "'WCSCDATE'", "]", "=", "fileutil", ".", "getLTime", "(", ")", "self", ".", "backup", "[", "'WCSCDATE'", "]", "=", "'WCSCDATE'", "self", ".", "revert", "[", "'WCSCDATE'", "]", "=", "'WCSCDATE'" ]
Extract a copy of WCS keywords from an open file header, if they have already been created and remember the prefix used for those keywords. Otherwise, setup the current WCS keywords as the archive values.
[ "Extract", "a", "copy", "of", "WCS", "keywords", "from", "an", "open", "file", "header", "if", "they", "have", "already", "been", "created", "and", "remember", "the", "prefix", "used", "for", "those", "keywords", ".", "Otherwise", "setup", "the", "current", "WCS", "keywords", "as", "the", "archive", "values", "." ]
train
https://github.com/spacetelescope/stsci.tools/blob/9a022503ad24ca54ce83331482dfa3ff6de9f403/lib/stsci/tools/wcsutil.py#L866-L920
spacetelescope/stsci.tools
lib/stsci/tools/wcsutil.py
WCSObject.write_archive
def write_archive(self,fitsname=None,overwrite=no,quiet=yes): """ Saves a copy of the WCS keywords from the image header as new keywords with the user-supplied 'prepend' character(s) prepended to the old keyword names. If the file is a GEIS image and 'fitsname' is not None, create a FITS copy and update that version; otherwise, raise an Exception and do not update anything. """ _fitsname = fitsname # Open image in update mode # Copying of GEIS images handled by 'openImage'. fimg = fileutil.openImage(self.rootname,mode='update',fitsname=_fitsname) if self.rootname.find('.fits') < 0 and _fitsname is not None: # A non-FITS image was provided, and openImage made a copy # Update attributes to point to new copy instead self.geisname = self.rootname self.rootname = _fitsname # extract the extension ID being updated _root,_iextn = fileutil.parseFilename(self.rootname) _extn = fileutil.getExtn(fimg,_iextn) if not quiet: print('Updating archive WCS keywords for ',_fitsname) # Write out values to header... for key in self.orig_wcs.keys(): _comment = None _dkey = self.revert[key] # Verify that archive keywords will not be overwritten, # unless overwrite=yes. _old_key = key in _extn.header if _old_key == True and overwrite == no: if not quiet: print('WCS keyword',key,' already exists! Not overwriting.') continue # No archive keywords exist yet in file, or overwrite=yes... # Extract the value for the original keyword if _dkey in _extn.header: # Extract any comment string for the keyword as well _indx_key = _extn.header.index(_dkey) _full_key = _extn.header.cards[_indx_key] if not quiet: print('updating ',key,' with value of: ',self.orig_wcs[key]) _extn.header[key] = (self.orig_wcs[key], _full_key.comment) key = 'WCSCDATE' if key not in _extn.header: # Print out history keywords to record when these keywords # were backed up. _extn.header[key] = (self.orig_wcs[key], "Time WCS keywords were copied.") # Close the now updated image fimg.close() del fimg
python
def write_archive(self,fitsname=None,overwrite=no,quiet=yes): """ Saves a copy of the WCS keywords from the image header as new keywords with the user-supplied 'prepend' character(s) prepended to the old keyword names. If the file is a GEIS image and 'fitsname' is not None, create a FITS copy and update that version; otherwise, raise an Exception and do not update anything. """ _fitsname = fitsname # Open image in update mode # Copying of GEIS images handled by 'openImage'. fimg = fileutil.openImage(self.rootname,mode='update',fitsname=_fitsname) if self.rootname.find('.fits') < 0 and _fitsname is not None: # A non-FITS image was provided, and openImage made a copy # Update attributes to point to new copy instead self.geisname = self.rootname self.rootname = _fitsname # extract the extension ID being updated _root,_iextn = fileutil.parseFilename(self.rootname) _extn = fileutil.getExtn(fimg,_iextn) if not quiet: print('Updating archive WCS keywords for ',_fitsname) # Write out values to header... for key in self.orig_wcs.keys(): _comment = None _dkey = self.revert[key] # Verify that archive keywords will not be overwritten, # unless overwrite=yes. _old_key = key in _extn.header if _old_key == True and overwrite == no: if not quiet: print('WCS keyword',key,' already exists! Not overwriting.') continue # No archive keywords exist yet in file, or overwrite=yes... # Extract the value for the original keyword if _dkey in _extn.header: # Extract any comment string for the keyword as well _indx_key = _extn.header.index(_dkey) _full_key = _extn.header.cards[_indx_key] if not quiet: print('updating ',key,' with value of: ',self.orig_wcs[key]) _extn.header[key] = (self.orig_wcs[key], _full_key.comment) key = 'WCSCDATE' if key not in _extn.header: # Print out history keywords to record when these keywords # were backed up. _extn.header[key] = (self.orig_wcs[key], "Time WCS keywords were copied.") # Close the now updated image fimg.close() del fimg
[ "def", "write_archive", "(", "self", ",", "fitsname", "=", "None", ",", "overwrite", "=", "no", ",", "quiet", "=", "yes", ")", ":", "_fitsname", "=", "fitsname", "# Open image in update mode", "# Copying of GEIS images handled by 'openImage'.", "fimg", "=", "fileutil", ".", "openImage", "(", "self", ".", "rootname", ",", "mode", "=", "'update'", ",", "fitsname", "=", "_fitsname", ")", "if", "self", ".", "rootname", ".", "find", "(", "'.fits'", ")", "<", "0", "and", "_fitsname", "is", "not", "None", ":", "# A non-FITS image was provided, and openImage made a copy", "# Update attributes to point to new copy instead", "self", ".", "geisname", "=", "self", ".", "rootname", "self", ".", "rootname", "=", "_fitsname", "# extract the extension ID being updated", "_root", ",", "_iextn", "=", "fileutil", ".", "parseFilename", "(", "self", ".", "rootname", ")", "_extn", "=", "fileutil", ".", "getExtn", "(", "fimg", ",", "_iextn", ")", "if", "not", "quiet", ":", "print", "(", "'Updating archive WCS keywords for '", ",", "_fitsname", ")", "# Write out values to header...", "for", "key", "in", "self", ".", "orig_wcs", ".", "keys", "(", ")", ":", "_comment", "=", "None", "_dkey", "=", "self", ".", "revert", "[", "key", "]", "# Verify that archive keywords will not be overwritten,", "# unless overwrite=yes.", "_old_key", "=", "key", "in", "_extn", ".", "header", "if", "_old_key", "==", "True", "and", "overwrite", "==", "no", ":", "if", "not", "quiet", ":", "print", "(", "'WCS keyword'", ",", "key", ",", "' already exists! Not overwriting.'", ")", "continue", "# No archive keywords exist yet in file, or overwrite=yes...", "# Extract the value for the original keyword", "if", "_dkey", "in", "_extn", ".", "header", ":", "# Extract any comment string for the keyword as well", "_indx_key", "=", "_extn", ".", "header", ".", "index", "(", "_dkey", ")", "_full_key", "=", "_extn", ".", "header", ".", "cards", "[", "_indx_key", "]", "if", "not", "quiet", ":", "print", "(", "'updating '", ",", "key", ",", "' with value of: '", ",", "self", ".", "orig_wcs", "[", "key", "]", ")", "_extn", ".", "header", "[", "key", "]", "=", "(", "self", ".", "orig_wcs", "[", "key", "]", ",", "_full_key", ".", "comment", ")", "key", "=", "'WCSCDATE'", "if", "key", "not", "in", "_extn", ".", "header", ":", "# Print out history keywords to record when these keywords", "# were backed up.", "_extn", ".", "header", "[", "key", "]", "=", "(", "self", ".", "orig_wcs", "[", "key", "]", ",", "\"Time WCS keywords were copied.\"", ")", "# Close the now updated image", "fimg", ".", "close", "(", ")", "del", "fimg" ]
Saves a copy of the WCS keywords from the image header as new keywords with the user-supplied 'prepend' character(s) prepended to the old keyword names. If the file is a GEIS image and 'fitsname' is not None, create a FITS copy and update that version; otherwise, raise an Exception and do not update anything.
[ "Saves", "a", "copy", "of", "the", "WCS", "keywords", "from", "the", "image", "header", "as", "new", "keywords", "with", "the", "user", "-", "supplied", "prepend", "character", "(", "s", ")", "prepended", "to", "the", "old", "keyword", "names", "." ]
train
https://github.com/spacetelescope/stsci.tools/blob/9a022503ad24ca54ce83331482dfa3ff6de9f403/lib/stsci/tools/wcsutil.py#L922-L981
spacetelescope/stsci.tools
lib/stsci/tools/wcsutil.py
WCSObject.restoreWCS
def restoreWCS(self,prepend=None): """ Resets the WCS values to the original values stored in the backup keywords recorded in self.backup. """ # Open header for image image = self.rootname if prepend: _prepend = prepend elif self.prepend: _prepend = self.prepend else: _prepend = None # Open image as writable FITS object fimg = fileutil.openImage(image, mode='update') # extract the extension ID being updated _root,_iextn = fileutil.parseFilename(self.rootname) _extn = fileutil.getExtn(fimg,_iextn) if len(self.backup) > 0: # If it knows about the backup keywords already, # use this to restore the original values to the original keywords for newkey in self.revert.keys(): if newkey != 'opscale': _orig_key = self.revert[newkey] _extn.header[_orig_key] = _extn.header[newkey] elif _prepend: for key in self.wcstrans.keys(): # Get new keyword name based on old keyname # and prepend string if key != 'pixel scale': _okey = self._buildNewKeyname(key,_prepend) if _okey in _extn.header: _extn.header[key] = _extn.header[_okey] else: print('No original WCS values found. Exiting...') break else: print('No original WCS values found. Exiting...') fimg.close() del fimg
python
def restoreWCS(self,prepend=None): """ Resets the WCS values to the original values stored in the backup keywords recorded in self.backup. """ # Open header for image image = self.rootname if prepend: _prepend = prepend elif self.prepend: _prepend = self.prepend else: _prepend = None # Open image as writable FITS object fimg = fileutil.openImage(image, mode='update') # extract the extension ID being updated _root,_iextn = fileutil.parseFilename(self.rootname) _extn = fileutil.getExtn(fimg,_iextn) if len(self.backup) > 0: # If it knows about the backup keywords already, # use this to restore the original values to the original keywords for newkey in self.revert.keys(): if newkey != 'opscale': _orig_key = self.revert[newkey] _extn.header[_orig_key] = _extn.header[newkey] elif _prepend: for key in self.wcstrans.keys(): # Get new keyword name based on old keyname # and prepend string if key != 'pixel scale': _okey = self._buildNewKeyname(key,_prepend) if _okey in _extn.header: _extn.header[key] = _extn.header[_okey] else: print('No original WCS values found. Exiting...') break else: print('No original WCS values found. Exiting...') fimg.close() del fimg
[ "def", "restoreWCS", "(", "self", ",", "prepend", "=", "None", ")", ":", "# Open header for image", "image", "=", "self", ".", "rootname", "if", "prepend", ":", "_prepend", "=", "prepend", "elif", "self", ".", "prepend", ":", "_prepend", "=", "self", ".", "prepend", "else", ":", "_prepend", "=", "None", "# Open image as writable FITS object", "fimg", "=", "fileutil", ".", "openImage", "(", "image", ",", "mode", "=", "'update'", ")", "# extract the extension ID being updated", "_root", ",", "_iextn", "=", "fileutil", ".", "parseFilename", "(", "self", ".", "rootname", ")", "_extn", "=", "fileutil", ".", "getExtn", "(", "fimg", ",", "_iextn", ")", "if", "len", "(", "self", ".", "backup", ")", ">", "0", ":", "# If it knows about the backup keywords already,", "# use this to restore the original values to the original keywords", "for", "newkey", "in", "self", ".", "revert", ".", "keys", "(", ")", ":", "if", "newkey", "!=", "'opscale'", ":", "_orig_key", "=", "self", ".", "revert", "[", "newkey", "]", "_extn", ".", "header", "[", "_orig_key", "]", "=", "_extn", ".", "header", "[", "newkey", "]", "elif", "_prepend", ":", "for", "key", "in", "self", ".", "wcstrans", ".", "keys", "(", ")", ":", "# Get new keyword name based on old keyname", "# and prepend string", "if", "key", "!=", "'pixel scale'", ":", "_okey", "=", "self", ".", "_buildNewKeyname", "(", "key", ",", "_prepend", ")", "if", "_okey", "in", "_extn", ".", "header", ":", "_extn", ".", "header", "[", "key", "]", "=", "_extn", ".", "header", "[", "_okey", "]", "else", ":", "print", "(", "'No original WCS values found. Exiting...'", ")", "break", "else", ":", "print", "(", "'No original WCS values found. Exiting...'", ")", "fimg", ".", "close", "(", ")", "del", "fimg" ]
Resets the WCS values to the original values stored in the backup keywords recorded in self.backup.
[ "Resets", "the", "WCS", "values", "to", "the", "original", "values", "stored", "in", "the", "backup", "keywords", "recorded", "in", "self", ".", "backup", "." ]
train
https://github.com/spacetelescope/stsci.tools/blob/9a022503ad24ca54ce83331482dfa3ff6de9f403/lib/stsci/tools/wcsutil.py#L983-L1023
spacetelescope/stsci.tools
lib/stsci/tools/wcsutil.py
WCSObject.createReferenceWCS
def createReferenceWCS(self,refname,overwrite=yes): """ Write out the values of the WCS keywords to the NEW specified image 'fitsname'. """ hdu = self.createWcsHDU() # If refname already exists, delete it to make way for new file if os.path.exists(refname): if overwrite==yes: # Remove previous version and re-create with new header os.remove(refname) hdu.writeto(refname) else: # Append header to existing file wcs_append = True oldhdu = fits.open(refname, mode='append') for e in oldhdu: if 'extname' in e.header and e.header['extname'] == 'WCS': wcs_append = False if wcs_append == True: oldhdu.append(hdu) oldhdu.close() del oldhdu else: # No previous file, so generate new one from scratch hdu.writeto(refname) # Clean up del hdu
python
def createReferenceWCS(self,refname,overwrite=yes): """ Write out the values of the WCS keywords to the NEW specified image 'fitsname'. """ hdu = self.createWcsHDU() # If refname already exists, delete it to make way for new file if os.path.exists(refname): if overwrite==yes: # Remove previous version and re-create with new header os.remove(refname) hdu.writeto(refname) else: # Append header to existing file wcs_append = True oldhdu = fits.open(refname, mode='append') for e in oldhdu: if 'extname' in e.header and e.header['extname'] == 'WCS': wcs_append = False if wcs_append == True: oldhdu.append(hdu) oldhdu.close() del oldhdu else: # No previous file, so generate new one from scratch hdu.writeto(refname) # Clean up del hdu
[ "def", "createReferenceWCS", "(", "self", ",", "refname", ",", "overwrite", "=", "yes", ")", ":", "hdu", "=", "self", ".", "createWcsHDU", "(", ")", "# If refname already exists, delete it to make way for new file", "if", "os", ".", "path", ".", "exists", "(", "refname", ")", ":", "if", "overwrite", "==", "yes", ":", "# Remove previous version and re-create with new header", "os", ".", "remove", "(", "refname", ")", "hdu", ".", "writeto", "(", "refname", ")", "else", ":", "# Append header to existing file", "wcs_append", "=", "True", "oldhdu", "=", "fits", ".", "open", "(", "refname", ",", "mode", "=", "'append'", ")", "for", "e", "in", "oldhdu", ":", "if", "'extname'", "in", "e", ".", "header", "and", "e", ".", "header", "[", "'extname'", "]", "==", "'WCS'", ":", "wcs_append", "=", "False", "if", "wcs_append", "==", "True", ":", "oldhdu", ".", "append", "(", "hdu", ")", "oldhdu", ".", "close", "(", ")", "del", "oldhdu", "else", ":", "# No previous file, so generate new one from scratch", "hdu", ".", "writeto", "(", "refname", ")", "# Clean up", "del", "hdu" ]
Write out the values of the WCS keywords to the NEW specified image 'fitsname'.
[ "Write", "out", "the", "values", "of", "the", "WCS", "keywords", "to", "the", "NEW", "specified", "image", "fitsname", "." ]
train
https://github.com/spacetelescope/stsci.tools/blob/9a022503ad24ca54ce83331482dfa3ff6de9f403/lib/stsci/tools/wcsutil.py#L1025-L1053
spacetelescope/stsci.tools
lib/stsci/tools/wcsutil.py
WCSObject.createWcsHDU
def createWcsHDU(self): """ Generate a WCS header object that can be used to populate a reference WCS HDU. """ hdu = fits.ImageHDU() hdu.header['EXTNAME'] = 'WCS' hdu.header['EXTVER'] = 1 # Now, update original image size information hdu.header['WCSAXES'] = (2, "number of World Coordinate System axes") hdu.header['NPIX1'] = (self.naxis1, "Length of array axis 1") hdu.header['NPIX2'] = (self.naxis2, "Length of array axis 2") hdu.header['PIXVALUE'] = (0.0, "values of pixels in array") # Write out values to header... excluded_keys = ['naxis1','naxis2'] for key in self.wcskeys: _dkey = self.wcstrans[key] if _dkey not in excluded_keys: hdu.header[key] = self.__dict__[_dkey] return hdu
python
def createWcsHDU(self): """ Generate a WCS header object that can be used to populate a reference WCS HDU. """ hdu = fits.ImageHDU() hdu.header['EXTNAME'] = 'WCS' hdu.header['EXTVER'] = 1 # Now, update original image size information hdu.header['WCSAXES'] = (2, "number of World Coordinate System axes") hdu.header['NPIX1'] = (self.naxis1, "Length of array axis 1") hdu.header['NPIX2'] = (self.naxis2, "Length of array axis 2") hdu.header['PIXVALUE'] = (0.0, "values of pixels in array") # Write out values to header... excluded_keys = ['naxis1','naxis2'] for key in self.wcskeys: _dkey = self.wcstrans[key] if _dkey not in excluded_keys: hdu.header[key] = self.__dict__[_dkey] return hdu
[ "def", "createWcsHDU", "(", "self", ")", ":", "hdu", "=", "fits", ".", "ImageHDU", "(", ")", "hdu", ".", "header", "[", "'EXTNAME'", "]", "=", "'WCS'", "hdu", ".", "header", "[", "'EXTVER'", "]", "=", "1", "# Now, update original image size information", "hdu", ".", "header", "[", "'WCSAXES'", "]", "=", "(", "2", ",", "\"number of World Coordinate System axes\"", ")", "hdu", ".", "header", "[", "'NPIX1'", "]", "=", "(", "self", ".", "naxis1", ",", "\"Length of array axis 1\"", ")", "hdu", ".", "header", "[", "'NPIX2'", "]", "=", "(", "self", ".", "naxis2", ",", "\"Length of array axis 2\"", ")", "hdu", ".", "header", "[", "'PIXVALUE'", "]", "=", "(", "0.0", ",", "\"values of pixels in array\"", ")", "# Write out values to header...", "excluded_keys", "=", "[", "'naxis1'", ",", "'naxis2'", "]", "for", "key", "in", "self", ".", "wcskeys", ":", "_dkey", "=", "self", ".", "wcstrans", "[", "key", "]", "if", "_dkey", "not", "in", "excluded_keys", ":", "hdu", ".", "header", "[", "key", "]", "=", "self", ".", "__dict__", "[", "_dkey", "]", "return", "hdu" ]
Generate a WCS header object that can be used to populate a reference WCS HDU.
[ "Generate", "a", "WCS", "header", "object", "that", "can", "be", "used", "to", "populate", "a", "reference", "WCS", "HDU", "." ]
train
https://github.com/spacetelescope/stsci.tools/blob/9a022503ad24ca54ce83331482dfa3ff6de9f403/lib/stsci/tools/wcsutil.py#L1055-L1076
3ll3d00d/vibe
backend/src/analyser/app.py
main
def main(args=None): """ The main routine. """ logger = cfg.configureLogger() for root, dirs, files in os.walk(cfg.dataDir): for dir in dirs: newDir = os.path.join(root, dir) try: os.removedirs(newDir) logger.info("Deleted empty dir " + str(newDir)) except: pass if cfg.useTwisted: import logging logger = logging.getLogger('analyser.twisted') from twisted.internet import reactor from twisted.web.resource import Resource from twisted.web import static, server from twisted.web.wsgi import WSGIResource from twisted.application import service from twisted.internet import endpoints class ReactApp: """ Handles the react app (excluding the static dir). """ def __init__(self, path): # TODO allow this to load when in debug mode even if the files don't exist self.publicFiles = {f: static.File(os.path.join(path, f)) for f in os.listdir(path) if os.path.exists(os.path.join(path, f))} self.indexHtml = ReactIndex(os.path.join(path, 'index.html')) def getFile(self, path): """ overrides getChild so it always just serves index.html unless the file does actually exist (i.e. is an icon or something like that) """ return self.publicFiles.get(path.decode('utf-8'), self.indexHtml) class ReactIndex(static.File): """ a twisted File which overrides getChild so it always just serves index.html (NB: this is a bit of a hack, there is probably a more correct way to do this but...) """ def getChild(self, path, request): return self class FlaskAppWrapper(Resource): """ wraps the flask app as a WSGI resource while allow the react index.html (and its associated static content) to be served as the default page. """ def __init__(self): super().__init__() self.wsgi = WSGIResource(reactor, reactor.getThreadPool(), app) import sys if getattr(sys, 'frozen', False): # pyinstaller lets you copy files to arbitrary locations under the _MEIPASS root dir uiRoot = sys._MEIPASS else: # release script moves the ui under the analyser package because setuptools doesn't seem to include # files from outside the package uiRoot = os.path.dirname(__file__) logger.info('Serving ui from ' + str(uiRoot)) self.react = ReactApp(os.path.join(uiRoot, 'ui')) self.static = static.File(os.path.join(uiRoot, 'ui', 'static')) def getChild(self, path, request): """ Overrides getChild to allow the request to be routed to the wsgi app (i.e. flask for the rest api calls), the static dir (i.e. for the packaged css/js etc), the various concrete files (i.e. the public dir from react-app) or to index.html (i.e. the react app) for everything else. :param path: :param request: :return: """ if path == b'api': request.prepath.pop() request.postpath.insert(0, path) return self.wsgi elif path == b'static': return self.static else: return self.react.getFile(path) def render(self, request): return self.wsgi.render(request) application = service.Application('analyser') site = server.Site(FlaskAppWrapper()) endpoint = endpoints.TCP4ServerEndpoint(reactor, cfg.getPort(), interface='0.0.0.0') endpoint.listen(site) reactor.run() else: # get config from a flask standard place not our config yml app.run(debug=cfg.runInDebug(), host='0.0.0.0', port=cfg.getPort())
python
def main(args=None): """ The main routine. """ logger = cfg.configureLogger() for root, dirs, files in os.walk(cfg.dataDir): for dir in dirs: newDir = os.path.join(root, dir) try: os.removedirs(newDir) logger.info("Deleted empty dir " + str(newDir)) except: pass if cfg.useTwisted: import logging logger = logging.getLogger('analyser.twisted') from twisted.internet import reactor from twisted.web.resource import Resource from twisted.web import static, server from twisted.web.wsgi import WSGIResource from twisted.application import service from twisted.internet import endpoints class ReactApp: """ Handles the react app (excluding the static dir). """ def __init__(self, path): # TODO allow this to load when in debug mode even if the files don't exist self.publicFiles = {f: static.File(os.path.join(path, f)) for f in os.listdir(path) if os.path.exists(os.path.join(path, f))} self.indexHtml = ReactIndex(os.path.join(path, 'index.html')) def getFile(self, path): """ overrides getChild so it always just serves index.html unless the file does actually exist (i.e. is an icon or something like that) """ return self.publicFiles.get(path.decode('utf-8'), self.indexHtml) class ReactIndex(static.File): """ a twisted File which overrides getChild so it always just serves index.html (NB: this is a bit of a hack, there is probably a more correct way to do this but...) """ def getChild(self, path, request): return self class FlaskAppWrapper(Resource): """ wraps the flask app as a WSGI resource while allow the react index.html (and its associated static content) to be served as the default page. """ def __init__(self): super().__init__() self.wsgi = WSGIResource(reactor, reactor.getThreadPool(), app) import sys if getattr(sys, 'frozen', False): # pyinstaller lets you copy files to arbitrary locations under the _MEIPASS root dir uiRoot = sys._MEIPASS else: # release script moves the ui under the analyser package because setuptools doesn't seem to include # files from outside the package uiRoot = os.path.dirname(__file__) logger.info('Serving ui from ' + str(uiRoot)) self.react = ReactApp(os.path.join(uiRoot, 'ui')) self.static = static.File(os.path.join(uiRoot, 'ui', 'static')) def getChild(self, path, request): """ Overrides getChild to allow the request to be routed to the wsgi app (i.e. flask for the rest api calls), the static dir (i.e. for the packaged css/js etc), the various concrete files (i.e. the public dir from react-app) or to index.html (i.e. the react app) for everything else. :param path: :param request: :return: """ if path == b'api': request.prepath.pop() request.postpath.insert(0, path) return self.wsgi elif path == b'static': return self.static else: return self.react.getFile(path) def render(self, request): return self.wsgi.render(request) application = service.Application('analyser') site = server.Site(FlaskAppWrapper()) endpoint = endpoints.TCP4ServerEndpoint(reactor, cfg.getPort(), interface='0.0.0.0') endpoint.listen(site) reactor.run() else: # get config from a flask standard place not our config yml app.run(debug=cfg.runInDebug(), host='0.0.0.0', port=cfg.getPort())
[ "def", "main", "(", "args", "=", "None", ")", ":", "logger", "=", "cfg", ".", "configureLogger", "(", ")", "for", "root", ",", "dirs", ",", "files", "in", "os", ".", "walk", "(", "cfg", ".", "dataDir", ")", ":", "for", "dir", "in", "dirs", ":", "newDir", "=", "os", ".", "path", ".", "join", "(", "root", ",", "dir", ")", "try", ":", "os", ".", "removedirs", "(", "newDir", ")", "logger", ".", "info", "(", "\"Deleted empty dir \"", "+", "str", "(", "newDir", ")", ")", "except", ":", "pass", "if", "cfg", ".", "useTwisted", ":", "import", "logging", "logger", "=", "logging", ".", "getLogger", "(", "'analyser.twisted'", ")", "from", "twisted", ".", "internet", "import", "reactor", "from", "twisted", ".", "web", ".", "resource", "import", "Resource", "from", "twisted", ".", "web", "import", "static", ",", "server", "from", "twisted", ".", "web", ".", "wsgi", "import", "WSGIResource", "from", "twisted", ".", "application", "import", "service", "from", "twisted", ".", "internet", "import", "endpoints", "class", "ReactApp", ":", "\"\"\"\n Handles the react app (excluding the static dir).\n \"\"\"", "def", "__init__", "(", "self", ",", "path", ")", ":", "# TODO allow this to load when in debug mode even if the files don't exist", "self", ".", "publicFiles", "=", "{", "f", ":", "static", ".", "File", "(", "os", ".", "path", ".", "join", "(", "path", ",", "f", ")", ")", "for", "f", "in", "os", ".", "listdir", "(", "path", ")", "if", "os", ".", "path", ".", "exists", "(", "os", ".", "path", ".", "join", "(", "path", ",", "f", ")", ")", "}", "self", ".", "indexHtml", "=", "ReactIndex", "(", "os", ".", "path", ".", "join", "(", "path", ",", "'index.html'", ")", ")", "def", "getFile", "(", "self", ",", "path", ")", ":", "\"\"\"\n overrides getChild so it always just serves index.html unless the file does actually exist (i.e. is an\n icon or something like that)\n \"\"\"", "return", "self", ".", "publicFiles", ".", "get", "(", "path", ".", "decode", "(", "'utf-8'", ")", ",", "self", ".", "indexHtml", ")", "class", "ReactIndex", "(", "static", ".", "File", ")", ":", "\"\"\"\n a twisted File which overrides getChild so it always just serves index.html (NB: this is a bit of a hack, \n there is probably a more correct way to do this but...)\n \"\"\"", "def", "getChild", "(", "self", ",", "path", ",", "request", ")", ":", "return", "self", "class", "FlaskAppWrapper", "(", "Resource", ")", ":", "\"\"\"\n wraps the flask app as a WSGI resource while allow the react index.html (and its associated static content)\n to be served as the default page.\n \"\"\"", "def", "__init__", "(", "self", ")", ":", "super", "(", ")", ".", "__init__", "(", ")", "self", ".", "wsgi", "=", "WSGIResource", "(", "reactor", ",", "reactor", ".", "getThreadPool", "(", ")", ",", "app", ")", "import", "sys", "if", "getattr", "(", "sys", ",", "'frozen'", ",", "False", ")", ":", "# pyinstaller lets you copy files to arbitrary locations under the _MEIPASS root dir", "uiRoot", "=", "sys", ".", "_MEIPASS", "else", ":", "# release script moves the ui under the analyser package because setuptools doesn't seem to include", "# files from outside the package", "uiRoot", "=", "os", ".", "path", ".", "dirname", "(", "__file__", ")", "logger", ".", "info", "(", "'Serving ui from '", "+", "str", "(", "uiRoot", ")", ")", "self", ".", "react", "=", "ReactApp", "(", "os", ".", "path", ".", "join", "(", "uiRoot", ",", "'ui'", ")", ")", "self", ".", "static", "=", "static", ".", "File", "(", "os", ".", "path", ".", "join", "(", "uiRoot", ",", "'ui'", ",", "'static'", ")", ")", "def", "getChild", "(", "self", ",", "path", ",", "request", ")", ":", "\"\"\"\n Overrides getChild to allow the request to be routed to the wsgi app (i.e. flask for the rest api \n calls),\n the static dir (i.e. for the packaged css/js etc), the various concrete files (i.e. the public \n dir from react-app) or to index.html (i.e. the react app) for everything else.\n :param path: \n :param request: \n :return: \n \"\"\"", "if", "path", "==", "b'api'", ":", "request", ".", "prepath", ".", "pop", "(", ")", "request", ".", "postpath", ".", "insert", "(", "0", ",", "path", ")", "return", "self", ".", "wsgi", "elif", "path", "==", "b'static'", ":", "return", "self", ".", "static", "else", ":", "return", "self", ".", "react", ".", "getFile", "(", "path", ")", "def", "render", "(", "self", ",", "request", ")", ":", "return", "self", ".", "wsgi", ".", "render", "(", "request", ")", "application", "=", "service", ".", "Application", "(", "'analyser'", ")", "site", "=", "server", ".", "Site", "(", "FlaskAppWrapper", "(", ")", ")", "endpoint", "=", "endpoints", ".", "TCP4ServerEndpoint", "(", "reactor", ",", "cfg", ".", "getPort", "(", ")", ",", "interface", "=", "'0.0.0.0'", ")", "endpoint", ".", "listen", "(", "site", ")", "reactor", ".", "run", "(", ")", "else", ":", "# get config from a flask standard place not our config yml", "app", ".", "run", "(", "debug", "=", "cfg", ".", "runInDebug", "(", ")", ",", "host", "=", "'0.0.0.0'", ",", "port", "=", "cfg", ".", "getPort", "(", ")", ")" ]
The main routine.
[ "The", "main", "routine", "." ]
train
https://github.com/3ll3d00d/vibe/blob/124b029f13ac746723e92cb47e9cb56edd2e54b5/backend/src/analyser/app.py#L120-L219
fitnr/convertdate
convertdate/french_republican.py
leap
def leap(year, method=None): ''' Determine if this is a leap year in the FR calendar using one of three methods: 4, 100, 128 (every 4th years, every 4th or 400th but not 100th, every 4th but not 128th) ''' method = method or 'equinox' if year in (3, 7, 11): return True elif year < 15: return False if method in (4, 'continuous') or (year <= 16 and method in (128, 'madler', 4, 'continuous')): return year % 4 == 3 elif method in (100, 'romme'): return (year % 4 == 0 and year % 100 != 0) or year % 400 == 0 elif method in (128, 'madler'): return year % 4 == 0 and year % 128 != 0 elif method == 'equinox': # Is equinox on 366th day after (year, 1, 1) startjd = to_jd(year, 1, 1, method='equinox') if premier_da_la_annee(startjd + 367) - startjd == 366.0: return True else: raise ValueError("Unknown leap year method. Try: continuous, romme, madler or equinox") return False
python
def leap(year, method=None): ''' Determine if this is a leap year in the FR calendar using one of three methods: 4, 100, 128 (every 4th years, every 4th or 400th but not 100th, every 4th but not 128th) ''' method = method or 'equinox' if year in (3, 7, 11): return True elif year < 15: return False if method in (4, 'continuous') or (year <= 16 and method in (128, 'madler', 4, 'continuous')): return year % 4 == 3 elif method in (100, 'romme'): return (year % 4 == 0 and year % 100 != 0) or year % 400 == 0 elif method in (128, 'madler'): return year % 4 == 0 and year % 128 != 0 elif method == 'equinox': # Is equinox on 366th day after (year, 1, 1) startjd = to_jd(year, 1, 1, method='equinox') if premier_da_la_annee(startjd + 367) - startjd == 366.0: return True else: raise ValueError("Unknown leap year method. Try: continuous, romme, madler or equinox") return False
[ "def", "leap", "(", "year", ",", "method", "=", "None", ")", ":", "method", "=", "method", "or", "'equinox'", "if", "year", "in", "(", "3", ",", "7", ",", "11", ")", ":", "return", "True", "elif", "year", "<", "15", ":", "return", "False", "if", "method", "in", "(", "4", ",", "'continuous'", ")", "or", "(", "year", "<=", "16", "and", "method", "in", "(", "128", ",", "'madler'", ",", "4", ",", "'continuous'", ")", ")", ":", "return", "year", "%", "4", "==", "3", "elif", "method", "in", "(", "100", ",", "'romme'", ")", ":", "return", "(", "year", "%", "4", "==", "0", "and", "year", "%", "100", "!=", "0", ")", "or", "year", "%", "400", "==", "0", "elif", "method", "in", "(", "128", ",", "'madler'", ")", ":", "return", "year", "%", "4", "==", "0", "and", "year", "%", "128", "!=", "0", "elif", "method", "==", "'equinox'", ":", "# Is equinox on 366th day after (year, 1, 1)", "startjd", "=", "to_jd", "(", "year", ",", "1", ",", "1", ",", "method", "=", "'equinox'", ")", "if", "premier_da_la_annee", "(", "startjd", "+", "367", ")", "-", "startjd", "==", "366.0", ":", "return", "True", "else", ":", "raise", "ValueError", "(", "\"Unknown leap year method. Try: continuous, romme, madler or equinox\"", ")", "return", "False" ]
Determine if this is a leap year in the FR calendar using one of three methods: 4, 100, 128 (every 4th years, every 4th or 400th but not 100th, every 4th but not 128th)
[ "Determine", "if", "this", "is", "a", "leap", "year", "in", "the", "FR", "calendar", "using", "one", "of", "three", "methods", ":", "4", "100", "128", "(", "every", "4th", "years", "every", "4th", "or", "400th", "but", "not", "100th", "every", "4th", "but", "not", "128th", ")" ]
train
https://github.com/fitnr/convertdate/blob/e920f168a87f99183b0aa7290d6c3af222582d43/convertdate/french_republican.py#L48-L78
fitnr/convertdate
convertdate/french_republican.py
premier_da_la_annee
def premier_da_la_annee(jd): '''Determine the year in the French revolutionary calendar in which a given Julian day falls. Returns Julian day number containing fall equinox (first day of the FR year)''' p = ephem.previous_fall_equinox(dublin.from_jd(jd)) previous = trunc(dublin.to_jd(p) - 0.5) + 0.5 if previous + 364 < jd: # test if current day is the equinox if the previous equinox was a long time ago n = ephem.next_fall_equinox(dublin.from_jd(jd)) nxt = trunc(dublin.to_jd(n) - 0.5) + 0.5 if nxt <= jd: return nxt return previous
python
def premier_da_la_annee(jd): '''Determine the year in the French revolutionary calendar in which a given Julian day falls. Returns Julian day number containing fall equinox (first day of the FR year)''' p = ephem.previous_fall_equinox(dublin.from_jd(jd)) previous = trunc(dublin.to_jd(p) - 0.5) + 0.5 if previous + 364 < jd: # test if current day is the equinox if the previous equinox was a long time ago n = ephem.next_fall_equinox(dublin.from_jd(jd)) nxt = trunc(dublin.to_jd(n) - 0.5) + 0.5 if nxt <= jd: return nxt return previous
[ "def", "premier_da_la_annee", "(", "jd", ")", ":", "p", "=", "ephem", ".", "previous_fall_equinox", "(", "dublin", ".", "from_jd", "(", "jd", ")", ")", "previous", "=", "trunc", "(", "dublin", ".", "to_jd", "(", "p", ")", "-", "0.5", ")", "+", "0.5", "if", "previous", "+", "364", "<", "jd", ":", "# test if current day is the equinox if the previous equinox was a long time ago", "n", "=", "ephem", ".", "next_fall_equinox", "(", "dublin", ".", "from_jd", "(", "jd", ")", ")", "nxt", "=", "trunc", "(", "dublin", ".", "to_jd", "(", "n", ")", "-", "0.5", ")", "+", "0.5", "if", "nxt", "<=", "jd", ":", "return", "nxt", "return", "previous" ]
Determine the year in the French revolutionary calendar in which a given Julian day falls. Returns Julian day number containing fall equinox (first day of the FR year)
[ "Determine", "the", "year", "in", "the", "French", "revolutionary", "calendar", "in", "which", "a", "given", "Julian", "day", "falls", ".", "Returns", "Julian", "day", "number", "containing", "fall", "equinox", "(", "first", "day", "of", "the", "FR", "year", ")" ]
train
https://github.com/fitnr/convertdate/blob/e920f168a87f99183b0aa7290d6c3af222582d43/convertdate/french_republican.py#L81-L95
fitnr/convertdate
convertdate/french_republican.py
to_jd
def to_jd(year, month, day, method=None): '''Obtain Julian day from a given French Revolutionary calendar date.''' method = method or 'equinox' if day < 1 or day > 30: raise ValueError("Invalid day for this calendar") if month > 13: raise ValueError("Invalid month for this calendar") if month == 13 and day > 5 + leap(year, method=method): raise ValueError("Invalid day for this month in this calendar") if method == 'equinox': return _to_jd_equinox(year, month, day) else: return _to_jd_schematic(year, month, day, method)
python
def to_jd(year, month, day, method=None): '''Obtain Julian day from a given French Revolutionary calendar date.''' method = method or 'equinox' if day < 1 or day > 30: raise ValueError("Invalid day for this calendar") if month > 13: raise ValueError("Invalid month for this calendar") if month == 13 and day > 5 + leap(year, method=method): raise ValueError("Invalid day for this month in this calendar") if method == 'equinox': return _to_jd_equinox(year, month, day) else: return _to_jd_schematic(year, month, day, method)
[ "def", "to_jd", "(", "year", ",", "month", ",", "day", ",", "method", "=", "None", ")", ":", "method", "=", "method", "or", "'equinox'", "if", "day", "<", "1", "or", "day", ">", "30", ":", "raise", "ValueError", "(", "\"Invalid day for this calendar\"", ")", "if", "month", ">", "13", ":", "raise", "ValueError", "(", "\"Invalid month for this calendar\"", ")", "if", "month", "==", "13", "and", "day", ">", "5", "+", "leap", "(", "year", ",", "method", "=", "method", ")", ":", "raise", "ValueError", "(", "\"Invalid day for this month in this calendar\"", ")", "if", "method", "==", "'equinox'", ":", "return", "_to_jd_equinox", "(", "year", ",", "month", ",", "day", ")", "else", ":", "return", "_to_jd_schematic", "(", "year", ",", "month", ",", "day", ",", "method", ")" ]
Obtain Julian day from a given French Revolutionary calendar date.
[ "Obtain", "Julian", "day", "from", "a", "given", "French", "Revolutionary", "calendar", "date", "." ]
train
https://github.com/fitnr/convertdate/blob/e920f168a87f99183b0aa7290d6c3af222582d43/convertdate/french_republican.py#L98-L115
fitnr/convertdate
convertdate/french_republican.py
_to_jd_schematic
def _to_jd_schematic(year, month, day, method): '''Calculate JD using various leap-year calculation methods''' y0, y1, y2, y3, y4, y5 = 0, 0, 0, 0, 0, 0 intercal_cycle_yrs, over_cycle_yrs, leap_suppression_yrs = None, None, None # Use the every-four-years method below year 16 (madler) or below 15 (romme) if ((method in (100, 'romme') and year < 15) or (method in (128, 'madler') and year < 17)): method = 4 if method in (4, 'continuous'): # Leap years: 15, 19, 23, ... y5 = -365 elif method in (100, 'romme'): year = year - 13 y5 = DAYS_IN_YEAR * 12 + 3 leap_suppression_yrs = 100. leap_suppression_days = 36524 # leap_cycle_days * 25 - 1 intercal_cycle_yrs = 400. intercal_cycle_days = 146097 # leap_suppression_days * 4 + 1 over_cycle_yrs = 4000. over_cycle_days = 1460969 # intercal_cycle_days * 10 - 1 elif method in (128, 'madler'): year = year - 17 y5 = DAYS_IN_YEAR * 16 + 4 leap_suppression_days = 46751 # 32 * leap_cycle_days - 1 leap_suppression_yrs = 128 else: raise ValueError("Unknown leap year method. Try: continuous, romme, madler or equinox") if over_cycle_yrs: y0 = trunc(year / over_cycle_yrs) * over_cycle_days year = year % over_cycle_yrs # count intercalary cycles in days (400 years long or None) if intercal_cycle_yrs: y1 = trunc(year / intercal_cycle_yrs) * intercal_cycle_days year = year % intercal_cycle_yrs # count leap suppresion cycles in days (100 or 128 years long) if leap_suppression_yrs: y2 = trunc(year / leap_suppression_yrs) * leap_suppression_days year = year % leap_suppression_yrs y3 = trunc(year / LEAP_CYCLE_YEARS) * LEAP_CYCLE_DAYS year = year % LEAP_CYCLE_YEARS # Adjust 'year' by one to account for lack of year 0 y4 = year * DAYS_IN_YEAR yj = y0 + y1 + y2 + y3 + y4 + y5 mj = (month - 1) * 30 return EPOCH + yj + mj + day - 1
python
def _to_jd_schematic(year, month, day, method): '''Calculate JD using various leap-year calculation methods''' y0, y1, y2, y3, y4, y5 = 0, 0, 0, 0, 0, 0 intercal_cycle_yrs, over_cycle_yrs, leap_suppression_yrs = None, None, None # Use the every-four-years method below year 16 (madler) or below 15 (romme) if ((method in (100, 'romme') and year < 15) or (method in (128, 'madler') and year < 17)): method = 4 if method in (4, 'continuous'): # Leap years: 15, 19, 23, ... y5 = -365 elif method in (100, 'romme'): year = year - 13 y5 = DAYS_IN_YEAR * 12 + 3 leap_suppression_yrs = 100. leap_suppression_days = 36524 # leap_cycle_days * 25 - 1 intercal_cycle_yrs = 400. intercal_cycle_days = 146097 # leap_suppression_days * 4 + 1 over_cycle_yrs = 4000. over_cycle_days = 1460969 # intercal_cycle_days * 10 - 1 elif method in (128, 'madler'): year = year - 17 y5 = DAYS_IN_YEAR * 16 + 4 leap_suppression_days = 46751 # 32 * leap_cycle_days - 1 leap_suppression_yrs = 128 else: raise ValueError("Unknown leap year method. Try: continuous, romme, madler or equinox") if over_cycle_yrs: y0 = trunc(year / over_cycle_yrs) * over_cycle_days year = year % over_cycle_yrs # count intercalary cycles in days (400 years long or None) if intercal_cycle_yrs: y1 = trunc(year / intercal_cycle_yrs) * intercal_cycle_days year = year % intercal_cycle_yrs # count leap suppresion cycles in days (100 or 128 years long) if leap_suppression_yrs: y2 = trunc(year / leap_suppression_yrs) * leap_suppression_days year = year % leap_suppression_yrs y3 = trunc(year / LEAP_CYCLE_YEARS) * LEAP_CYCLE_DAYS year = year % LEAP_CYCLE_YEARS # Adjust 'year' by one to account for lack of year 0 y4 = year * DAYS_IN_YEAR yj = y0 + y1 + y2 + y3 + y4 + y5 mj = (month - 1) * 30 return EPOCH + yj + mj + day - 1
[ "def", "_to_jd_schematic", "(", "year", ",", "month", ",", "day", ",", "method", ")", ":", "y0", ",", "y1", ",", "y2", ",", "y3", ",", "y4", ",", "y5", "=", "0", ",", "0", ",", "0", ",", "0", ",", "0", ",", "0", "intercal_cycle_yrs", ",", "over_cycle_yrs", ",", "leap_suppression_yrs", "=", "None", ",", "None", ",", "None", "# Use the every-four-years method below year 16 (madler) or below 15 (romme)", "if", "(", "(", "method", "in", "(", "100", ",", "'romme'", ")", "and", "year", "<", "15", ")", "or", "(", "method", "in", "(", "128", ",", "'madler'", ")", "and", "year", "<", "17", ")", ")", ":", "method", "=", "4", "if", "method", "in", "(", "4", ",", "'continuous'", ")", ":", "# Leap years: 15, 19, 23, ...", "y5", "=", "-", "365", "elif", "method", "in", "(", "100", ",", "'romme'", ")", ":", "year", "=", "year", "-", "13", "y5", "=", "DAYS_IN_YEAR", "*", "12", "+", "3", "leap_suppression_yrs", "=", "100.", "leap_suppression_days", "=", "36524", "# leap_cycle_days * 25 - 1", "intercal_cycle_yrs", "=", "400.", "intercal_cycle_days", "=", "146097", "# leap_suppression_days * 4 + 1", "over_cycle_yrs", "=", "4000.", "over_cycle_days", "=", "1460969", "# intercal_cycle_days * 10 - 1", "elif", "method", "in", "(", "128", ",", "'madler'", ")", ":", "year", "=", "year", "-", "17", "y5", "=", "DAYS_IN_YEAR", "*", "16", "+", "4", "leap_suppression_days", "=", "46751", "# 32 * leap_cycle_days - 1", "leap_suppression_yrs", "=", "128", "else", ":", "raise", "ValueError", "(", "\"Unknown leap year method. Try: continuous, romme, madler or equinox\"", ")", "if", "over_cycle_yrs", ":", "y0", "=", "trunc", "(", "year", "/", "over_cycle_yrs", ")", "*", "over_cycle_days", "year", "=", "year", "%", "over_cycle_yrs", "# count intercalary cycles in days (400 years long or None)", "if", "intercal_cycle_yrs", ":", "y1", "=", "trunc", "(", "year", "/", "intercal_cycle_yrs", ")", "*", "intercal_cycle_days", "year", "=", "year", "%", "intercal_cycle_yrs", "# count leap suppresion cycles in days (100 or 128 years long)", "if", "leap_suppression_yrs", ":", "y2", "=", "trunc", "(", "year", "/", "leap_suppression_yrs", ")", "*", "leap_suppression_days", "year", "=", "year", "%", "leap_suppression_yrs", "y3", "=", "trunc", "(", "year", "/", "LEAP_CYCLE_YEARS", ")", "*", "LEAP_CYCLE_DAYS", "year", "=", "year", "%", "LEAP_CYCLE_YEARS", "# Adjust 'year' by one to account for lack of year 0", "y4", "=", "year", "*", "DAYS_IN_YEAR", "yj", "=", "y0", "+", "y1", "+", "y2", "+", "y3", "+", "y4", "+", "y5", "mj", "=", "(", "month", "-", "1", ")", "*", "30", "return", "EPOCH", "+", "yj", "+", "mj", "+", "day", "-", "1" ]
Calculate JD using various leap-year calculation methods
[ "Calculate", "JD", "using", "various", "leap", "-", "year", "calculation", "methods" ]
train
https://github.com/fitnr/convertdate/blob/e920f168a87f99183b0aa7290d6c3af222582d43/convertdate/french_republican.py#L118-L181
fitnr/convertdate
convertdate/french_republican.py
from_jd
def from_jd(jd, method=None): '''Calculate date in the French Revolutionary calendar from Julian day. The five or six "sansculottides" are considered a thirteenth month in the results of this function.''' method = method or 'equinox' if method == 'equinox': return _from_jd_equinox(jd) else: return _from_jd_schematic(jd, method)
python
def from_jd(jd, method=None): '''Calculate date in the French Revolutionary calendar from Julian day. The five or six "sansculottides" are considered a thirteenth month in the results of this function.''' method = method or 'equinox' if method == 'equinox': return _from_jd_equinox(jd) else: return _from_jd_schematic(jd, method)
[ "def", "from_jd", "(", "jd", ",", "method", "=", "None", ")", ":", "method", "=", "method", "or", "'equinox'", "if", "method", "==", "'equinox'", ":", "return", "_from_jd_equinox", "(", "jd", ")", "else", ":", "return", "_from_jd_schematic", "(", "jd", ",", "method", ")" ]
Calculate date in the French Revolutionary calendar from Julian day. The five or six "sansculottides" are considered a thirteenth month in the results of this function.
[ "Calculate", "date", "in", "the", "French", "Revolutionary", "calendar", "from", "Julian", "day", ".", "The", "five", "or", "six", "sansculottides", "are", "considered", "a", "thirteenth", "month", "in", "the", "results", "of", "this", "function", "." ]
train
https://github.com/fitnr/convertdate/blob/e920f168a87f99183b0aa7290d6c3af222582d43/convertdate/french_republican.py#L190-L201
fitnr/convertdate
convertdate/french_republican.py
_from_jd_schematic
def _from_jd_schematic(jd, method): '''Convert from JD using various leap-year calculation methods''' if jd < EPOCH: raise ValueError("Can't convert days before the French Revolution") # days since Epoch J = trunc(jd) + 0.5 - EPOCH y0, y1, y2, y3, y4, y5 = 0, 0, 0, 0, 0, 0 intercal_cycle_days = leap_suppression_days = over_cycle_days = None # Use the every-four-years method below year 17 if (J <= DAYS_IN_YEAR * 12 + 3 and method in (100, 'romme')) or (J <= DAYS_IN_YEAR * 17 + 4 and method in (128, 'madler')): method = 4 # set p and r in Hatcher algorithm if method in (4, 'continuous'): # Leap years: 15, 19, 23, ... # Reorganize so that leap day is last day of cycle J = J + 365 y5 = - 1 elif method in (100, 'romme'): # Year 15 is not a leap year # Year 16 is leap, then multiples of 4, not multiples of 100, yes multiples of 400 y5 = 12 J = J - DAYS_IN_YEAR * 12 - 3 leap_suppression_yrs = 100. leap_suppression_days = 36524 # LEAP_CYCLE_DAYS * 25 - 1 intercal_cycle_yrs = 400. intercal_cycle_days = 146097 # leap_suppression_days * 4 + 1 over_cycle_yrs = 4000. over_cycle_days = 1460969 # intercal_cycle_days * 10 - 1 elif method in (128, 'madler'): # Year 15 is a leap year, then year 20 and multiples of 4, not multiples of 128 y5 = 16 J = J - DAYS_IN_YEAR * 16 - 4 leap_suppression_yrs = 128 leap_suppression_days = 46751 # 32 * leap_cycle_days - 1 else: raise ValueError("Unknown leap year method. Try: continuous, romme, madler or equinox") if over_cycle_days: y0 = trunc(J / over_cycle_days) * over_cycle_yrs J = J % over_cycle_days if intercal_cycle_days: y1 = trunc(J / intercal_cycle_days) * intercal_cycle_yrs J = J % intercal_cycle_days if leap_suppression_days: y2 = trunc(J / leap_suppression_days) * leap_suppression_yrs J = J % leap_suppression_days y3 = trunc(J / LEAP_CYCLE_DAYS) * LEAP_CYCLE_YEARS if J % LEAP_CYCLE_DAYS == LEAP_CYCLE_DAYS - 1: J = 1460 else: J = J % LEAP_CYCLE_DAYS # 0 <= J <= 1460 # J needs to be 365 here on leap days ONLY y4 = trunc(J / DAYS_IN_YEAR) if J == DAYS_IN_YEAR * 4: y4 = y4 - 1 J = 365.0 else: J = J % DAYS_IN_YEAR year = y0 + y1 + y2 + y3 + y4 + y5 month = trunc(J / 30.) J = J - month * 30 return year + 1, month + 1, trunc(J) + 1
python
def _from_jd_schematic(jd, method): '''Convert from JD using various leap-year calculation methods''' if jd < EPOCH: raise ValueError("Can't convert days before the French Revolution") # days since Epoch J = trunc(jd) + 0.5 - EPOCH y0, y1, y2, y3, y4, y5 = 0, 0, 0, 0, 0, 0 intercal_cycle_days = leap_suppression_days = over_cycle_days = None # Use the every-four-years method below year 17 if (J <= DAYS_IN_YEAR * 12 + 3 and method in (100, 'romme')) or (J <= DAYS_IN_YEAR * 17 + 4 and method in (128, 'madler')): method = 4 # set p and r in Hatcher algorithm if method in (4, 'continuous'): # Leap years: 15, 19, 23, ... # Reorganize so that leap day is last day of cycle J = J + 365 y5 = - 1 elif method in (100, 'romme'): # Year 15 is not a leap year # Year 16 is leap, then multiples of 4, not multiples of 100, yes multiples of 400 y5 = 12 J = J - DAYS_IN_YEAR * 12 - 3 leap_suppression_yrs = 100. leap_suppression_days = 36524 # LEAP_CYCLE_DAYS * 25 - 1 intercal_cycle_yrs = 400. intercal_cycle_days = 146097 # leap_suppression_days * 4 + 1 over_cycle_yrs = 4000. over_cycle_days = 1460969 # intercal_cycle_days * 10 - 1 elif method in (128, 'madler'): # Year 15 is a leap year, then year 20 and multiples of 4, not multiples of 128 y5 = 16 J = J - DAYS_IN_YEAR * 16 - 4 leap_suppression_yrs = 128 leap_suppression_days = 46751 # 32 * leap_cycle_days - 1 else: raise ValueError("Unknown leap year method. Try: continuous, romme, madler or equinox") if over_cycle_days: y0 = trunc(J / over_cycle_days) * over_cycle_yrs J = J % over_cycle_days if intercal_cycle_days: y1 = trunc(J / intercal_cycle_days) * intercal_cycle_yrs J = J % intercal_cycle_days if leap_suppression_days: y2 = trunc(J / leap_suppression_days) * leap_suppression_yrs J = J % leap_suppression_days y3 = trunc(J / LEAP_CYCLE_DAYS) * LEAP_CYCLE_YEARS if J % LEAP_CYCLE_DAYS == LEAP_CYCLE_DAYS - 1: J = 1460 else: J = J % LEAP_CYCLE_DAYS # 0 <= J <= 1460 # J needs to be 365 here on leap days ONLY y4 = trunc(J / DAYS_IN_YEAR) if J == DAYS_IN_YEAR * 4: y4 = y4 - 1 J = 365.0 else: J = J % DAYS_IN_YEAR year = y0 + y1 + y2 + y3 + y4 + y5 month = trunc(J / 30.) J = J - month * 30 return year + 1, month + 1, trunc(J) + 1
[ "def", "_from_jd_schematic", "(", "jd", ",", "method", ")", ":", "if", "jd", "<", "EPOCH", ":", "raise", "ValueError", "(", "\"Can't convert days before the French Revolution\"", ")", "# days since Epoch", "J", "=", "trunc", "(", "jd", ")", "+", "0.5", "-", "EPOCH", "y0", ",", "y1", ",", "y2", ",", "y3", ",", "y4", ",", "y5", "=", "0", ",", "0", ",", "0", ",", "0", ",", "0", ",", "0", "intercal_cycle_days", "=", "leap_suppression_days", "=", "over_cycle_days", "=", "None", "# Use the every-four-years method below year 17", "if", "(", "J", "<=", "DAYS_IN_YEAR", "*", "12", "+", "3", "and", "method", "in", "(", "100", ",", "'romme'", ")", ")", "or", "(", "J", "<=", "DAYS_IN_YEAR", "*", "17", "+", "4", "and", "method", "in", "(", "128", ",", "'madler'", ")", ")", ":", "method", "=", "4", "# set p and r in Hatcher algorithm", "if", "method", "in", "(", "4", ",", "'continuous'", ")", ":", "# Leap years: 15, 19, 23, ...", "# Reorganize so that leap day is last day of cycle", "J", "=", "J", "+", "365", "y5", "=", "-", "1", "elif", "method", "in", "(", "100", ",", "'romme'", ")", ":", "# Year 15 is not a leap year", "# Year 16 is leap, then multiples of 4, not multiples of 100, yes multiples of 400", "y5", "=", "12", "J", "=", "J", "-", "DAYS_IN_YEAR", "*", "12", "-", "3", "leap_suppression_yrs", "=", "100.", "leap_suppression_days", "=", "36524", "# LEAP_CYCLE_DAYS * 25 - 1", "intercal_cycle_yrs", "=", "400.", "intercal_cycle_days", "=", "146097", "# leap_suppression_days * 4 + 1", "over_cycle_yrs", "=", "4000.", "over_cycle_days", "=", "1460969", "# intercal_cycle_days * 10 - 1", "elif", "method", "in", "(", "128", ",", "'madler'", ")", ":", "# Year 15 is a leap year, then year 20 and multiples of 4, not multiples of 128", "y5", "=", "16", "J", "=", "J", "-", "DAYS_IN_YEAR", "*", "16", "-", "4", "leap_suppression_yrs", "=", "128", "leap_suppression_days", "=", "46751", "# 32 * leap_cycle_days - 1", "else", ":", "raise", "ValueError", "(", "\"Unknown leap year method. Try: continuous, romme, madler or equinox\"", ")", "if", "over_cycle_days", ":", "y0", "=", "trunc", "(", "J", "/", "over_cycle_days", ")", "*", "over_cycle_yrs", "J", "=", "J", "%", "over_cycle_days", "if", "intercal_cycle_days", ":", "y1", "=", "trunc", "(", "J", "/", "intercal_cycle_days", ")", "*", "intercal_cycle_yrs", "J", "=", "J", "%", "intercal_cycle_days", "if", "leap_suppression_days", ":", "y2", "=", "trunc", "(", "J", "/", "leap_suppression_days", ")", "*", "leap_suppression_yrs", "J", "=", "J", "%", "leap_suppression_days", "y3", "=", "trunc", "(", "J", "/", "LEAP_CYCLE_DAYS", ")", "*", "LEAP_CYCLE_YEARS", "if", "J", "%", "LEAP_CYCLE_DAYS", "==", "LEAP_CYCLE_DAYS", "-", "1", ":", "J", "=", "1460", "else", ":", "J", "=", "J", "%", "LEAP_CYCLE_DAYS", "# 0 <= J <= 1460", "# J needs to be 365 here on leap days ONLY", "y4", "=", "trunc", "(", "J", "/", "DAYS_IN_YEAR", ")", "if", "J", "==", "DAYS_IN_YEAR", "*", "4", ":", "y4", "=", "y4", "-", "1", "J", "=", "365.0", "else", ":", "J", "=", "J", "%", "DAYS_IN_YEAR", "year", "=", "y0", "+", "y1", "+", "y2", "+", "y3", "+", "y4", "+", "y5", "month", "=", "trunc", "(", "J", "/", "30.", ")", "J", "=", "J", "-", "month", "*", "30", "return", "year", "+", "1", ",", "month", "+", "1", ",", "trunc", "(", "J", ")", "+", "1" ]
Convert from JD using various leap-year calculation methods
[ "Convert", "from", "JD", "using", "various", "leap", "-", "year", "calculation", "methods" ]
train
https://github.com/fitnr/convertdate/blob/e920f168a87f99183b0aa7290d6c3af222582d43/convertdate/french_republican.py#L204-L288
fitnr/convertdate
convertdate/french_republican.py
_from_jd_equinox
def _from_jd_equinox(jd): '''Calculate the FR day using the equinox as day 1''' jd = trunc(jd) + 0.5 equinoxe = premier_da_la_annee(jd) an = gregorian.from_jd(equinoxe)[0] - YEAR_EPOCH mois = trunc((jd - equinoxe) / 30.) + 1 jour = int((jd - equinoxe) % 30) + 1 return (an, mois, jour)
python
def _from_jd_equinox(jd): '''Calculate the FR day using the equinox as day 1''' jd = trunc(jd) + 0.5 equinoxe = premier_da_la_annee(jd) an = gregorian.from_jd(equinoxe)[0] - YEAR_EPOCH mois = trunc((jd - equinoxe) / 30.) + 1 jour = int((jd - equinoxe) % 30) + 1 return (an, mois, jour)
[ "def", "_from_jd_equinox", "(", "jd", ")", ":", "jd", "=", "trunc", "(", "jd", ")", "+", "0.5", "equinoxe", "=", "premier_da_la_annee", "(", "jd", ")", "an", "=", "gregorian", ".", "from_jd", "(", "equinoxe", ")", "[", "0", "]", "-", "YEAR_EPOCH", "mois", "=", "trunc", "(", "(", "jd", "-", "equinoxe", ")", "/", "30.", ")", "+", "1", "jour", "=", "int", "(", "(", "jd", "-", "equinoxe", ")", "%", "30", ")", "+", "1", "return", "(", "an", ",", "mois", ",", "jour", ")" ]
Calculate the FR day using the equinox as day 1
[ "Calculate", "the", "FR", "day", "using", "the", "equinox", "as", "day", "1" ]
train
https://github.com/fitnr/convertdate/blob/e920f168a87f99183b0aa7290d6c3af222582d43/convertdate/french_republican.py#L291-L300
fitnr/convertdate
convertdate/indian_civil.py
to_jd
def to_jd(year, month, day): '''Obtain Julian day for Indian Civil date''' gyear = year + 78 leap = isleap(gyear) # // Is this a leap year ? # 22 - leap = 21 if leap, 22 non-leap start = gregorian.to_jd(gyear, 3, 22 - leap) if leap: Caitra = 31 else: Caitra = 30 if month == 1: jd = start + (day - 1) else: jd = start + Caitra m = month - 2 m = min(m, 5) jd += m * 31 if month >= 8: m = month - 7 jd += m * 30 jd += day - 1 return jd
python
def to_jd(year, month, day): '''Obtain Julian day for Indian Civil date''' gyear = year + 78 leap = isleap(gyear) # // Is this a leap year ? # 22 - leap = 21 if leap, 22 non-leap start = gregorian.to_jd(gyear, 3, 22 - leap) if leap: Caitra = 31 else: Caitra = 30 if month == 1: jd = start + (day - 1) else: jd = start + Caitra m = month - 2 m = min(m, 5) jd += m * 31 if month >= 8: m = month - 7 jd += m * 30 jd += day - 1 return jd
[ "def", "to_jd", "(", "year", ",", "month", ",", "day", ")", ":", "gyear", "=", "year", "+", "78", "leap", "=", "isleap", "(", "gyear", ")", "# // Is this a leap year ?", "# 22 - leap = 21 if leap, 22 non-leap", "start", "=", "gregorian", ".", "to_jd", "(", "gyear", ",", "3", ",", "22", "-", "leap", ")", "if", "leap", ":", "Caitra", "=", "31", "else", ":", "Caitra", "=", "30", "if", "month", "==", "1", ":", "jd", "=", "start", "+", "(", "day", "-", "1", ")", "else", ":", "jd", "=", "start", "+", "Caitra", "m", "=", "month", "-", "2", "m", "=", "min", "(", "m", ",", "5", ")", "jd", "+=", "m", "*", "31", "if", "month", ">=", "8", ":", "m", "=", "month", "-", "7", "jd", "+=", "m", "*", "30", "jd", "+=", "day", "-", "1", "return", "jd" ]
Obtain Julian day for Indian Civil date
[ "Obtain", "Julian", "day", "for", "Indian", "Civil", "date" ]
train
https://github.com/fitnr/convertdate/blob/e920f168a87f99183b0aa7290d6c3af222582d43/convertdate/indian_civil.py#L47-L74
fitnr/convertdate
convertdate/indian_civil.py
from_jd
def from_jd(jd): '''Calculate Indian Civil date from Julian day Offset in years from Saka era to Gregorian epoch''' start = 80 # Day offset between Saka and Gregorian jd = trunc(jd) + 0.5 greg = gregorian.from_jd(jd) # Gregorian date for Julian day leap = isleap(greg[0]) # Is this a leap year? # Tentative year in Saka era year = greg[0] - SAKA_EPOCH # JD at start of Gregorian year greg0 = gregorian.to_jd(greg[0], 1, 1) yday = jd - greg0 # Day number (0 based) in Gregorian year if leap: Caitra = 31 # Days in Caitra this year else: Caitra = 30 if yday < start: # Day is at the end of the preceding Saka year year -= 1 yday += Caitra + (31 * 5) + (30 * 3) + 10 + start yday -= start if yday < Caitra: month = 1 day = yday + 1 else: mday = yday - Caitra if (mday < (31 * 5)): month = trunc(mday / 31) + 2 day = (mday % 31) + 1 else: mday -= 31 * 5 month = trunc(mday / 30) + 7 day = (mday % 30) + 1 return (year, month, int(day))
python
def from_jd(jd): '''Calculate Indian Civil date from Julian day Offset in years from Saka era to Gregorian epoch''' start = 80 # Day offset between Saka and Gregorian jd = trunc(jd) + 0.5 greg = gregorian.from_jd(jd) # Gregorian date for Julian day leap = isleap(greg[0]) # Is this a leap year? # Tentative year in Saka era year = greg[0] - SAKA_EPOCH # JD at start of Gregorian year greg0 = gregorian.to_jd(greg[0], 1, 1) yday = jd - greg0 # Day number (0 based) in Gregorian year if leap: Caitra = 31 # Days in Caitra this year else: Caitra = 30 if yday < start: # Day is at the end of the preceding Saka year year -= 1 yday += Caitra + (31 * 5) + (30 * 3) + 10 + start yday -= start if yday < Caitra: month = 1 day = yday + 1 else: mday = yday - Caitra if (mday < (31 * 5)): month = trunc(mday / 31) + 2 day = (mday % 31) + 1 else: mday -= 31 * 5 month = trunc(mday / 30) + 7 day = (mday % 30) + 1 return (year, month, int(day))
[ "def", "from_jd", "(", "jd", ")", ":", "start", "=", "80", "# Day offset between Saka and Gregorian", "jd", "=", "trunc", "(", "jd", ")", "+", "0.5", "greg", "=", "gregorian", ".", "from_jd", "(", "jd", ")", "# Gregorian date for Julian day", "leap", "=", "isleap", "(", "greg", "[", "0", "]", ")", "# Is this a leap year?", "# Tentative year in Saka era", "year", "=", "greg", "[", "0", "]", "-", "SAKA_EPOCH", "# JD at start of Gregorian year", "greg0", "=", "gregorian", ".", "to_jd", "(", "greg", "[", "0", "]", ",", "1", ",", "1", ")", "yday", "=", "jd", "-", "greg0", "# Day number (0 based) in Gregorian year", "if", "leap", ":", "Caitra", "=", "31", "# Days in Caitra this year", "else", ":", "Caitra", "=", "30", "if", "yday", "<", "start", ":", "# Day is at the end of the preceding Saka year", "year", "-=", "1", "yday", "+=", "Caitra", "+", "(", "31", "*", "5", ")", "+", "(", "30", "*", "3", ")", "+", "10", "+", "start", "yday", "-=", "start", "if", "yday", "<", "Caitra", ":", "month", "=", "1", "day", "=", "yday", "+", "1", "else", ":", "mday", "=", "yday", "-", "Caitra", "if", "(", "mday", "<", "(", "31", "*", "5", ")", ")", ":", "month", "=", "trunc", "(", "mday", "/", "31", ")", "+", "2", "day", "=", "(", "mday", "%", "31", ")", "+", "1", "else", ":", "mday", "-=", "31", "*", "5", "month", "=", "trunc", "(", "mday", "/", "30", ")", "+", "7", "day", "=", "(", "mday", "%", "30", ")", "+", "1", "return", "(", "year", ",", "month", ",", "int", "(", "day", ")", ")" ]
Calculate Indian Civil date from Julian day Offset in years from Saka era to Gregorian epoch
[ "Calculate", "Indian", "Civil", "date", "from", "Julian", "day", "Offset", "in", "years", "from", "Saka", "era", "to", "Gregorian", "epoch" ]
train
https://github.com/fitnr/convertdate/blob/e920f168a87f99183b0aa7290d6c3af222582d43/convertdate/indian_civil.py#L77-L117
ionelmc/python-aspectlib
src/aspectlib/debug.py
format_stack
def format_stack(skip=0, length=6, _sep=os.path.sep): """ Returns a one-line string with the current callstack. """ return ' < '.join("%s:%s:%s" % ( '/'.join(f.f_code.co_filename.split(_sep)[-2:]), f.f_lineno, f.f_code.co_name ) for f in islice(frame_iterator(sys._getframe(1 + skip)), length))
python
def format_stack(skip=0, length=6, _sep=os.path.sep): """ Returns a one-line string with the current callstack. """ return ' < '.join("%s:%s:%s" % ( '/'.join(f.f_code.co_filename.split(_sep)[-2:]), f.f_lineno, f.f_code.co_name ) for f in islice(frame_iterator(sys._getframe(1 + skip)), length))
[ "def", "format_stack", "(", "skip", "=", "0", ",", "length", "=", "6", ",", "_sep", "=", "os", ".", "path", ".", "sep", ")", ":", "return", "' < '", ".", "join", "(", "\"%s:%s:%s\"", "%", "(", "'/'", ".", "join", "(", "f", ".", "f_code", ".", "co_filename", ".", "split", "(", "_sep", ")", "[", "-", "2", ":", "]", ")", ",", "f", ".", "f_lineno", ",", "f", ".", "f_code", ".", "co_name", ")", "for", "f", "in", "islice", "(", "frame_iterator", "(", "sys", ".", "_getframe", "(", "1", "+", "skip", ")", ")", ",", "length", ")", ")" ]
Returns a one-line string with the current callstack.
[ "Returns", "a", "one", "-", "line", "string", "with", "the", "current", "callstack", "." ]
train
https://github.com/ionelmc/python-aspectlib/blob/7dba7651ab76160b9ce66c9c11e770f7844e95ec/src/aspectlib/debug.py#L27-L35
ionelmc/python-aspectlib
src/aspectlib/debug.py
log
def log(func=None, stacktrace=10, stacktrace_align=60, attributes=(), module=True, call=True, call_args=True, call_args_repr=repr, result=True, exception=True, exception_repr=repr, result_repr=strip_non_ascii, use_logging='CRITICAL', print_to=None): """ Decorates `func` to have logging. Args func (function): Function to decorate. If missing log returns a partial which you can use as a decorator. stacktrace (int): Number of frames to show. stacktrace_align (int): Column to align the framelist to. attributes (list): List of instance attributes to show, in case the function is a instance method. module (bool): Show the module. call (bool): If ``True``, then show calls. If ``False`` only show the call details on exceptions (if ``exception`` is enabled) (default: ``True``) call_args (bool): If ``True``, then show call arguments. (default: ``True``) call_args_repr (bool): Function to convert one argument to a string. (default: ``repr``) result (bool): If ``True``, then show result. (default: ``True``) exception (bool): If ``True``, then show exceptions. (default: ``True``) exception_repr (function): Function to convert an exception to a string. (default: ``repr``) result_repr (function): Function to convert the result object to a string. (default: ``strip_non_ascii`` - like ``str`` but nonascii characters are replaced with dots.) use_logging (string): Emit log messages with the given loglevel. (default: ``"CRITICAL"``) print_to (fileobject): File object to write to, in case you don't want to use logging module. (default: ``None`` - printing is disabled) Returns: A decorator or a wrapper. Example:: >>> @log(print_to=sys.stdout) ... def a(weird=False): ... if weird: ... raise RuntimeError('BOOM!') >>> a() a() <<< ... a => None >>> try: ... a(weird=True) ... except Exception: ... pass # naughty code! a(weird=True) <<< ... a ~ raised RuntimeError('BOOM!',) You can conveniently use this to logs just errors, or just results, example:: >>> import aspectlib >>> with aspectlib.weave(float, log(call=False, result=False, print_to=sys.stdout)): ... try: ... float('invalid') ... except Exception as e: ... pass # naughty code! float('invalid') <<< ... float ~ raised ValueError(...float...invalid...) This makes debugging naughty code easier. PS: Without the weaving it looks like this:: >>> try: ... log(call=False, result=False, print_to=sys.stdout)(float)('invalid') ... except Exception: ... pass # naughty code! float('invalid') <<< ... float ~ raised ValueError(...float...invalid...) .. versionchanged:: 0.5.0 Renamed `arguments` to `call_args`. Renamed `arguments_repr` to `call_args_repr`. Added `call` option. """ loglevel = use_logging and ( logging._levelNames if hasattr(logging, '_levelNames') else logging._nameToLevel ).get(use_logging, logging.CRITICAL) _missing = object() def dump(buf): try: if use_logging: logger._log(loglevel, buf, ()) if print_to: buf += '\n' print_to.write(buf) except Exception as exc: logger.critical('Failed to log a message: %s', exc, exc_info=True) class __logged__(Aspect): __slots__ = 'cutpoint_function', 'final_function', 'binding', '__name__', '__weakref__' bind = False def __init__(self, cutpoint_function, binding=None): mimic(self, cutpoint_function) self.cutpoint_function = cutpoint_function self.final_function = super(__logged__, self).__call__(cutpoint_function) self.binding = binding def __get__(self, instance, owner): return __logged__(self.cutpoint_function.__get__(instance, owner), instance) def __call__(self, *args, **kwargs): return self.final_function(*args, **kwargs) def advising_function(self, *args, **kwargs): name = self.cutpoint_function.__name__ instance = self.binding if instance is not None: if isinstance(instance, InstanceType): instance_type = instance.__class__ else: instance_type = type(instance) info = [] for key in attributes: if key.endswith('()'): callarg = key = key.rstrip('()') else: callarg = False val = getattr(instance, key, _missing) if val is not _missing and key != name: info.append(' %s=%s' % ( key, call_args_repr(val() if callarg else val) )) sig = buf = '{%s%s%s}.%s' % ( instance_type.__module__ + '.' if module else '', instance_type.__name__, ''.join(info), name ) else: sig = buf = name if call_args: buf += '(%s%s)' % ( ', '.join(repr(i) for i in (args if call_args is True else args[:call_args])), ((', ' if args else '') + ', '.join('%s=%r' % i for i in kwargs.items())) if kwargs and call_args is True else '', ) if stacktrace: buf = ("%%-%ds <<< %%s" % stacktrace_align) % (buf, format_stack(skip=1, length=stacktrace)) if call: dump(buf) try: res = yield except Exception as exc: if exception: if not call: dump(buf) dump('%s ~ raised %s' % (sig, exception_repr(exc))) raise if result: dump('%s => %s' % (sig, result_repr(res))) if func: return __logged__(func) else: return __logged__
python
def log(func=None, stacktrace=10, stacktrace_align=60, attributes=(), module=True, call=True, call_args=True, call_args_repr=repr, result=True, exception=True, exception_repr=repr, result_repr=strip_non_ascii, use_logging='CRITICAL', print_to=None): """ Decorates `func` to have logging. Args func (function): Function to decorate. If missing log returns a partial which you can use as a decorator. stacktrace (int): Number of frames to show. stacktrace_align (int): Column to align the framelist to. attributes (list): List of instance attributes to show, in case the function is a instance method. module (bool): Show the module. call (bool): If ``True``, then show calls. If ``False`` only show the call details on exceptions (if ``exception`` is enabled) (default: ``True``) call_args (bool): If ``True``, then show call arguments. (default: ``True``) call_args_repr (bool): Function to convert one argument to a string. (default: ``repr``) result (bool): If ``True``, then show result. (default: ``True``) exception (bool): If ``True``, then show exceptions. (default: ``True``) exception_repr (function): Function to convert an exception to a string. (default: ``repr``) result_repr (function): Function to convert the result object to a string. (default: ``strip_non_ascii`` - like ``str`` but nonascii characters are replaced with dots.) use_logging (string): Emit log messages with the given loglevel. (default: ``"CRITICAL"``) print_to (fileobject): File object to write to, in case you don't want to use logging module. (default: ``None`` - printing is disabled) Returns: A decorator or a wrapper. Example:: >>> @log(print_to=sys.stdout) ... def a(weird=False): ... if weird: ... raise RuntimeError('BOOM!') >>> a() a() <<< ... a => None >>> try: ... a(weird=True) ... except Exception: ... pass # naughty code! a(weird=True) <<< ... a ~ raised RuntimeError('BOOM!',) You can conveniently use this to logs just errors, or just results, example:: >>> import aspectlib >>> with aspectlib.weave(float, log(call=False, result=False, print_to=sys.stdout)): ... try: ... float('invalid') ... except Exception as e: ... pass # naughty code! float('invalid') <<< ... float ~ raised ValueError(...float...invalid...) This makes debugging naughty code easier. PS: Without the weaving it looks like this:: >>> try: ... log(call=False, result=False, print_to=sys.stdout)(float)('invalid') ... except Exception: ... pass # naughty code! float('invalid') <<< ... float ~ raised ValueError(...float...invalid...) .. versionchanged:: 0.5.0 Renamed `arguments` to `call_args`. Renamed `arguments_repr` to `call_args_repr`. Added `call` option. """ loglevel = use_logging and ( logging._levelNames if hasattr(logging, '_levelNames') else logging._nameToLevel ).get(use_logging, logging.CRITICAL) _missing = object() def dump(buf): try: if use_logging: logger._log(loglevel, buf, ()) if print_to: buf += '\n' print_to.write(buf) except Exception as exc: logger.critical('Failed to log a message: %s', exc, exc_info=True) class __logged__(Aspect): __slots__ = 'cutpoint_function', 'final_function', 'binding', '__name__', '__weakref__' bind = False def __init__(self, cutpoint_function, binding=None): mimic(self, cutpoint_function) self.cutpoint_function = cutpoint_function self.final_function = super(__logged__, self).__call__(cutpoint_function) self.binding = binding def __get__(self, instance, owner): return __logged__(self.cutpoint_function.__get__(instance, owner), instance) def __call__(self, *args, **kwargs): return self.final_function(*args, **kwargs) def advising_function(self, *args, **kwargs): name = self.cutpoint_function.__name__ instance = self.binding if instance is not None: if isinstance(instance, InstanceType): instance_type = instance.__class__ else: instance_type = type(instance) info = [] for key in attributes: if key.endswith('()'): callarg = key = key.rstrip('()') else: callarg = False val = getattr(instance, key, _missing) if val is not _missing and key != name: info.append(' %s=%s' % ( key, call_args_repr(val() if callarg else val) )) sig = buf = '{%s%s%s}.%s' % ( instance_type.__module__ + '.' if module else '', instance_type.__name__, ''.join(info), name ) else: sig = buf = name if call_args: buf += '(%s%s)' % ( ', '.join(repr(i) for i in (args if call_args is True else args[:call_args])), ((', ' if args else '') + ', '.join('%s=%r' % i for i in kwargs.items())) if kwargs and call_args is True else '', ) if stacktrace: buf = ("%%-%ds <<< %%s" % stacktrace_align) % (buf, format_stack(skip=1, length=stacktrace)) if call: dump(buf) try: res = yield except Exception as exc: if exception: if not call: dump(buf) dump('%s ~ raised %s' % (sig, exception_repr(exc))) raise if result: dump('%s => %s' % (sig, result_repr(res))) if func: return __logged__(func) else: return __logged__
[ "def", "log", "(", "func", "=", "None", ",", "stacktrace", "=", "10", ",", "stacktrace_align", "=", "60", ",", "attributes", "=", "(", ")", ",", "module", "=", "True", ",", "call", "=", "True", ",", "call_args", "=", "True", ",", "call_args_repr", "=", "repr", ",", "result", "=", "True", ",", "exception", "=", "True", ",", "exception_repr", "=", "repr", ",", "result_repr", "=", "strip_non_ascii", ",", "use_logging", "=", "'CRITICAL'", ",", "print_to", "=", "None", ")", ":", "loglevel", "=", "use_logging", "and", "(", "logging", ".", "_levelNames", "if", "hasattr", "(", "logging", ",", "'_levelNames'", ")", "else", "logging", ".", "_nameToLevel", ")", ".", "get", "(", "use_logging", ",", "logging", ".", "CRITICAL", ")", "_missing", "=", "object", "(", ")", "def", "dump", "(", "buf", ")", ":", "try", ":", "if", "use_logging", ":", "logger", ".", "_log", "(", "loglevel", ",", "buf", ",", "(", ")", ")", "if", "print_to", ":", "buf", "+=", "'\\n'", "print_to", ".", "write", "(", "buf", ")", "except", "Exception", "as", "exc", ":", "logger", ".", "critical", "(", "'Failed to log a message: %s'", ",", "exc", ",", "exc_info", "=", "True", ")", "class", "__logged__", "(", "Aspect", ")", ":", "__slots__", "=", "'cutpoint_function'", ",", "'final_function'", ",", "'binding'", ",", "'__name__'", ",", "'__weakref__'", "bind", "=", "False", "def", "__init__", "(", "self", ",", "cutpoint_function", ",", "binding", "=", "None", ")", ":", "mimic", "(", "self", ",", "cutpoint_function", ")", "self", ".", "cutpoint_function", "=", "cutpoint_function", "self", ".", "final_function", "=", "super", "(", "__logged__", ",", "self", ")", ".", "__call__", "(", "cutpoint_function", ")", "self", ".", "binding", "=", "binding", "def", "__get__", "(", "self", ",", "instance", ",", "owner", ")", ":", "return", "__logged__", "(", "self", ".", "cutpoint_function", ".", "__get__", "(", "instance", ",", "owner", ")", ",", "instance", ")", "def", "__call__", "(", "self", ",", "*", "args", ",", "*", "*", "kwargs", ")", ":", "return", "self", ".", "final_function", "(", "*", "args", ",", "*", "*", "kwargs", ")", "def", "advising_function", "(", "self", ",", "*", "args", ",", "*", "*", "kwargs", ")", ":", "name", "=", "self", ".", "cutpoint_function", ".", "__name__", "instance", "=", "self", ".", "binding", "if", "instance", "is", "not", "None", ":", "if", "isinstance", "(", "instance", ",", "InstanceType", ")", ":", "instance_type", "=", "instance", ".", "__class__", "else", ":", "instance_type", "=", "type", "(", "instance", ")", "info", "=", "[", "]", "for", "key", "in", "attributes", ":", "if", "key", ".", "endswith", "(", "'()'", ")", ":", "callarg", "=", "key", "=", "key", ".", "rstrip", "(", "'()'", ")", "else", ":", "callarg", "=", "False", "val", "=", "getattr", "(", "instance", ",", "key", ",", "_missing", ")", "if", "val", "is", "not", "_missing", "and", "key", "!=", "name", ":", "info", ".", "append", "(", "' %s=%s'", "%", "(", "key", ",", "call_args_repr", "(", "val", "(", ")", "if", "callarg", "else", "val", ")", ")", ")", "sig", "=", "buf", "=", "'{%s%s%s}.%s'", "%", "(", "instance_type", ".", "__module__", "+", "'.'", "if", "module", "else", "''", ",", "instance_type", ".", "__name__", ",", "''", ".", "join", "(", "info", ")", ",", "name", ")", "else", ":", "sig", "=", "buf", "=", "name", "if", "call_args", ":", "buf", "+=", "'(%s%s)'", "%", "(", "', '", ".", "join", "(", "repr", "(", "i", ")", "for", "i", "in", "(", "args", "if", "call_args", "is", "True", "else", "args", "[", ":", "call_args", "]", ")", ")", ",", "(", "(", "', '", "if", "args", "else", "''", ")", "+", "', '", ".", "join", "(", "'%s=%r'", "%", "i", "for", "i", "in", "kwargs", ".", "items", "(", ")", ")", ")", "if", "kwargs", "and", "call_args", "is", "True", "else", "''", ",", ")", "if", "stacktrace", ":", "buf", "=", "(", "\"%%-%ds <<< %%s\"", "%", "stacktrace_align", ")", "%", "(", "buf", ",", "format_stack", "(", "skip", "=", "1", ",", "length", "=", "stacktrace", ")", ")", "if", "call", ":", "dump", "(", "buf", ")", "try", ":", "res", "=", "yield", "except", "Exception", "as", "exc", ":", "if", "exception", ":", "if", "not", "call", ":", "dump", "(", "buf", ")", "dump", "(", "'%s ~ raised %s'", "%", "(", "sig", ",", "exception_repr", "(", "exc", ")", ")", ")", "raise", "if", "result", ":", "dump", "(", "'%s => %s'", "%", "(", "sig", ",", "result_repr", "(", "res", ")", ")", ")", "if", "func", ":", "return", "__logged__", "(", "func", ")", "else", ":", "return", "__logged__" ]
Decorates `func` to have logging. Args func (function): Function to decorate. If missing log returns a partial which you can use as a decorator. stacktrace (int): Number of frames to show. stacktrace_align (int): Column to align the framelist to. attributes (list): List of instance attributes to show, in case the function is a instance method. module (bool): Show the module. call (bool): If ``True``, then show calls. If ``False`` only show the call details on exceptions (if ``exception`` is enabled) (default: ``True``) call_args (bool): If ``True``, then show call arguments. (default: ``True``) call_args_repr (bool): Function to convert one argument to a string. (default: ``repr``) result (bool): If ``True``, then show result. (default: ``True``) exception (bool): If ``True``, then show exceptions. (default: ``True``) exception_repr (function): Function to convert an exception to a string. (default: ``repr``) result_repr (function): Function to convert the result object to a string. (default: ``strip_non_ascii`` - like ``str`` but nonascii characters are replaced with dots.) use_logging (string): Emit log messages with the given loglevel. (default: ``"CRITICAL"``) print_to (fileobject): File object to write to, in case you don't want to use logging module. (default: ``None`` - printing is disabled) Returns: A decorator or a wrapper. Example:: >>> @log(print_to=sys.stdout) ... def a(weird=False): ... if weird: ... raise RuntimeError('BOOM!') >>> a() a() <<< ... a => None >>> try: ... a(weird=True) ... except Exception: ... pass # naughty code! a(weird=True) <<< ... a ~ raised RuntimeError('BOOM!',) You can conveniently use this to logs just errors, or just results, example:: >>> import aspectlib >>> with aspectlib.weave(float, log(call=False, result=False, print_to=sys.stdout)): ... try: ... float('invalid') ... except Exception as e: ... pass # naughty code! float('invalid') <<< ... float ~ raised ValueError(...float...invalid...) This makes debugging naughty code easier. PS: Without the weaving it looks like this:: >>> try: ... log(call=False, result=False, print_to=sys.stdout)(float)('invalid') ... except Exception: ... pass # naughty code! float('invalid') <<< ... float ~ raised ValueError(...float...invalid...) .. versionchanged:: 0.5.0 Renamed `arguments` to `call_args`. Renamed `arguments_repr` to `call_args_repr`. Added `call` option.
[ "Decorates", "func", "to", "have", "logging", "." ]
train
https://github.com/ionelmc/python-aspectlib/blob/7dba7651ab76160b9ce66c9c11e770f7844e95ec/src/aspectlib/debug.py#L48-L233
3ll3d00d/vibe
backend/src/recorder/app.py
wireHandlers
def wireHandlers(cfg): """ If the device is configured to run against a remote server, ping that device on a scheduled basis with our current state. :param cfg: the config object. :return: """ logger = logging.getLogger('recorder') httpPoster = cfg.handlers.get('remote') csvLogger = cfg.handlers.get('local') activeHandler = None if httpPoster is None: if csvLogger is None: logger.warning("App is running with discard handler only, ALL DATA WILL BE DISCARDED!!!") else: logger.info("App is running in standalone mode, logging data to local filesystem") activeHandler = csvLogger else: logger.info("App is running against remote server, logging data to " + httpPoster.target) activeHandler = httpPoster heartbeater.serverURL = httpPoster.target heartbeater.ping() if activeHandler is not None: for device in cfg.recordingDevices.values(): if activeHandler is httpPoster: httpPoster.deviceName = device.name copied = copy.copy(activeHandler) device.dataHandler = copied if not cfg.useAsyncHandlers else AsyncHandler('recorder', copied)
python
def wireHandlers(cfg): """ If the device is configured to run against a remote server, ping that device on a scheduled basis with our current state. :param cfg: the config object. :return: """ logger = logging.getLogger('recorder') httpPoster = cfg.handlers.get('remote') csvLogger = cfg.handlers.get('local') activeHandler = None if httpPoster is None: if csvLogger is None: logger.warning("App is running with discard handler only, ALL DATA WILL BE DISCARDED!!!") else: logger.info("App is running in standalone mode, logging data to local filesystem") activeHandler = csvLogger else: logger.info("App is running against remote server, logging data to " + httpPoster.target) activeHandler = httpPoster heartbeater.serverURL = httpPoster.target heartbeater.ping() if activeHandler is not None: for device in cfg.recordingDevices.values(): if activeHandler is httpPoster: httpPoster.deviceName = device.name copied = copy.copy(activeHandler) device.dataHandler = copied if not cfg.useAsyncHandlers else AsyncHandler('recorder', copied)
[ "def", "wireHandlers", "(", "cfg", ")", ":", "logger", "=", "logging", ".", "getLogger", "(", "'recorder'", ")", "httpPoster", "=", "cfg", ".", "handlers", ".", "get", "(", "'remote'", ")", "csvLogger", "=", "cfg", ".", "handlers", ".", "get", "(", "'local'", ")", "activeHandler", "=", "None", "if", "httpPoster", "is", "None", ":", "if", "csvLogger", "is", "None", ":", "logger", ".", "warning", "(", "\"App is running with discard handler only, ALL DATA WILL BE DISCARDED!!!\"", ")", "else", ":", "logger", ".", "info", "(", "\"App is running in standalone mode, logging data to local filesystem\"", ")", "activeHandler", "=", "csvLogger", "else", ":", "logger", ".", "info", "(", "\"App is running against remote server, logging data to \"", "+", "httpPoster", ".", "target", ")", "activeHandler", "=", "httpPoster", "heartbeater", ".", "serverURL", "=", "httpPoster", ".", "target", "heartbeater", ".", "ping", "(", ")", "if", "activeHandler", "is", "not", "None", ":", "for", "device", "in", "cfg", ".", "recordingDevices", ".", "values", "(", ")", ":", "if", "activeHandler", "is", "httpPoster", ":", "httpPoster", ".", "deviceName", "=", "device", ".", "name", "copied", "=", "copy", ".", "copy", "(", "activeHandler", ")", "device", ".", "dataHandler", "=", "copied", "if", "not", "cfg", ".", "useAsyncHandlers", "else", "AsyncHandler", "(", "'recorder'", ",", "copied", ")" ]
If the device is configured to run against a remote server, ping that device on a scheduled basis with our current state. :param cfg: the config object. :return:
[ "If", "the", "device", "is", "configured", "to", "run", "against", "a", "remote", "server", "ping", "that", "device", "on", "a", "scheduled", "basis", "with", "our", "current", "state", ".", ":", "param", "cfg", ":", "the", "config", "object", ".", ":", "return", ":" ]
train
https://github.com/3ll3d00d/vibe/blob/124b029f13ac746723e92cb47e9cb56edd2e54b5/backend/src/recorder/app.py#L55-L83
3ll3d00d/vibe
backend/src/recorder/app.py
main
def main(args=None): """ The main routine. """ cfg.configureLogger() wireHandlers(cfg) # get config from a flask standard place not our config yml app.run(debug=cfg.runInDebug(), host='0.0.0.0', port=cfg.getPort())
python
def main(args=None): """ The main routine. """ cfg.configureLogger() wireHandlers(cfg) # get config from a flask standard place not our config yml app.run(debug=cfg.runInDebug(), host='0.0.0.0', port=cfg.getPort())
[ "def", "main", "(", "args", "=", "None", ")", ":", "cfg", ".", "configureLogger", "(", ")", "wireHandlers", "(", "cfg", ")", "# get config from a flask standard place not our config yml", "app", ".", "run", "(", "debug", "=", "cfg", ".", "runInDebug", "(", ")", ",", "host", "=", "'0.0.0.0'", ",", "port", "=", "cfg", ".", "getPort", "(", ")", ")" ]
The main routine.
[ "The", "main", "routine", "." ]
train
https://github.com/3ll3d00d/vibe/blob/124b029f13ac746723e92cb47e9cb56edd2e54b5/backend/src/recorder/app.py#L86-L91
3ll3d00d/vibe
backend/src/recorder/resources/measurements.py
Measurements.get
def get(self, deviceId): """ lists all known active measurements. """ measurementsByName = self.measurements.get(deviceId) if measurementsByName is None: return [] else: return list(measurementsByName.values())
python
def get(self, deviceId): """ lists all known active measurements. """ measurementsByName = self.measurements.get(deviceId) if measurementsByName is None: return [] else: return list(measurementsByName.values())
[ "def", "get", "(", "self", ",", "deviceId", ")", ":", "measurementsByName", "=", "self", ".", "measurements", ".", "get", "(", "deviceId", ")", "if", "measurementsByName", "is", "None", ":", "return", "[", "]", "else", ":", "return", "list", "(", "measurementsByName", ".", "values", "(", ")", ")" ]
lists all known active measurements.
[ "lists", "all", "known", "active", "measurements", "." ]
train
https://github.com/3ll3d00d/vibe/blob/124b029f13ac746723e92cb47e9cb56edd2e54b5/backend/src/recorder/resources/measurements.py#L25-L33
3ll3d00d/vibe
backend/src/recorder/resources/measurements.py
Measurement.get
def get(self, deviceId, measurementId): """ details the specific measurement. """ record = self.measurements.get(deviceId) if record is not None: return record.get(measurementId) return None
python
def get(self, deviceId, measurementId): """ details the specific measurement. """ record = self.measurements.get(deviceId) if record is not None: return record.get(measurementId) return None
[ "def", "get", "(", "self", ",", "deviceId", ",", "measurementId", ")", ":", "record", "=", "self", ".", "measurements", ".", "get", "(", "deviceId", ")", "if", "record", "is", "not", "None", ":", "return", "record", ".", "get", "(", "measurementId", ")", "return", "None" ]
details the specific measurement.
[ "details", "the", "specific", "measurement", "." ]
train
https://github.com/3ll3d00d/vibe/blob/124b029f13ac746723e92cb47e9cb56edd2e54b5/backend/src/recorder/resources/measurements.py#L42-L49
3ll3d00d/vibe
backend/src/recorder/resources/measurements.py
Measurement.put
def put(self, deviceId, measurementId): """ Schedules a new measurement at the specified time. :param deviceId: the device to measure. :param measurementId: the name of the measurement. :return: 200 if it was scheduled, 400 if the device is busy, 500 if the device is bad. """ record = self.measurements.get(deviceId) if record is not None: measurement = record.get(measurementId) if measurement is not None: if len([x.name for x in measurement.statuses if x.name is 'COMPLETE' or x.name is 'FAILED']) > 0: logger.info('Overwriting existing completed measurement ' + x.name) measurement = None if measurement is None: logger.info('Initiating measurement ' + measurementId) measurement = ScheduledMeasurement(measurementId, self.recordingDevices.get(deviceId)) body = request.get_json() duration_ = body['duration'] def _cleanup(): logger.info('Removing ' + measurementId + ' from ' + deviceId) record.pop(measurementId) measurement.schedule(duration_, at=body.get('at'), delay=body.get('delay'), callback=_cleanup) # a quick hack to enable the measurement to be cleaned up by the ScheduledMeasurement record[measurementId] = measurement return measurement, 200 else: return measurement, 400 else: return 'unknown device ' + deviceId, 400
python
def put(self, deviceId, measurementId): """ Schedules a new measurement at the specified time. :param deviceId: the device to measure. :param measurementId: the name of the measurement. :return: 200 if it was scheduled, 400 if the device is busy, 500 if the device is bad. """ record = self.measurements.get(deviceId) if record is not None: measurement = record.get(measurementId) if measurement is not None: if len([x.name for x in measurement.statuses if x.name is 'COMPLETE' or x.name is 'FAILED']) > 0: logger.info('Overwriting existing completed measurement ' + x.name) measurement = None if measurement is None: logger.info('Initiating measurement ' + measurementId) measurement = ScheduledMeasurement(measurementId, self.recordingDevices.get(deviceId)) body = request.get_json() duration_ = body['duration'] def _cleanup(): logger.info('Removing ' + measurementId + ' from ' + deviceId) record.pop(measurementId) measurement.schedule(duration_, at=body.get('at'), delay=body.get('delay'), callback=_cleanup) # a quick hack to enable the measurement to be cleaned up by the ScheduledMeasurement record[measurementId] = measurement return measurement, 200 else: return measurement, 400 else: return 'unknown device ' + deviceId, 400
[ "def", "put", "(", "self", ",", "deviceId", ",", "measurementId", ")", ":", "record", "=", "self", ".", "measurements", ".", "get", "(", "deviceId", ")", "if", "record", "is", "not", "None", ":", "measurement", "=", "record", ".", "get", "(", "measurementId", ")", "if", "measurement", "is", "not", "None", ":", "if", "len", "(", "[", "x", ".", "name", "for", "x", "in", "measurement", ".", "statuses", "if", "x", ".", "name", "is", "'COMPLETE'", "or", "x", ".", "name", "is", "'FAILED'", "]", ")", ">", "0", ":", "logger", ".", "info", "(", "'Overwriting existing completed measurement '", "+", "x", ".", "name", ")", "measurement", "=", "None", "if", "measurement", "is", "None", ":", "logger", ".", "info", "(", "'Initiating measurement '", "+", "measurementId", ")", "measurement", "=", "ScheduledMeasurement", "(", "measurementId", ",", "self", ".", "recordingDevices", ".", "get", "(", "deviceId", ")", ")", "body", "=", "request", ".", "get_json", "(", ")", "duration_", "=", "body", "[", "'duration'", "]", "def", "_cleanup", "(", ")", ":", "logger", ".", "info", "(", "'Removing '", "+", "measurementId", "+", "' from '", "+", "deviceId", ")", "record", ".", "pop", "(", "measurementId", ")", "measurement", ".", "schedule", "(", "duration_", ",", "at", "=", "body", ".", "get", "(", "'at'", ")", ",", "delay", "=", "body", ".", "get", "(", "'delay'", ")", ",", "callback", "=", "_cleanup", ")", "# a quick hack to enable the measurement to be cleaned up by the ScheduledMeasurement", "record", "[", "measurementId", "]", "=", "measurement", "return", "measurement", ",", "200", "else", ":", "return", "measurement", ",", "400", "else", ":", "return", "'unknown device '", "+", "deviceId", ",", "400" ]
Schedules a new measurement at the specified time. :param deviceId: the device to measure. :param measurementId: the name of the measurement. :return: 200 if it was scheduled, 400 if the device is busy, 500 if the device is bad.
[ "Schedules", "a", "new", "measurement", "at", "the", "specified", "time", ".", ":", "param", "deviceId", ":", "the", "device", "to", "measure", ".", ":", "param", "measurementId", ":", "the", "name", "of", "the", "measurement", ".", ":", "return", ":", "200", "if", "it", "was", "scheduled", "400", "if", "the", "device", "is", "busy", "500", "if", "the", "device", "is", "bad", "." ]
train
https://github.com/3ll3d00d/vibe/blob/124b029f13ac746723e92cb47e9cb56edd2e54b5/backend/src/recorder/resources/measurements.py#L52-L81
3ll3d00d/vibe
backend/src/recorder/resources/measurements.py
Measurement.delete
def delete(self, deviceId, measurementId): """ Deletes a stored measurement. :param deviceId: the device to measure. :param measurementId: the name of the measurement. :return: 200 if it was deleted, 400 if no such measurement (or device). """ record = self.measurements.get(deviceId) if record is not None: popped = record.pop(measurementId, None) return popped, 200 if popped else 400 return None, 400
python
def delete(self, deviceId, measurementId): """ Deletes a stored measurement. :param deviceId: the device to measure. :param measurementId: the name of the measurement. :return: 200 if it was deleted, 400 if no such measurement (or device). """ record = self.measurements.get(deviceId) if record is not None: popped = record.pop(measurementId, None) return popped, 200 if popped else 400 return None, 400
[ "def", "delete", "(", "self", ",", "deviceId", ",", "measurementId", ")", ":", "record", "=", "self", ".", "measurements", ".", "get", "(", "deviceId", ")", "if", "record", "is", "not", "None", ":", "popped", "=", "record", ".", "pop", "(", "measurementId", ",", "None", ")", "return", "popped", ",", "200", "if", "popped", "else", "400", "return", "None", ",", "400" ]
Deletes a stored measurement. :param deviceId: the device to measure. :param measurementId: the name of the measurement. :return: 200 if it was deleted, 400 if no such measurement (or device).
[ "Deletes", "a", "stored", "measurement", ".", ":", "param", "deviceId", ":", "the", "device", "to", "measure", ".", ":", "param", "measurementId", ":", "the", "name", "of", "the", "measurement", ".", ":", "return", ":", "200", "if", "it", "was", "deleted", "400", "if", "no", "such", "measurement", "(", "or", "device", ")", "." ]
train
https://github.com/3ll3d00d/vibe/blob/124b029f13ac746723e92cb47e9cb56edd2e54b5/backend/src/recorder/resources/measurements.py#L84-L95
3ll3d00d/vibe
backend/src/recorder/resources/measurements.py
AbortMeasurement.get
def get(self, deviceId, measurementId): """ signals a stop for the given measurement. :param deviceId: the device to measure. :param measurementId: the name of the measurement. :return: 200 if stop is signalled, 400 if it doesn't exist or is not running. """ record = self.measurements.get(deviceId) if record is not None: measurement = record.get(measurementId) if measurement.recording: device = self.recordingDevices.get(deviceId) device.signalStop() return measurement, 200 else: return measurement, 400 return '', 400
python
def get(self, deviceId, measurementId): """ signals a stop for the given measurement. :param deviceId: the device to measure. :param measurementId: the name of the measurement. :return: 200 if stop is signalled, 400 if it doesn't exist or is not running. """ record = self.measurements.get(deviceId) if record is not None: measurement = record.get(measurementId) if measurement.recording: device = self.recordingDevices.get(deviceId) device.signalStop() return measurement, 200 else: return measurement, 400 return '', 400
[ "def", "get", "(", "self", ",", "deviceId", ",", "measurementId", ")", ":", "record", "=", "self", ".", "measurements", ".", "get", "(", "deviceId", ")", "if", "record", "is", "not", "None", ":", "measurement", "=", "record", ".", "get", "(", "measurementId", ")", "if", "measurement", ".", "recording", ":", "device", "=", "self", ".", "recordingDevices", ".", "get", "(", "deviceId", ")", "device", ".", "signalStop", "(", ")", "return", "measurement", ",", "200", "else", ":", "return", "measurement", ",", "400", "return", "''", ",", "400" ]
signals a stop for the given measurement. :param deviceId: the device to measure. :param measurementId: the name of the measurement. :return: 200 if stop is signalled, 400 if it doesn't exist or is not running.
[ "signals", "a", "stop", "for", "the", "given", "measurement", ".", ":", "param", "deviceId", ":", "the", "device", "to", "measure", ".", ":", "param", "measurementId", ":", "the", "name", "of", "the", "measurement", ".", ":", "return", ":", "200", "if", "stop", "is", "signalled", "400", "if", "it", "doesn", "t", "exist", "or", "is", "not", "running", "." ]
train
https://github.com/3ll3d00d/vibe/blob/124b029f13ac746723e92cb47e9cb56edd2e54b5/backend/src/recorder/resources/measurements.py#L104-L120
3ll3d00d/vibe
backend/src/recorder/resources/measurements.py
ScheduledMeasurement.schedule
def schedule(self, duration, at=None, delay=None, callback=None): """ schedules the measurement (to execute asynchronously). :param duration: how long to run for. :param at: the time to start at. :param delay: the time to wait til starting (use at or delay). :param callback: a callback. :return: nothing. """ delay = self.calculateDelay(at, delay) self.callback = callback logger.info('Initiating measurement ' + self.name + ' for ' + str(duration) + 's in ' + str(delay) + 's') self.statuses.append({'name': ScheduledMeasurementStatus.SCHEDULED.name, 'time': datetime.utcnow()}) threading.Timer(delay, self.execute, [duration]).start()
python
def schedule(self, duration, at=None, delay=None, callback=None): """ schedules the measurement (to execute asynchronously). :param duration: how long to run for. :param at: the time to start at. :param delay: the time to wait til starting (use at or delay). :param callback: a callback. :return: nothing. """ delay = self.calculateDelay(at, delay) self.callback = callback logger.info('Initiating measurement ' + self.name + ' for ' + str(duration) + 's in ' + str(delay) + 's') self.statuses.append({'name': ScheduledMeasurementStatus.SCHEDULED.name, 'time': datetime.utcnow()}) threading.Timer(delay, self.execute, [duration]).start()
[ "def", "schedule", "(", "self", ",", "duration", ",", "at", "=", "None", ",", "delay", "=", "None", ",", "callback", "=", "None", ")", ":", "delay", "=", "self", ".", "calculateDelay", "(", "at", ",", "delay", ")", "self", ".", "callback", "=", "callback", "logger", ".", "info", "(", "'Initiating measurement '", "+", "self", ".", "name", "+", "' for '", "+", "str", "(", "duration", ")", "+", "'s in '", "+", "str", "(", "delay", ")", "+", "'s'", ")", "self", ".", "statuses", ".", "append", "(", "{", "'name'", ":", "ScheduledMeasurementStatus", ".", "SCHEDULED", ".", "name", ",", "'time'", ":", "datetime", ".", "utcnow", "(", ")", "}", ")", "threading", ".", "Timer", "(", "delay", ",", "self", ".", "execute", ",", "[", "duration", "]", ")", ".", "start", "(", ")" ]
schedules the measurement (to execute asynchronously). :param duration: how long to run for. :param at: the time to start at. :param delay: the time to wait til starting (use at or delay). :param callback: a callback. :return: nothing.
[ "schedules", "the", "measurement", "(", "to", "execute", "asynchronously", ")", ".", ":", "param", "duration", ":", "how", "long", "to", "run", "for", ".", ":", "param", "at", ":", "the", "time", "to", "start", "at", ".", ":", "param", "delay", ":", "the", "time", "to", "wait", "til", "starting", "(", "use", "at", "or", "delay", ")", ".", ":", "param", "callback", ":", "a", "callback", ".", ":", "return", ":", "nothing", "." ]
train
https://github.com/3ll3d00d/vibe/blob/124b029f13ac746723e92cb47e9cb56edd2e54b5/backend/src/recorder/resources/measurements.py#L148-L161
3ll3d00d/vibe
backend/src/recorder/resources/measurements.py
ScheduledMeasurement.execute
def execute(self, duration): """ Executes the measurement, recording the event status. :param duration: the time to run for. :return: nothing. """ self.statuses.append({'name': ScheduledMeasurementStatus.RUNNING.name, 'time': datetime.utcnow()}) try: self.recording = True self.device.start(self.name, durationInSeconds=duration) finally: self.recording = False if self.device.status == RecordingDeviceStatus.FAILED: self.statuses.append({'name': ScheduledMeasurementStatus.FAILED.name, 'time': datetime.utcnow(), 'reason': self.device.failureCode}) else: self.statuses.append({'name': ScheduledMeasurementStatus.COMPLETE.name, 'time': datetime.utcnow()}) # this is a bit of a hack, need to remove this at some point by refactoring the way measurements are stored if self.callback is not None: self.callback()
python
def execute(self, duration): """ Executes the measurement, recording the event status. :param duration: the time to run for. :return: nothing. """ self.statuses.append({'name': ScheduledMeasurementStatus.RUNNING.name, 'time': datetime.utcnow()}) try: self.recording = True self.device.start(self.name, durationInSeconds=duration) finally: self.recording = False if self.device.status == RecordingDeviceStatus.FAILED: self.statuses.append({'name': ScheduledMeasurementStatus.FAILED.name, 'time': datetime.utcnow(), 'reason': self.device.failureCode}) else: self.statuses.append({'name': ScheduledMeasurementStatus.COMPLETE.name, 'time': datetime.utcnow()}) # this is a bit of a hack, need to remove this at some point by refactoring the way measurements are stored if self.callback is not None: self.callback()
[ "def", "execute", "(", "self", ",", "duration", ")", ":", "self", ".", "statuses", ".", "append", "(", "{", "'name'", ":", "ScheduledMeasurementStatus", ".", "RUNNING", ".", "name", ",", "'time'", ":", "datetime", ".", "utcnow", "(", ")", "}", ")", "try", ":", "self", ".", "recording", "=", "True", "self", ".", "device", ".", "start", "(", "self", ".", "name", ",", "durationInSeconds", "=", "duration", ")", "finally", ":", "self", ".", "recording", "=", "False", "if", "self", ".", "device", ".", "status", "==", "RecordingDeviceStatus", ".", "FAILED", ":", "self", ".", "statuses", ".", "append", "(", "{", "'name'", ":", "ScheduledMeasurementStatus", ".", "FAILED", ".", "name", ",", "'time'", ":", "datetime", ".", "utcnow", "(", ")", ",", "'reason'", ":", "self", ".", "device", ".", "failureCode", "}", ")", "else", ":", "self", ".", "statuses", ".", "append", "(", "{", "'name'", ":", "ScheduledMeasurementStatus", ".", "COMPLETE", ".", "name", ",", "'time'", ":", "datetime", ".", "utcnow", "(", ")", "}", ")", "# this is a bit of a hack, need to remove this at some point by refactoring the way measurements are stored", "if", "self", ".", "callback", "is", "not", "None", ":", "self", ".", "callback", "(", ")" ]
Executes the measurement, recording the event status. :param duration: the time to run for. :return: nothing.
[ "Executes", "the", "measurement", "recording", "the", "event", "status", ".", ":", "param", "duration", ":", "the", "time", "to", "run", "for", ".", ":", "return", ":", "nothing", "." ]
train
https://github.com/3ll3d00d/vibe/blob/124b029f13ac746723e92cb47e9cb56edd2e54b5/backend/src/recorder/resources/measurements.py#L163-L183
3ll3d00d/vibe
backend/src/recorder/resources/measurements.py
ScheduledMeasurement.calculateDelay
def calculateDelay(self, at, delay): """ Creates the delay from now til the specified start time, uses "at" if available. :param at: the start time in %a %b %d %H:%M:%S %Y format. :param delay: the delay from now til start. :return: the delay. """ if at is not None: return max((datetime.strptime(at, DATETIME_FORMAT) - datetime.utcnow()).total_seconds(), 0) elif delay is not None: return delay else: return 0
python
def calculateDelay(self, at, delay): """ Creates the delay from now til the specified start time, uses "at" if available. :param at: the start time in %a %b %d %H:%M:%S %Y format. :param delay: the delay from now til start. :return: the delay. """ if at is not None: return max((datetime.strptime(at, DATETIME_FORMAT) - datetime.utcnow()).total_seconds(), 0) elif delay is not None: return delay else: return 0
[ "def", "calculateDelay", "(", "self", ",", "at", ",", "delay", ")", ":", "if", "at", "is", "not", "None", ":", "return", "max", "(", "(", "datetime", ".", "strptime", "(", "at", ",", "DATETIME_FORMAT", ")", "-", "datetime", ".", "utcnow", "(", ")", ")", ".", "total_seconds", "(", ")", ",", "0", ")", "elif", "delay", "is", "not", "None", ":", "return", "delay", "else", ":", "return", "0" ]
Creates the delay from now til the specified start time, uses "at" if available. :param at: the start time in %a %b %d %H:%M:%S %Y format. :param delay: the delay from now til start. :return: the delay.
[ "Creates", "the", "delay", "from", "now", "til", "the", "specified", "start", "time", "uses", "at", "if", "available", ".", ":", "param", "at", ":", "the", "start", "time", "in", "%a", "%b", "%d", "%H", ":", "%M", ":", "%S", "%Y", "format", ".", ":", "param", "delay", ":", "the", "delay", "from", "now", "til", "start", ".", ":", "return", ":", "the", "delay", "." ]
train
https://github.com/3ll3d00d/vibe/blob/124b029f13ac746723e92cb47e9cb56edd2e54b5/backend/src/recorder/resources/measurements.py#L185-L197
spacetelescope/stsci.tools
lib/stsci/tools/mputil.py
launch_and_wait
def launch_and_wait(mp_proc_list, pool_size): """ Given a list of multiprocessing.Process objects which have not yet been started, this function launches them and blocks until the last finishes. This makes sure that only <pool_size> processes are ever working at any one time (this number does not include the main process which called this function, since that will not tax the CPU). The idea here is roughly analogous to multiprocessing.Pool with the exceptions that: 1 - The caller will get to use the multiprocessing.Process model of using shared memory (inheritance) to pass arg data to the child, 2 - maxtasksperchild is always 1, 3 - no function return value is kept/tranferred (not yet implemented) """ # Sanity check if len(mp_proc_list) < 1: return # Create or own list with easy state watching procs = [] for p in mp_proc_list: procs.append(WatchedProcess(p)) # Launch all of them, but only so pool_size are running at any time keep_going = True while (keep_going): # Before we start any more, find out how many are running. First go # through the list of those started and see if alive. Update state. for p in procs: if p.state == 1: # been started if not p.process.is_alive(): p.state = 2 # process has finished or been terminated assert p.process.exitcode is not None, \ "Process is not alive but has no exitcode? "+ \ str(p.process) # now figure num_running num_running = len([p for p in procs if p.state == 1]) # Start some. Only as many as pool_size should ever be running. num_avail_cpus = pool_size - num_running num_to_start = len([p for p in procs if p.state == 0]) if num_to_start < 1: # all have been started, can finally leave loop and go wait break if num_avail_cpus > 0 and num_to_start > 0: num_to_start_now = min(num_avail_cpus, num_to_start) started_now = 0 for p in procs: if started_now < num_to_start_now and p.state == 0: p.start_process() # debug "launch_and_wait: started: "+str(p.process) started_now += 1 # else: otherwise, all cpus are in use, just wait ... # sleep to tame loop activity, but also must sleep a bit after each # start call so that the call to is_alive() woorks correctly time.sleep(1) # Out of the launching loop, can now wait on all procs left. for p in procs: p.join_process() # Check all exit codes before returning for p in procs: if 0 != p.process.exitcode: raise RuntimeError("Problem during: "+str(p.process.name)+ \ ', exitcode: '+str(p.process.exitcode)+'. Check log.')
python
def launch_and_wait(mp_proc_list, pool_size): """ Given a list of multiprocessing.Process objects which have not yet been started, this function launches them and blocks until the last finishes. This makes sure that only <pool_size> processes are ever working at any one time (this number does not include the main process which called this function, since that will not tax the CPU). The idea here is roughly analogous to multiprocessing.Pool with the exceptions that: 1 - The caller will get to use the multiprocessing.Process model of using shared memory (inheritance) to pass arg data to the child, 2 - maxtasksperchild is always 1, 3 - no function return value is kept/tranferred (not yet implemented) """ # Sanity check if len(mp_proc_list) < 1: return # Create or own list with easy state watching procs = [] for p in mp_proc_list: procs.append(WatchedProcess(p)) # Launch all of them, but only so pool_size are running at any time keep_going = True while (keep_going): # Before we start any more, find out how many are running. First go # through the list of those started and see if alive. Update state. for p in procs: if p.state == 1: # been started if not p.process.is_alive(): p.state = 2 # process has finished or been terminated assert p.process.exitcode is not None, \ "Process is not alive but has no exitcode? "+ \ str(p.process) # now figure num_running num_running = len([p for p in procs if p.state == 1]) # Start some. Only as many as pool_size should ever be running. num_avail_cpus = pool_size - num_running num_to_start = len([p for p in procs if p.state == 0]) if num_to_start < 1: # all have been started, can finally leave loop and go wait break if num_avail_cpus > 0 and num_to_start > 0: num_to_start_now = min(num_avail_cpus, num_to_start) started_now = 0 for p in procs: if started_now < num_to_start_now and p.state == 0: p.start_process() # debug "launch_and_wait: started: "+str(p.process) started_now += 1 # else: otherwise, all cpus are in use, just wait ... # sleep to tame loop activity, but also must sleep a bit after each # start call so that the call to is_alive() woorks correctly time.sleep(1) # Out of the launching loop, can now wait on all procs left. for p in procs: p.join_process() # Check all exit codes before returning for p in procs: if 0 != p.process.exitcode: raise RuntimeError("Problem during: "+str(p.process.name)+ \ ', exitcode: '+str(p.process.exitcode)+'. Check log.')
[ "def", "launch_and_wait", "(", "mp_proc_list", ",", "pool_size", ")", ":", "# Sanity check", "if", "len", "(", "mp_proc_list", ")", "<", "1", ":", "return", "# Create or own list with easy state watching", "procs", "=", "[", "]", "for", "p", "in", "mp_proc_list", ":", "procs", ".", "append", "(", "WatchedProcess", "(", "p", ")", ")", "# Launch all of them, but only so pool_size are running at any time", "keep_going", "=", "True", "while", "(", "keep_going", ")", ":", "# Before we start any more, find out how many are running. First go", "# through the list of those started and see if alive. Update state.", "for", "p", "in", "procs", ":", "if", "p", ".", "state", "==", "1", ":", "# been started", "if", "not", "p", ".", "process", ".", "is_alive", "(", ")", ":", "p", ".", "state", "=", "2", "# process has finished or been terminated", "assert", "p", ".", "process", ".", "exitcode", "is", "not", "None", ",", "\"Process is not alive but has no exitcode? \"", "+", "str", "(", "p", ".", "process", ")", "# now figure num_running", "num_running", "=", "len", "(", "[", "p", "for", "p", "in", "procs", "if", "p", ".", "state", "==", "1", "]", ")", "# Start some. Only as many as pool_size should ever be running.", "num_avail_cpus", "=", "pool_size", "-", "num_running", "num_to_start", "=", "len", "(", "[", "p", "for", "p", "in", "procs", "if", "p", ".", "state", "==", "0", "]", ")", "if", "num_to_start", "<", "1", ":", "# all have been started, can finally leave loop and go wait", "break", "if", "num_avail_cpus", ">", "0", "and", "num_to_start", ">", "0", ":", "num_to_start_now", "=", "min", "(", "num_avail_cpus", ",", "num_to_start", ")", "started_now", "=", "0", "for", "p", "in", "procs", ":", "if", "started_now", "<", "num_to_start_now", "and", "p", ".", "state", "==", "0", ":", "p", ".", "start_process", "(", ")", "# debug \"launch_and_wait: started: \"+str(p.process)", "started_now", "+=", "1", "# else: otherwise, all cpus are in use, just wait ...", "# sleep to tame loop activity, but also must sleep a bit after each", "# start call so that the call to is_alive() woorks correctly", "time", ".", "sleep", "(", "1", ")", "# Out of the launching loop, can now wait on all procs left.", "for", "p", "in", "procs", ":", "p", ".", "join_process", "(", ")", "# Check all exit codes before returning", "for", "p", "in", "procs", ":", "if", "0", "!=", "p", ".", "process", ".", "exitcode", ":", "raise", "RuntimeError", "(", "\"Problem during: \"", "+", "str", "(", "p", ".", "process", ".", "name", ")", "+", "', exitcode: '", "+", "str", "(", "p", ".", "process", ".", "exitcode", ")", "+", "'. Check log.'", ")" ]
Given a list of multiprocessing.Process objects which have not yet been started, this function launches them and blocks until the last finishes. This makes sure that only <pool_size> processes are ever working at any one time (this number does not include the main process which called this function, since that will not tax the CPU). The idea here is roughly analogous to multiprocessing.Pool with the exceptions that: 1 - The caller will get to use the multiprocessing.Process model of using shared memory (inheritance) to pass arg data to the child, 2 - maxtasksperchild is always 1, 3 - no function return value is kept/tranferred (not yet implemented)
[ "Given", "a", "list", "of", "multiprocessing", ".", "Process", "objects", "which", "have", "not", "yet", "been", "started", "this", "function", "launches", "them", "and", "blocks", "until", "the", "last", "finishes", ".", "This", "makes", "sure", "that", "only", "<pool_size", ">", "processes", "are", "ever", "working", "at", "any", "one", "time", "(", "this", "number", "does", "not", "include", "the", "main", "process", "which", "called", "this", "function", "since", "that", "will", "not", "tax", "the", "CPU", ")", ".", "The", "idea", "here", "is", "roughly", "analogous", "to", "multiprocessing", ".", "Pool", "with", "the", "exceptions", "that", ":", "1", "-", "The", "caller", "will", "get", "to", "use", "the", "multiprocessing", ".", "Process", "model", "of", "using", "shared", "memory", "(", "inheritance", ")", "to", "pass", "arg", "data", "to", "the", "child", "2", "-", "maxtasksperchild", "is", "always", "1", "3", "-", "no", "function", "return", "value", "is", "kept", "/", "tranferred", "(", "not", "yet", "implemented", ")" ]
train
https://github.com/spacetelescope/stsci.tools/blob/9a022503ad24ca54ce83331482dfa3ff6de9f403/lib/stsci/tools/mputil.py#L36-L103
spacetelescope/stsci.tools
lib/stsci/tools/mputil.py
best_tile_layout
def best_tile_layout(pool_size): """ Determine and return the best layout of "tiles" for fastest overall parallel processing of a rectangular image broken up into N smaller equally-sized rectangular tiles, given as input the number of processes/chunks which can be run/worked at the same time (pool_size). This attempts to return a layout whose total number of tiles is as close as possible to pool_size, without going over (and thus not really taking advantage of pooling). Since we can vary the size of the rectangles, there is not much (any?) benefit to pooling. Returns a tuple of ( <num tiles in X dir>, <num in Y direction> ) This assumes the image in question is relatively close to square, and so the returned tuple attempts to give a layout which is as squarishly-blocked as possible, except in cases where speed would be sacrificed. EXAMPLES: For pool_size of 4, the best result is 2x2. For pool_size of 6, the best result is 2x3. For pool_size of 5, a result of 1x5 is better than a result of 2x2 (which would leave one core unused), and 1x5 is also better than a result of 2x3 (which would require one core to work twice while all others wait). For higher, odd pool_size values (say 39), it is deemed best to sacrifice a few unused cores to satisfy our other constraints, and thus the result of 6x6 is best (giving 36 tiles and 3 unused cores). """ # Easy answer sanity-checks if pool_size < 2: return (1, 1) # Next, use a small mapping of hard-coded results. While we agree # that many of these are unlikely pool_size values, they are easy # to accomodate. mapping = { 0:(1,1), 1:(1,1), 2:(1,2), 3:(1,3), 4:(2,2), 5:(1,5), 6:(2,3), 7:(2,3), 8:(2,4), 9:(3,3), 10:(2,5), 11:(2,5), 14:(2,7), 18:(3,6), 19:(3,6), 28:(4,7), 29:(4,7), 32:(4,8), 33:(4,8), 34:(4,8), 40:(4,10), 41:(4,10) } if pool_size in mapping: return mapping[pool_size] # Next, take a guess using the square root and (for the sake of # simplicity), go with it. We *could* get much fancier here... # Use floor-rounding (not ceil) so that the total number of resulting # tiles is <= pool_size. xnum = int(math.sqrt(pool_size)) ynum = int((1.*pool_size)/xnum) return (xnum, ynum)
python
def best_tile_layout(pool_size): """ Determine and return the best layout of "tiles" for fastest overall parallel processing of a rectangular image broken up into N smaller equally-sized rectangular tiles, given as input the number of processes/chunks which can be run/worked at the same time (pool_size). This attempts to return a layout whose total number of tiles is as close as possible to pool_size, without going over (and thus not really taking advantage of pooling). Since we can vary the size of the rectangles, there is not much (any?) benefit to pooling. Returns a tuple of ( <num tiles in X dir>, <num in Y direction> ) This assumes the image in question is relatively close to square, and so the returned tuple attempts to give a layout which is as squarishly-blocked as possible, except in cases where speed would be sacrificed. EXAMPLES: For pool_size of 4, the best result is 2x2. For pool_size of 6, the best result is 2x3. For pool_size of 5, a result of 1x5 is better than a result of 2x2 (which would leave one core unused), and 1x5 is also better than a result of 2x3 (which would require one core to work twice while all others wait). For higher, odd pool_size values (say 39), it is deemed best to sacrifice a few unused cores to satisfy our other constraints, and thus the result of 6x6 is best (giving 36 tiles and 3 unused cores). """ # Easy answer sanity-checks if pool_size < 2: return (1, 1) # Next, use a small mapping of hard-coded results. While we agree # that many of these are unlikely pool_size values, they are easy # to accomodate. mapping = { 0:(1,1), 1:(1,1), 2:(1,2), 3:(1,3), 4:(2,2), 5:(1,5), 6:(2,3), 7:(2,3), 8:(2,4), 9:(3,3), 10:(2,5), 11:(2,5), 14:(2,7), 18:(3,6), 19:(3,6), 28:(4,7), 29:(4,7), 32:(4,8), 33:(4,8), 34:(4,8), 40:(4,10), 41:(4,10) } if pool_size in mapping: return mapping[pool_size] # Next, take a guess using the square root and (for the sake of # simplicity), go with it. We *could* get much fancier here... # Use floor-rounding (not ceil) so that the total number of resulting # tiles is <= pool_size. xnum = int(math.sqrt(pool_size)) ynum = int((1.*pool_size)/xnum) return (xnum, ynum)
[ "def", "best_tile_layout", "(", "pool_size", ")", ":", "# Easy answer sanity-checks", "if", "pool_size", "<", "2", ":", "return", "(", "1", ",", "1", ")", "# Next, use a small mapping of hard-coded results. While we agree", "# that many of these are unlikely pool_size values, they are easy", "# to accomodate.", "mapping", "=", "{", "0", ":", "(", "1", ",", "1", ")", ",", "1", ":", "(", "1", ",", "1", ")", ",", "2", ":", "(", "1", ",", "2", ")", ",", "3", ":", "(", "1", ",", "3", ")", ",", "4", ":", "(", "2", ",", "2", ")", ",", "5", ":", "(", "1", ",", "5", ")", ",", "6", ":", "(", "2", ",", "3", ")", ",", "7", ":", "(", "2", ",", "3", ")", ",", "8", ":", "(", "2", ",", "4", ")", ",", "9", ":", "(", "3", ",", "3", ")", ",", "10", ":", "(", "2", ",", "5", ")", ",", "11", ":", "(", "2", ",", "5", ")", ",", "14", ":", "(", "2", ",", "7", ")", ",", "18", ":", "(", "3", ",", "6", ")", ",", "19", ":", "(", "3", ",", "6", ")", ",", "28", ":", "(", "4", ",", "7", ")", ",", "29", ":", "(", "4", ",", "7", ")", ",", "32", ":", "(", "4", ",", "8", ")", ",", "33", ":", "(", "4", ",", "8", ")", ",", "34", ":", "(", "4", ",", "8", ")", ",", "40", ":", "(", "4", ",", "10", ")", ",", "41", ":", "(", "4", ",", "10", ")", "}", "if", "pool_size", "in", "mapping", ":", "return", "mapping", "[", "pool_size", "]", "# Next, take a guess using the square root and (for the sake of", "# simplicity), go with it. We *could* get much fancier here...", "# Use floor-rounding (not ceil) so that the total number of resulting", "# tiles is <= pool_size.", "xnum", "=", "int", "(", "math", ".", "sqrt", "(", "pool_size", ")", ")", "ynum", "=", "int", "(", "(", "1.", "*", "pool_size", ")", "/", "xnum", ")", "return", "(", "xnum", ",", "ynum", ")" ]
Determine and return the best layout of "tiles" for fastest overall parallel processing of a rectangular image broken up into N smaller equally-sized rectangular tiles, given as input the number of processes/chunks which can be run/worked at the same time (pool_size). This attempts to return a layout whose total number of tiles is as close as possible to pool_size, without going over (and thus not really taking advantage of pooling). Since we can vary the size of the rectangles, there is not much (any?) benefit to pooling. Returns a tuple of ( <num tiles in X dir>, <num in Y direction> ) This assumes the image in question is relatively close to square, and so the returned tuple attempts to give a layout which is as squarishly-blocked as possible, except in cases where speed would be sacrificed. EXAMPLES: For pool_size of 4, the best result is 2x2. For pool_size of 6, the best result is 2x3. For pool_size of 5, a result of 1x5 is better than a result of 2x2 (which would leave one core unused), and 1x5 is also better than a result of 2x3 (which would require one core to work twice while all others wait). For higher, odd pool_size values (say 39), it is deemed best to sacrifice a few unused cores to satisfy our other constraints, and thus the result of 6x6 is best (giving 36 tiles and 3 unused cores).
[ "Determine", "and", "return", "the", "best", "layout", "of", "tiles", "for", "fastest", "overall", "parallel", "processing", "of", "a", "rectangular", "image", "broken", "up", "into", "N", "smaller", "equally", "-", "sized", "rectangular", "tiles", "given", "as", "input", "the", "number", "of", "processes", "/", "chunks", "which", "can", "be", "run", "/", "worked", "at", "the", "same", "time", "(", "pool_size", ")", "." ]
train
https://github.com/spacetelescope/stsci.tools/blob/9a022503ad24ca54ce83331482dfa3ff6de9f403/lib/stsci/tools/mputil.py#L107-L160
3ll3d00d/vibe
backend/src/core/reactor.py
Reactor._accept
def _accept(self): """ Work loop runs forever (or until running is False) :return: """ logger.warning("Reactor " + self._name + " is starting") while self.running: try: self._completeTask() except: logger.exception("Unexpected exception during request processing") logger.warning("Reactor " + self._name + " is terminating")
python
def _accept(self): """ Work loop runs forever (or until running is False) :return: """ logger.warning("Reactor " + self._name + " is starting") while self.running: try: self._completeTask() except: logger.exception("Unexpected exception during request processing") logger.warning("Reactor " + self._name + " is terminating")
[ "def", "_accept", "(", "self", ")", ":", "logger", ".", "warning", "(", "\"Reactor \"", "+", "self", ".", "_name", "+", "\" is starting\"", ")", "while", "self", ".", "running", ":", "try", ":", "self", ".", "_completeTask", "(", ")", "except", ":", "logger", ".", "exception", "(", "\"Unexpected exception during request processing\"", ")", "logger", ".", "warning", "(", "\"Reactor \"", "+", "self", ".", "_name", "+", "\" is terminating\"", ")" ]
Work loop runs forever (or until running is False) :return:
[ "Work", "loop", "runs", "forever", "(", "or", "until", "running", "is", "False", ")", ":", "return", ":" ]
train
https://github.com/3ll3d00d/vibe/blob/124b029f13ac746723e92cb47e9cb56edd2e54b5/backend/src/core/reactor.py#L23-L34
3ll3d00d/vibe
backend/src/core/reactor.py
Reactor.offer
def offer(self, requestType, *args): """ public interface to the reactor. :param requestType: :param args: :return: """ if self._funcsByRequest.get(requestType) is not None: self._workQueue.put((requestType, list(*args))) else: logger.error("Ignoring unknown request on reactor " + self._name + " " + requestType)
python
def offer(self, requestType, *args): """ public interface to the reactor. :param requestType: :param args: :return: """ if self._funcsByRequest.get(requestType) is not None: self._workQueue.put((requestType, list(*args))) else: logger.error("Ignoring unknown request on reactor " + self._name + " " + requestType)
[ "def", "offer", "(", "self", ",", "requestType", ",", "*", "args", ")", ":", "if", "self", ".", "_funcsByRequest", ".", "get", "(", "requestType", ")", "is", "not", "None", ":", "self", ".", "_workQueue", ".", "put", "(", "(", "requestType", ",", "list", "(", "*", "args", ")", ")", ")", "else", ":", "logger", ".", "error", "(", "\"Ignoring unknown request on reactor \"", "+", "self", ".", "_name", "+", "\" \"", "+", "requestType", ")" ]
public interface to the reactor. :param requestType: :param args: :return:
[ "public", "interface", "to", "the", "reactor", ".", ":", "param", "requestType", ":", ":", "param", "args", ":", ":", "return", ":" ]
train
https://github.com/3ll3d00d/vibe/blob/124b029f13ac746723e92cb47e9cb56edd2e54b5/backend/src/core/reactor.py#L58-L68
spacetelescope/stsci.tools
lib/stsci/tools/teal_bttn.py
TealActionParButton.clicked
def clicked(self): """ Called when this button is clicked. Execute code from .cfgspc """ try: from . import teal except: teal = None try: # start drilling down into the tpo to get the code tealGui = self._mainGuiObj tealGui.showStatus('Clicked "'+self.getButtonLabel()+'"', keep=1) pscope = self.paramInfo.scope pname = self.paramInfo.name tpo = tealGui._taskParsObj tup = tpo.getExecuteStrings(pscope, pname) code = '' if not tup: if teal: teal.popUpErr(tealGui.top, "No action to perform", "Action Button Error") return for exname in tup: if '_RULES_' in tpo and exname in tpo['_RULES_'].configspec: ruleSig = tpo['_RULES_'].configspec[exname] chkArgsDict = vtor_checks.sigStrToKwArgsDict(ruleSig) code = chkArgsDict.get('code') # a string or None # now go ahead and execute it teal.execEmbCode(pscope, pname, self.getButtonLabel(), tealGui, code) # done tealGui.debug('Finished: "'+self.getButtonLabel()+'"') except Exception as ex: msg = 'Error executing: "'+self.getButtonLabel()+'"\n'+ex.message msgFull = msg+'\n'+''.join(traceback.format_exc()) msgFull+= "CODE:\n"+code if tealGui: if teal: teal.popUpErr(tealGui.top, msg, "Action Button Error") tealGui.debug(msgFull) else: if teal: teal.popUpErr(None, msg, "Action Button Error") print(msgFull)
python
def clicked(self): """ Called when this button is clicked. Execute code from .cfgspc """ try: from . import teal except: teal = None try: # start drilling down into the tpo to get the code tealGui = self._mainGuiObj tealGui.showStatus('Clicked "'+self.getButtonLabel()+'"', keep=1) pscope = self.paramInfo.scope pname = self.paramInfo.name tpo = tealGui._taskParsObj tup = tpo.getExecuteStrings(pscope, pname) code = '' if not tup: if teal: teal.popUpErr(tealGui.top, "No action to perform", "Action Button Error") return for exname in tup: if '_RULES_' in tpo and exname in tpo['_RULES_'].configspec: ruleSig = tpo['_RULES_'].configspec[exname] chkArgsDict = vtor_checks.sigStrToKwArgsDict(ruleSig) code = chkArgsDict.get('code') # a string or None # now go ahead and execute it teal.execEmbCode(pscope, pname, self.getButtonLabel(), tealGui, code) # done tealGui.debug('Finished: "'+self.getButtonLabel()+'"') except Exception as ex: msg = 'Error executing: "'+self.getButtonLabel()+'"\n'+ex.message msgFull = msg+'\n'+''.join(traceback.format_exc()) msgFull+= "CODE:\n"+code if tealGui: if teal: teal.popUpErr(tealGui.top, msg, "Action Button Error") tealGui.debug(msgFull) else: if teal: teal.popUpErr(None, msg, "Action Button Error") print(msgFull)
[ "def", "clicked", "(", "self", ")", ":", "try", ":", "from", ".", "import", "teal", "except", ":", "teal", "=", "None", "try", ":", "# start drilling down into the tpo to get the code", "tealGui", "=", "self", ".", "_mainGuiObj", "tealGui", ".", "showStatus", "(", "'Clicked \"'", "+", "self", ".", "getButtonLabel", "(", ")", "+", "'\"'", ",", "keep", "=", "1", ")", "pscope", "=", "self", ".", "paramInfo", ".", "scope", "pname", "=", "self", ".", "paramInfo", ".", "name", "tpo", "=", "tealGui", ".", "_taskParsObj", "tup", "=", "tpo", ".", "getExecuteStrings", "(", "pscope", ",", "pname", ")", "code", "=", "''", "if", "not", "tup", ":", "if", "teal", ":", "teal", ".", "popUpErr", "(", "tealGui", ".", "top", ",", "\"No action to perform\"", ",", "\"Action Button Error\"", ")", "return", "for", "exname", "in", "tup", ":", "if", "'_RULES_'", "in", "tpo", "and", "exname", "in", "tpo", "[", "'_RULES_'", "]", ".", "configspec", ":", "ruleSig", "=", "tpo", "[", "'_RULES_'", "]", ".", "configspec", "[", "exname", "]", "chkArgsDict", "=", "vtor_checks", ".", "sigStrToKwArgsDict", "(", "ruleSig", ")", "code", "=", "chkArgsDict", ".", "get", "(", "'code'", ")", "# a string or None", "# now go ahead and execute it", "teal", ".", "execEmbCode", "(", "pscope", ",", "pname", ",", "self", ".", "getButtonLabel", "(", ")", ",", "tealGui", ",", "code", ")", "# done", "tealGui", ".", "debug", "(", "'Finished: \"'", "+", "self", ".", "getButtonLabel", "(", ")", "+", "'\"'", ")", "except", "Exception", "as", "ex", ":", "msg", "=", "'Error executing: \"'", "+", "self", ".", "getButtonLabel", "(", ")", "+", "'\"\\n'", "+", "ex", ".", "message", "msgFull", "=", "msg", "+", "'\\n'", "+", "''", ".", "join", "(", "traceback", ".", "format_exc", "(", ")", ")", "msgFull", "+=", "\"CODE:\\n\"", "+", "code", "if", "tealGui", ":", "if", "teal", ":", "teal", ".", "popUpErr", "(", "tealGui", ".", "top", ",", "msg", ",", "\"Action Button Error\"", ")", "tealGui", ".", "debug", "(", "msgFull", ")", "else", ":", "if", "teal", ":", "teal", ".", "popUpErr", "(", "None", ",", "msg", ",", "\"Action Button Error\"", ")", "print", "(", "msgFull", ")" ]
Called when this button is clicked. Execute code from .cfgspc
[ "Called", "when", "this", "button", "is", "clicked", ".", "Execute", "code", "from", ".", "cfgspc" ]
train
https://github.com/spacetelescope/stsci.tools/blob/9a022503ad24ca54ce83331482dfa3ff6de9f403/lib/stsci/tools/teal_bttn.py#L31-L70
spacetelescope/stsci.tools
lib/stsci/tools/for2to3.py
ndarr2str
def ndarr2str(arr, encoding='ascii'): """ This is used to ensure that the return value of arr.tostring() is actually a string. This will prevent lots of if-checks in calling code. As of numpy v1.6.1 (in Python 3.2.3), the tostring() function still returns type 'bytes', not 'str' as it advertises. """ # be fast, don't check - just assume 'arr' is a numpy array - the tostring # call will fail anyway if not retval = arr.tostring() # would rather check "if isinstance(retval, bytes)", but support 2.5. # could rm the if PY3K check, but it makes this faster on 2.x. if PY3K and not isinstance(retval, str): return retval.decode(encoding) else: # is str return retval
python
def ndarr2str(arr, encoding='ascii'): """ This is used to ensure that the return value of arr.tostring() is actually a string. This will prevent lots of if-checks in calling code. As of numpy v1.6.1 (in Python 3.2.3), the tostring() function still returns type 'bytes', not 'str' as it advertises. """ # be fast, don't check - just assume 'arr' is a numpy array - the tostring # call will fail anyway if not retval = arr.tostring() # would rather check "if isinstance(retval, bytes)", but support 2.5. # could rm the if PY3K check, but it makes this faster on 2.x. if PY3K and not isinstance(retval, str): return retval.decode(encoding) else: # is str return retval
[ "def", "ndarr2str", "(", "arr", ",", "encoding", "=", "'ascii'", ")", ":", "# be fast, don't check - just assume 'arr' is a numpy array - the tostring", "# call will fail anyway if not", "retval", "=", "arr", ".", "tostring", "(", ")", "# would rather check \"if isinstance(retval, bytes)\", but support 2.5.", "# could rm the if PY3K check, but it makes this faster on 2.x.", "if", "PY3K", "and", "not", "isinstance", "(", "retval", ",", "str", ")", ":", "return", "retval", ".", "decode", "(", "encoding", ")", "else", ":", "# is str", "return", "retval" ]
This is used to ensure that the return value of arr.tostring() is actually a string. This will prevent lots of if-checks in calling code. As of numpy v1.6.1 (in Python 3.2.3), the tostring() function still returns type 'bytes', not 'str' as it advertises.
[ "This", "is", "used", "to", "ensure", "that", "the", "return", "value", "of", "arr", ".", "tostring", "()", "is", "actually", "a", "string", ".", "This", "will", "prevent", "lots", "of", "if", "-", "checks", "in", "calling", "code", ".", "As", "of", "numpy", "v1", ".", "6", ".", "1", "(", "in", "Python", "3", ".", "2", ".", "3", ")", "the", "tostring", "()", "function", "still", "returns", "type", "bytes", "not", "str", "as", "it", "advertises", "." ]
train
https://github.com/spacetelescope/stsci.tools/blob/9a022503ad24ca54ce83331482dfa3ff6de9f403/lib/stsci/tools/for2to3.py#L15-L28
spacetelescope/stsci.tools
lib/stsci/tools/for2to3.py
ndarr2bytes
def ndarr2bytes(arr, encoding='ascii'): """ This is used to ensure that the return value of arr.tostring() is actually a *bytes* array in PY3K. See notes in ndarr2str above. Even though we consider it a bug that numpy's tostring() function returns a bytes array in PY3K, there are actually many instances where that is what we want - bytes, not unicode. So we use this function in those instances to ensure that when/if this numpy "bug" is "fixed", that our calling code still gets bytes where it needs/expects them. """ # be fast, don't check - just assume 'arr' is a numpy array - the tostring # call will fail anyway if not retval = arr.tostring() # would rather check "if not isinstance(retval, bytes)", but support 2.5. if PY3K and isinstance(retval, str): # Take note if this ever gets used. If this ever occurs, it # is likely wildly inefficient since numpy.tostring() is now # returning unicode and numpy surely has a tobytes() func by now. # If so, add a code path to call its tobytes() func at our start. return retval.encode(encoding) else: # is str==bytes in 2.x return retval
python
def ndarr2bytes(arr, encoding='ascii'): """ This is used to ensure that the return value of arr.tostring() is actually a *bytes* array in PY3K. See notes in ndarr2str above. Even though we consider it a bug that numpy's tostring() function returns a bytes array in PY3K, there are actually many instances where that is what we want - bytes, not unicode. So we use this function in those instances to ensure that when/if this numpy "bug" is "fixed", that our calling code still gets bytes where it needs/expects them. """ # be fast, don't check - just assume 'arr' is a numpy array - the tostring # call will fail anyway if not retval = arr.tostring() # would rather check "if not isinstance(retval, bytes)", but support 2.5. if PY3K and isinstance(retval, str): # Take note if this ever gets used. If this ever occurs, it # is likely wildly inefficient since numpy.tostring() is now # returning unicode and numpy surely has a tobytes() func by now. # If so, add a code path to call its tobytes() func at our start. return retval.encode(encoding) else: # is str==bytes in 2.x return retval
[ "def", "ndarr2bytes", "(", "arr", ",", "encoding", "=", "'ascii'", ")", ":", "# be fast, don't check - just assume 'arr' is a numpy array - the tostring", "# call will fail anyway if not", "retval", "=", "arr", ".", "tostring", "(", ")", "# would rather check \"if not isinstance(retval, bytes)\", but support 2.5.", "if", "PY3K", "and", "isinstance", "(", "retval", ",", "str", ")", ":", "# Take note if this ever gets used. If this ever occurs, it", "# is likely wildly inefficient since numpy.tostring() is now", "# returning unicode and numpy surely has a tobytes() func by now.", "# If so, add a code path to call its tobytes() func at our start.", "return", "retval", ".", "encode", "(", "encoding", ")", "else", ":", "# is str==bytes in 2.x", "return", "retval" ]
This is used to ensure that the return value of arr.tostring() is actually a *bytes* array in PY3K. See notes in ndarr2str above. Even though we consider it a bug that numpy's tostring() function returns a bytes array in PY3K, there are actually many instances where that is what we want - bytes, not unicode. So we use this function in those instances to ensure that when/if this numpy "bug" is "fixed", that our calling code still gets bytes where it needs/expects them.
[ "This", "is", "used", "to", "ensure", "that", "the", "return", "value", "of", "arr", ".", "tostring", "()", "is", "actually", "a", "*", "bytes", "*", "array", "in", "PY3K", ".", "See", "notes", "in", "ndarr2str", "above", ".", "Even", "though", "we", "consider", "it", "a", "bug", "that", "numpy", "s", "tostring", "()", "function", "returns", "a", "bytes", "array", "in", "PY3K", "there", "are", "actually", "many", "instances", "where", "that", "is", "what", "we", "want", "-", "bytes", "not", "unicode", ".", "So", "we", "use", "this", "function", "in", "those", "instances", "to", "ensure", "that", "when", "/", "if", "this", "numpy", "bug", "is", "fixed", "that", "our", "calling", "code", "still", "gets", "bytes", "where", "it", "needs", "/", "expects", "them", "." ]
train
https://github.com/spacetelescope/stsci.tools/blob/9a022503ad24ca54ce83331482dfa3ff6de9f403/lib/stsci/tools/for2to3.py#L31-L50
spacetelescope/stsci.tools
lib/stsci/tools/for2to3.py
tobytes
def tobytes(s, encoding='ascii'): """ Convert string s to the 'bytes' type, in all Pythons, even back before Python 2.6. What 'str' means varies by PY3K or not. In Pythons before 3.0, this is technically the same as the str type in terms of the character data in memory. """ # NOTE: after we abandon 2.5, we might simply instead use "bytes(s)" # NOTE: after we abandon all 2.*, del this and prepend byte strings with 'b' if PY3K: if isinstance(s, bytes): return s else: return s.encode(encoding) else: # for py2.6 on (before 3.0), bytes is same as str; 2.5 has no bytes # but handle if unicode is passed if isinstance(s, unicode): return s.encode(encoding) else: return s
python
def tobytes(s, encoding='ascii'): """ Convert string s to the 'bytes' type, in all Pythons, even back before Python 2.6. What 'str' means varies by PY3K or not. In Pythons before 3.0, this is technically the same as the str type in terms of the character data in memory. """ # NOTE: after we abandon 2.5, we might simply instead use "bytes(s)" # NOTE: after we abandon all 2.*, del this and prepend byte strings with 'b' if PY3K: if isinstance(s, bytes): return s else: return s.encode(encoding) else: # for py2.6 on (before 3.0), bytes is same as str; 2.5 has no bytes # but handle if unicode is passed if isinstance(s, unicode): return s.encode(encoding) else: return s
[ "def", "tobytes", "(", "s", ",", "encoding", "=", "'ascii'", ")", ":", "# NOTE: after we abandon 2.5, we might simply instead use \"bytes(s)\"", "# NOTE: after we abandon all 2.*, del this and prepend byte strings with 'b'", "if", "PY3K", ":", "if", "isinstance", "(", "s", ",", "bytes", ")", ":", "return", "s", "else", ":", "return", "s", ".", "encode", "(", "encoding", ")", "else", ":", "# for py2.6 on (before 3.0), bytes is same as str; 2.5 has no bytes", "# but handle if unicode is passed", "if", "isinstance", "(", "s", ",", "unicode", ")", ":", "return", "s", ".", "encode", "(", "encoding", ")", "else", ":", "return", "s" ]
Convert string s to the 'bytes' type, in all Pythons, even back before Python 2.6. What 'str' means varies by PY3K or not. In Pythons before 3.0, this is technically the same as the str type in terms of the character data in memory.
[ "Convert", "string", "s", "to", "the", "bytes", "type", "in", "all", "Pythons", "even", "back", "before", "Python", "2", ".", "6", ".", "What", "str", "means", "varies", "by", "PY3K", "or", "not", ".", "In", "Pythons", "before", "3", ".", "0", "this", "is", "technically", "the", "same", "as", "the", "str", "type", "in", "terms", "of", "the", "character", "data", "in", "memory", "." ]
train
https://github.com/spacetelescope/stsci.tools/blob/9a022503ad24ca54ce83331482dfa3ff6de9f403/lib/stsci/tools/for2to3.py#L53-L71
spacetelescope/stsci.tools
lib/stsci/tools/for2to3.py
tostr
def tostr(s, encoding='ascii'): """ Convert string-like-thing s to the 'str' type, in all Pythons, even back before Python 2.6. What 'str' means varies by PY3K or not. In Pythons before 3.0, str and bytes are the same type. In Python 3+, this may require a decoding step. """ if PY3K: if isinstance(s, str): # str == unicode in PY3K return s else: # s is type bytes return s.decode(encoding) else: # for py2.6 on (before 3.0), bytes is same as str; 2.5 has no bytes # but handle if unicode is passed if isinstance(s, unicode): return s.encode(encoding) else: return s
python
def tostr(s, encoding='ascii'): """ Convert string-like-thing s to the 'str' type, in all Pythons, even back before Python 2.6. What 'str' means varies by PY3K or not. In Pythons before 3.0, str and bytes are the same type. In Python 3+, this may require a decoding step. """ if PY3K: if isinstance(s, str): # str == unicode in PY3K return s else: # s is type bytes return s.decode(encoding) else: # for py2.6 on (before 3.0), bytes is same as str; 2.5 has no bytes # but handle if unicode is passed if isinstance(s, unicode): return s.encode(encoding) else: return s
[ "def", "tostr", "(", "s", ",", "encoding", "=", "'ascii'", ")", ":", "if", "PY3K", ":", "if", "isinstance", "(", "s", ",", "str", ")", ":", "# str == unicode in PY3K", "return", "s", "else", ":", "# s is type bytes", "return", "s", ".", "decode", "(", "encoding", ")", "else", ":", "# for py2.6 on (before 3.0), bytes is same as str; 2.5 has no bytes", "# but handle if unicode is passed", "if", "isinstance", "(", "s", ",", "unicode", ")", ":", "return", "s", ".", "encode", "(", "encoding", ")", "else", ":", "return", "s" ]
Convert string-like-thing s to the 'str' type, in all Pythons, even back before Python 2.6. What 'str' means varies by PY3K or not. In Pythons before 3.0, str and bytes are the same type. In Python 3+, this may require a decoding step.
[ "Convert", "string", "-", "like", "-", "thing", "s", "to", "the", "str", "type", "in", "all", "Pythons", "even", "back", "before", "Python", "2", ".", "6", ".", "What", "str", "means", "varies", "by", "PY3K", "or", "not", ".", "In", "Pythons", "before", "3", ".", "0", "str", "and", "bytes", "are", "the", "same", "type", ".", "In", "Python", "3", "+", "this", "may", "require", "a", "decoding", "step", "." ]
train
https://github.com/spacetelescope/stsci.tools/blob/9a022503ad24ca54ce83331482dfa3ff6de9f403/lib/stsci/tools/for2to3.py#L73-L89
ionelmc/python-aspectlib
src/aspectlib/contrib.py
retry
def retry(func=None, retries=5, backoff=None, exceptions=(IOError, OSError, EOFError), cleanup=None, sleep=time.sleep): """ Decorator that retries the call ``retries`` times if ``func`` raises ``exceptions``. Can use a ``backoff`` function to sleep till next retry. Example:: >>> should_fail = lambda foo=[1,2,3]: foo and foo.pop() >>> @retry ... def flaky_func(): ... if should_fail(): ... raise OSError('Tough luck!') ... print("Success!") ... >>> flaky_func() Success! If it reaches the retry limit:: >>> @retry ... def bad_func(): ... raise OSError('Tough luck!') ... >>> bad_func() Traceback (most recent call last): ... OSError: Tough luck! """ @Aspect(bind=True) def retry_aspect(cutpoint, *args, **kwargs): for count in range(retries + 1): try: if count and cleanup: cleanup(*args, **kwargs) yield break except exceptions as exc: if count == retries: raise if not backoff: timeout = 0 elif isinstance(backoff, (int, float)): timeout = backoff else: timeout = backoff(count) logger.exception("%s(%s, %s) raised exception %s. %s retries left. Sleeping %s secs.", cutpoint.__name__, args, kwargs, exc, retries - count, timeout) sleep(timeout) return retry_aspect if func is None else retry_aspect(func)
python
def retry(func=None, retries=5, backoff=None, exceptions=(IOError, OSError, EOFError), cleanup=None, sleep=time.sleep): """ Decorator that retries the call ``retries`` times if ``func`` raises ``exceptions``. Can use a ``backoff`` function to sleep till next retry. Example:: >>> should_fail = lambda foo=[1,2,3]: foo and foo.pop() >>> @retry ... def flaky_func(): ... if should_fail(): ... raise OSError('Tough luck!') ... print("Success!") ... >>> flaky_func() Success! If it reaches the retry limit:: >>> @retry ... def bad_func(): ... raise OSError('Tough luck!') ... >>> bad_func() Traceback (most recent call last): ... OSError: Tough luck! """ @Aspect(bind=True) def retry_aspect(cutpoint, *args, **kwargs): for count in range(retries + 1): try: if count and cleanup: cleanup(*args, **kwargs) yield break except exceptions as exc: if count == retries: raise if not backoff: timeout = 0 elif isinstance(backoff, (int, float)): timeout = backoff else: timeout = backoff(count) logger.exception("%s(%s, %s) raised exception %s. %s retries left. Sleeping %s secs.", cutpoint.__name__, args, kwargs, exc, retries - count, timeout) sleep(timeout) return retry_aspect if func is None else retry_aspect(func)
[ "def", "retry", "(", "func", "=", "None", ",", "retries", "=", "5", ",", "backoff", "=", "None", ",", "exceptions", "=", "(", "IOError", ",", "OSError", ",", "EOFError", ")", ",", "cleanup", "=", "None", ",", "sleep", "=", "time", ".", "sleep", ")", ":", "@", "Aspect", "(", "bind", "=", "True", ")", "def", "retry_aspect", "(", "cutpoint", ",", "*", "args", ",", "*", "*", "kwargs", ")", ":", "for", "count", "in", "range", "(", "retries", "+", "1", ")", ":", "try", ":", "if", "count", "and", "cleanup", ":", "cleanup", "(", "*", "args", ",", "*", "*", "kwargs", ")", "yield", "break", "except", "exceptions", "as", "exc", ":", "if", "count", "==", "retries", ":", "raise", "if", "not", "backoff", ":", "timeout", "=", "0", "elif", "isinstance", "(", "backoff", ",", "(", "int", ",", "float", ")", ")", ":", "timeout", "=", "backoff", "else", ":", "timeout", "=", "backoff", "(", "count", ")", "logger", ".", "exception", "(", "\"%s(%s, %s) raised exception %s. %s retries left. Sleeping %s secs.\"", ",", "cutpoint", ".", "__name__", ",", "args", ",", "kwargs", ",", "exc", ",", "retries", "-", "count", ",", "timeout", ")", "sleep", "(", "timeout", ")", "return", "retry_aspect", "if", "func", "is", "None", "else", "retry_aspect", "(", "func", ")" ]
Decorator that retries the call ``retries`` times if ``func`` raises ``exceptions``. Can use a ``backoff`` function to sleep till next retry. Example:: >>> should_fail = lambda foo=[1,2,3]: foo and foo.pop() >>> @retry ... def flaky_func(): ... if should_fail(): ... raise OSError('Tough luck!') ... print("Success!") ... >>> flaky_func() Success! If it reaches the retry limit:: >>> @retry ... def bad_func(): ... raise OSError('Tough luck!') ... >>> bad_func() Traceback (most recent call last): ... OSError: Tough luck!
[ "Decorator", "that", "retries", "the", "call", "retries", "times", "if", "func", "raises", "exceptions", ".", "Can", "use", "a", "backoff", "function", "to", "sleep", "till", "next", "retry", "." ]
train
https://github.com/ionelmc/python-aspectlib/blob/7dba7651ab76160b9ce66c9c11e770f7844e95ec/src/aspectlib/contrib.py#L9-L60
3ll3d00d/vibe
backend/src/analyser/common/uploadcontroller.py
UploadController.loadSignal
def loadSignal(self, name, start=None, end=None): """ Loads the named entry from the upload cache as a signal. :param name: the name. :param start: the time to start from in HH:mm:ss.SSS format :param end: the time to end at in HH:mm:ss.SSS format. :return: the signal if the named upload exists. """ entry = self._getCacheEntry(name) if entry is not None: from analyser.common.signal import loadSignalFromWav return loadSignalFromWav(entry['path'], start=start, end=end) else: return None
python
def loadSignal(self, name, start=None, end=None): """ Loads the named entry from the upload cache as a signal. :param name: the name. :param start: the time to start from in HH:mm:ss.SSS format :param end: the time to end at in HH:mm:ss.SSS format. :return: the signal if the named upload exists. """ entry = self._getCacheEntry(name) if entry is not None: from analyser.common.signal import loadSignalFromWav return loadSignalFromWav(entry['path'], start=start, end=end) else: return None
[ "def", "loadSignal", "(", "self", ",", "name", ",", "start", "=", "None", ",", "end", "=", "None", ")", ":", "entry", "=", "self", ".", "_getCacheEntry", "(", "name", ")", "if", "entry", "is", "not", "None", ":", "from", "analyser", ".", "common", ".", "signal", "import", "loadSignalFromWav", "return", "loadSignalFromWav", "(", "entry", "[", "'path'", "]", ",", "start", "=", "start", ",", "end", "=", "end", ")", "else", ":", "return", "None" ]
Loads the named entry from the upload cache as a signal. :param name: the name. :param start: the time to start from in HH:mm:ss.SSS format :param end: the time to end at in HH:mm:ss.SSS format. :return: the signal if the named upload exists.
[ "Loads", "the", "named", "entry", "from", "the", "upload", "cache", "as", "a", "signal", ".", ":", "param", "name", ":", "the", "name", ".", ":", "param", "start", ":", "the", "time", "to", "start", "from", "in", "HH", ":", "mm", ":", "ss", ".", "SSS", "format", ":", "param", "end", ":", "the", "time", "to", "end", "at", "in", "HH", ":", "mm", ":", "ss", ".", "SSS", "format", ".", ":", "return", ":", "the", "signal", "if", "the", "named", "upload", "exists", "." ]
train
https://github.com/3ll3d00d/vibe/blob/124b029f13ac746723e92cb47e9cb56edd2e54b5/backend/src/analyser/common/uploadcontroller.py#L41-L54
3ll3d00d/vibe
backend/src/analyser/common/uploadcontroller.py
UploadController._getCacheEntry
def _getCacheEntry(self, name): """ :param name: the name of the cache entry. :return: the entry or none. """ return next((x for x in self._uploadCache if x['name'] == name), None)
python
def _getCacheEntry(self, name): """ :param name: the name of the cache entry. :return: the entry or none. """ return next((x for x in self._uploadCache if x['name'] == name), None)
[ "def", "_getCacheEntry", "(", "self", ",", "name", ")", ":", "return", "next", "(", "(", "x", "for", "x", "in", "self", ".", "_uploadCache", "if", "x", "[", "'name'", "]", "==", "name", ")", ",", "None", ")" ]
:param name: the name of the cache entry. :return: the entry or none.
[ ":", "param", "name", ":", "the", "name", "of", "the", "cache", "entry", ".", ":", "return", ":", "the", "entry", "or", "none", "." ]
train
https://github.com/3ll3d00d/vibe/blob/124b029f13ac746723e92cb47e9cb56edd2e54b5/backend/src/analyser/common/uploadcontroller.py#L56-L61
3ll3d00d/vibe
backend/src/analyser/common/uploadcontroller.py
UploadController._convertTmp
def _convertTmp(self, tmpCacheEntry): """ Moves a tmp file to the upload dir, resampling it if necessary, and then deleting the tmp entries. :param tmpCacheEntry: the cache entry. :return: """ from analyser.common.signal import loadSignalFromWav tmpCacheEntry['status'] = 'converting' logger.info("Loading " + tmpCacheEntry['path']) signal = loadSignalFromWav(tmpCacheEntry['path']) logger.info("Loaded " + tmpCacheEntry['path']) if Path(tmpCacheEntry['path']).exists(): logger.info('Deleting ' + tmpCacheEntry['path']) os.remove(tmpCacheEntry['path']) else: logger.warning('Tmp cache file does not exist: ' + tmpCacheEntry['path']) self._tmpCache.remove(tmpCacheEntry) self._conversionCache.append(tmpCacheEntry) srcFs = signal.fs completeSamples = signal.samples outputFileName = os.path.join(self._uploadDir, tmpCacheEntry['name']) if srcFs > 1024: self.writeOutput(outputFileName, completeSamples, srcFs, 1000) else: self.writeOutput(outputFileName, completeSamples, srcFs, srcFs) tmpCacheEntry['status'] = 'loaded' self._conversionCache.remove(tmpCacheEntry) self._uploadCache.append(self._extractMeta(outputFileName, 'loaded'))
python
def _convertTmp(self, tmpCacheEntry): """ Moves a tmp file to the upload dir, resampling it if necessary, and then deleting the tmp entries. :param tmpCacheEntry: the cache entry. :return: """ from analyser.common.signal import loadSignalFromWav tmpCacheEntry['status'] = 'converting' logger.info("Loading " + tmpCacheEntry['path']) signal = loadSignalFromWav(tmpCacheEntry['path']) logger.info("Loaded " + tmpCacheEntry['path']) if Path(tmpCacheEntry['path']).exists(): logger.info('Deleting ' + tmpCacheEntry['path']) os.remove(tmpCacheEntry['path']) else: logger.warning('Tmp cache file does not exist: ' + tmpCacheEntry['path']) self._tmpCache.remove(tmpCacheEntry) self._conversionCache.append(tmpCacheEntry) srcFs = signal.fs completeSamples = signal.samples outputFileName = os.path.join(self._uploadDir, tmpCacheEntry['name']) if srcFs > 1024: self.writeOutput(outputFileName, completeSamples, srcFs, 1000) else: self.writeOutput(outputFileName, completeSamples, srcFs, srcFs) tmpCacheEntry['status'] = 'loaded' self._conversionCache.remove(tmpCacheEntry) self._uploadCache.append(self._extractMeta(outputFileName, 'loaded'))
[ "def", "_convertTmp", "(", "self", ",", "tmpCacheEntry", ")", ":", "from", "analyser", ".", "common", ".", "signal", "import", "loadSignalFromWav", "tmpCacheEntry", "[", "'status'", "]", "=", "'converting'", "logger", ".", "info", "(", "\"Loading \"", "+", "tmpCacheEntry", "[", "'path'", "]", ")", "signal", "=", "loadSignalFromWav", "(", "tmpCacheEntry", "[", "'path'", "]", ")", "logger", ".", "info", "(", "\"Loaded \"", "+", "tmpCacheEntry", "[", "'path'", "]", ")", "if", "Path", "(", "tmpCacheEntry", "[", "'path'", "]", ")", ".", "exists", "(", ")", ":", "logger", ".", "info", "(", "'Deleting '", "+", "tmpCacheEntry", "[", "'path'", "]", ")", "os", ".", "remove", "(", "tmpCacheEntry", "[", "'path'", "]", ")", "else", ":", "logger", ".", "warning", "(", "'Tmp cache file does not exist: '", "+", "tmpCacheEntry", "[", "'path'", "]", ")", "self", ".", "_tmpCache", ".", "remove", "(", "tmpCacheEntry", ")", "self", ".", "_conversionCache", ".", "append", "(", "tmpCacheEntry", ")", "srcFs", "=", "signal", ".", "fs", "completeSamples", "=", "signal", ".", "samples", "outputFileName", "=", "os", ".", "path", ".", "join", "(", "self", ".", "_uploadDir", ",", "tmpCacheEntry", "[", "'name'", "]", ")", "if", "srcFs", ">", "1024", ":", "self", ".", "writeOutput", "(", "outputFileName", ",", "completeSamples", ",", "srcFs", ",", "1000", ")", "else", ":", "self", ".", "writeOutput", "(", "outputFileName", ",", "completeSamples", ",", "srcFs", ",", "srcFs", ")", "tmpCacheEntry", "[", "'status'", "]", "=", "'loaded'", "self", ".", "_conversionCache", ".", "remove", "(", "tmpCacheEntry", ")", "self", ".", "_uploadCache", ".", "append", "(", "self", ".", "_extractMeta", "(", "outputFileName", ",", "'loaded'", ")", ")" ]
Moves a tmp file to the upload dir, resampling it if necessary, and then deleting the tmp entries. :param tmpCacheEntry: the cache entry. :return:
[ "Moves", "a", "tmp", "file", "to", "the", "upload", "dir", "resampling", "it", "if", "necessary", "and", "then", "deleting", "the", "tmp", "entries", ".", ":", "param", "tmpCacheEntry", ":", "the", "cache", "entry", ".", ":", "return", ":" ]
train
https://github.com/3ll3d00d/vibe/blob/124b029f13ac746723e92cb47e9cb56edd2e54b5/backend/src/analyser/common/uploadcontroller.py#L94-L121
3ll3d00d/vibe
backend/src/analyser/common/uploadcontroller.py
UploadController.writeChunk
def writeChunk(self, stream, filename, chunkIdx=None): """ Streams an uploaded chunk to a file. :param stream: the binary stream that contains the file. :param filename: the name of the file. :param chunkIdx: optional chunk index (for writing to a tmp dir) :return: no of bytes written or -1 if there was an error. """ import io more = True outputFileName = filename if chunkIdx is None else filename + '.' + str(chunkIdx) outputDir = self._uploadDir if chunkIdx is None else self._tmpDir chunkFilePath = os.path.join(outputDir, outputFileName) if os.path.exists(chunkFilePath) and os.path.isfile(chunkFilePath): logger.error('Uploaded file already exists: ' + chunkFilePath) return -1 else: chunkFile = open(chunkFilePath, 'xb') count = 0 while more: chunk = stream.read(io.DEFAULT_BUFFER_SIZE) chunkLen = len(chunk) count += chunkLen if chunkLen == 0: more = False else: chunkFile.write(chunk) return count
python
def writeChunk(self, stream, filename, chunkIdx=None): """ Streams an uploaded chunk to a file. :param stream: the binary stream that contains the file. :param filename: the name of the file. :param chunkIdx: optional chunk index (for writing to a tmp dir) :return: no of bytes written or -1 if there was an error. """ import io more = True outputFileName = filename if chunkIdx is None else filename + '.' + str(chunkIdx) outputDir = self._uploadDir if chunkIdx is None else self._tmpDir chunkFilePath = os.path.join(outputDir, outputFileName) if os.path.exists(chunkFilePath) and os.path.isfile(chunkFilePath): logger.error('Uploaded file already exists: ' + chunkFilePath) return -1 else: chunkFile = open(chunkFilePath, 'xb') count = 0 while more: chunk = stream.read(io.DEFAULT_BUFFER_SIZE) chunkLen = len(chunk) count += chunkLen if chunkLen == 0: more = False else: chunkFile.write(chunk) return count
[ "def", "writeChunk", "(", "self", ",", "stream", ",", "filename", ",", "chunkIdx", "=", "None", ")", ":", "import", "io", "more", "=", "True", "outputFileName", "=", "filename", "if", "chunkIdx", "is", "None", "else", "filename", "+", "'.'", "+", "str", "(", "chunkIdx", ")", "outputDir", "=", "self", ".", "_uploadDir", "if", "chunkIdx", "is", "None", "else", "self", ".", "_tmpDir", "chunkFilePath", "=", "os", ".", "path", ".", "join", "(", "outputDir", ",", "outputFileName", ")", "if", "os", ".", "path", ".", "exists", "(", "chunkFilePath", ")", "and", "os", ".", "path", ".", "isfile", "(", "chunkFilePath", ")", ":", "logger", ".", "error", "(", "'Uploaded file already exists: '", "+", "chunkFilePath", ")", "return", "-", "1", "else", ":", "chunkFile", "=", "open", "(", "chunkFilePath", ",", "'xb'", ")", "count", "=", "0", "while", "more", ":", "chunk", "=", "stream", ".", "read", "(", "io", ".", "DEFAULT_BUFFER_SIZE", ")", "chunkLen", "=", "len", "(", "chunk", ")", "count", "+=", "chunkLen", "if", "chunkLen", "==", "0", ":", "more", "=", "False", "else", ":", "chunkFile", ".", "write", "(", "chunk", ")", "return", "count" ]
Streams an uploaded chunk to a file. :param stream: the binary stream that contains the file. :param filename: the name of the file. :param chunkIdx: optional chunk index (for writing to a tmp dir) :return: no of bytes written or -1 if there was an error.
[ "Streams", "an", "uploaded", "chunk", "to", "a", "file", ".", ":", "param", "stream", ":", "the", "binary", "stream", "that", "contains", "the", "file", ".", ":", "param", "filename", ":", "the", "name", "of", "the", "file", ".", ":", "param", "chunkIdx", ":", "optional", "chunk", "index", "(", "for", "writing", "to", "a", "tmp", "dir", ")", ":", "return", ":", "no", "of", "bytes", "written", "or", "-", "1", "if", "there", "was", "an", "error", "." ]
train
https://github.com/3ll3d00d/vibe/blob/124b029f13ac746723e92cb47e9cb56edd2e54b5/backend/src/analyser/common/uploadcontroller.py#L126-L153
3ll3d00d/vibe
backend/src/analyser/common/uploadcontroller.py
UploadController.finalise
def finalise(self, filename, totalChunks, status): """ Completes the upload which means converting to a single 1kHz sample rate file output file. :param filename: :param totalChunks: :param status: :return: """ def getChunkIdx(x): try: return int(x.suffix[1:]) except ValueError: return -1 def isChunkFile(x): return x.is_file() and -1 < getChunkIdx(x) <= totalChunks asSingleFile = os.path.join(self._tmpDir, filename) if status.lower() == 'true': chunks = [(getChunkIdx(file), str(file)) for file in Path(self._tmpDir).glob(filename + '.*') if isChunkFile(file)] # TODO if len(chunks) != totalChunks then error with open(asSingleFile, 'xb') as wfd: for f in [x[1] for x in sorted(chunks, key=lambda tup: tup[0])]: with open(f, 'rb') as fd: logger.info("cat " + f + " with " + asSingleFile) shutil.copyfileobj(fd, wfd, 1024 * 1024 * 10) self.cleanupChunks(filename, isChunkFile, status)
python
def finalise(self, filename, totalChunks, status): """ Completes the upload which means converting to a single 1kHz sample rate file output file. :param filename: :param totalChunks: :param status: :return: """ def getChunkIdx(x): try: return int(x.suffix[1:]) except ValueError: return -1 def isChunkFile(x): return x.is_file() and -1 < getChunkIdx(x) <= totalChunks asSingleFile = os.path.join(self._tmpDir, filename) if status.lower() == 'true': chunks = [(getChunkIdx(file), str(file)) for file in Path(self._tmpDir).glob(filename + '.*') if isChunkFile(file)] # TODO if len(chunks) != totalChunks then error with open(asSingleFile, 'xb') as wfd: for f in [x[1] for x in sorted(chunks, key=lambda tup: tup[0])]: with open(f, 'rb') as fd: logger.info("cat " + f + " with " + asSingleFile) shutil.copyfileobj(fd, wfd, 1024 * 1024 * 10) self.cleanupChunks(filename, isChunkFile, status)
[ "def", "finalise", "(", "self", ",", "filename", ",", "totalChunks", ",", "status", ")", ":", "def", "getChunkIdx", "(", "x", ")", ":", "try", ":", "return", "int", "(", "x", ".", "suffix", "[", "1", ":", "]", ")", "except", "ValueError", ":", "return", "-", "1", "def", "isChunkFile", "(", "x", ")", ":", "return", "x", ".", "is_file", "(", ")", "and", "-", "1", "<", "getChunkIdx", "(", "x", ")", "<=", "totalChunks", "asSingleFile", "=", "os", ".", "path", ".", "join", "(", "self", ".", "_tmpDir", ",", "filename", ")", "if", "status", ".", "lower", "(", ")", "==", "'true'", ":", "chunks", "=", "[", "(", "getChunkIdx", "(", "file", ")", ",", "str", "(", "file", ")", ")", "for", "file", "in", "Path", "(", "self", ".", "_tmpDir", ")", ".", "glob", "(", "filename", "+", "'.*'", ")", "if", "isChunkFile", "(", "file", ")", "]", "# TODO if len(chunks) != totalChunks then error", "with", "open", "(", "asSingleFile", ",", "'xb'", ")", "as", "wfd", ":", "for", "f", "in", "[", "x", "[", "1", "]", "for", "x", "in", "sorted", "(", "chunks", ",", "key", "=", "lambda", "tup", ":", "tup", "[", "0", "]", ")", "]", ":", "with", "open", "(", "f", ",", "'rb'", ")", "as", "fd", ":", "logger", ".", "info", "(", "\"cat \"", "+", "f", "+", "\" with \"", "+", "asSingleFile", ")", "shutil", ".", "copyfileobj", "(", "fd", ",", "wfd", ",", "1024", "*", "1024", "*", "10", ")", "self", ".", "cleanupChunks", "(", "filename", ",", "isChunkFile", ",", "status", ")" ]
Completes the upload which means converting to a single 1kHz sample rate file output file. :param filename: :param totalChunks: :param status: :return:
[ "Completes", "the", "upload", "which", "means", "converting", "to", "a", "single", "1kHz", "sample", "rate", "file", "output", "file", ".", ":", "param", "filename", ":", ":", "param", "totalChunks", ":", ":", "param", "status", ":", ":", "return", ":" ]
train
https://github.com/3ll3d00d/vibe/blob/124b029f13ac746723e92cb47e9cb56edd2e54b5/backend/src/analyser/common/uploadcontroller.py#L155-L183