repo
stringlengths
7
55
path
stringlengths
4
127
func_name
stringlengths
1
88
original_string
stringlengths
75
19.8k
language
stringclasses
1 value
code
stringlengths
75
19.8k
code_tokens
list
docstring
stringlengths
3
17.3k
docstring_tokens
list
sha
stringlengths
40
40
url
stringlengths
87
242
partition
stringclasses
1 value
acutesoftware/AIKIF
aikif/toolbox/image_tools.py
add_text_to_image
def add_text_to_image(fname, txt, opFilename): """ convert an image by adding text """ ft = ImageFont.load("T://user//dev//src//python//_AS_LIB//timR24.pil") #wh = ft.getsize(txt) print("Adding text ", txt, " to ", fname, " pixels wide to file " , opFilename) im = Image.open(fname) draw = ImageD...
python
def add_text_to_image(fname, txt, opFilename): """ convert an image by adding text """ ft = ImageFont.load("T://user//dev//src//python//_AS_LIB//timR24.pil") #wh = ft.getsize(txt) print("Adding text ", txt, " to ", fname, " pixels wide to file " , opFilename) im = Image.open(fname) draw = ImageD...
[ "def", "add_text_to_image", "(", "fname", ",", "txt", ",", "opFilename", ")", ":", "ft", "=", "ImageFont", ".", "load", "(", "\"T://user//dev//src//python//_AS_LIB//timR24.pil\"", ")", "print", "(", "\"Adding text \"", ",", "txt", ",", "\" to \"", ",", "fname", ...
convert an image by adding text
[ "convert", "an", "image", "by", "adding", "text" ]
fcf1582dc5f884b9a4fa7c6e20e9de9d94d21d03
https://github.com/acutesoftware/AIKIF/blob/fcf1582dc5f884b9a4fa7c6e20e9de9d94d21d03/aikif/toolbox/image_tools.py#L304-L313
train
acutesoftware/AIKIF
aikif/toolbox/image_tools.py
add_crosshair_to_image
def add_crosshair_to_image(fname, opFilename): """ convert an image by adding a cross hair """ im = Image.open(fname) draw = ImageDraw.Draw(im) draw.line((0, 0) + im.size, fill=(255, 255, 255)) draw.line((0, im.size[1], im.size[0], 0), fill=(255, 255, 255)) del draw im.save(opFilename)
python
def add_crosshair_to_image(fname, opFilename): """ convert an image by adding a cross hair """ im = Image.open(fname) draw = ImageDraw.Draw(im) draw.line((0, 0) + im.size, fill=(255, 255, 255)) draw.line((0, im.size[1], im.size[0], 0), fill=(255, 255, 255)) del draw im.save(opFilename)
[ "def", "add_crosshair_to_image", "(", "fname", ",", "opFilename", ")", ":", "im", "=", "Image", ".", "open", "(", "fname", ")", "draw", "=", "ImageDraw", ".", "Draw", "(", "im", ")", "draw", ".", "line", "(", "(", "0", ",", "0", ")", "+", "im", "...
convert an image by adding a cross hair
[ "convert", "an", "image", "by", "adding", "a", "cross", "hair" ]
fcf1582dc5f884b9a4fa7c6e20e9de9d94d21d03
https://github.com/acutesoftware/AIKIF/blob/fcf1582dc5f884b9a4fa7c6e20e9de9d94d21d03/aikif/toolbox/image_tools.py#L315-L322
train
acutesoftware/AIKIF
aikif/toolbox/image_tools.py
filter_contour
def filter_contour(imageFile, opFile): """ convert an image by applying a contour """ im = Image.open(imageFile) im1 = im.filter(ImageFilter.CONTOUR) im1.save(opFile)
python
def filter_contour(imageFile, opFile): """ convert an image by applying a contour """ im = Image.open(imageFile) im1 = im.filter(ImageFilter.CONTOUR) im1.save(opFile)
[ "def", "filter_contour", "(", "imageFile", ",", "opFile", ")", ":", "im", "=", "Image", ".", "open", "(", "imageFile", ")", "im1", "=", "im", ".", "filter", "(", "ImageFilter", ".", "CONTOUR", ")", "im1", ".", "save", "(", "opFile", ")" ]
convert an image by applying a contour
[ "convert", "an", "image", "by", "applying", "a", "contour" ]
fcf1582dc5f884b9a4fa7c6e20e9de9d94d21d03
https://github.com/acutesoftware/AIKIF/blob/fcf1582dc5f884b9a4fa7c6e20e9de9d94d21d03/aikif/toolbox/image_tools.py#L324-L328
train
acutesoftware/AIKIF
aikif/toolbox/image_tools.py
get_img_hash
def get_img_hash(image, hash_size = 8): """ Grayscale and shrink the image in one step """ image = image.resize((hash_size + 1, hash_size), Image.ANTIALIAS, ) pixels = list(image.getdata()) #print('get_img_hash: pixels=', pixels) # Compare adjacent pixels. difference = [] for row in range...
python
def get_img_hash(image, hash_size = 8): """ Grayscale and shrink the image in one step """ image = image.resize((hash_size + 1, hash_size), Image.ANTIALIAS, ) pixels = list(image.getdata()) #print('get_img_hash: pixels=', pixels) # Compare adjacent pixels. difference = [] for row in range...
[ "def", "get_img_hash", "(", "image", ",", "hash_size", "=", "8", ")", ":", "image", "=", "image", ".", "resize", "(", "(", "hash_size", "+", "1", ",", "hash_size", ")", ",", "Image", ".", "ANTIALIAS", ",", ")", "pixels", "=", "list", "(", "image", ...
Grayscale and shrink the image in one step
[ "Grayscale", "and", "shrink", "the", "image", "in", "one", "step" ]
fcf1582dc5f884b9a4fa7c6e20e9de9d94d21d03
https://github.com/acutesoftware/AIKIF/blob/fcf1582dc5f884b9a4fa7c6e20e9de9d94d21d03/aikif/toolbox/image_tools.py#L374-L400
train
acutesoftware/AIKIF
aikif/toolbox/image_tools.py
load_image
def load_image(fname): """ read an image from file - PIL doesnt close nicely """ with open(fname, "rb") as f: i = Image.open(fname) #i.load() return i
python
def load_image(fname): """ read an image from file - PIL doesnt close nicely """ with open(fname, "rb") as f: i = Image.open(fname) #i.load() return i
[ "def", "load_image", "(", "fname", ")", ":", "with", "open", "(", "fname", ",", "\"rb\"", ")", "as", "f", ":", "i", "=", "Image", ".", "open", "(", "fname", ")", "return", "i" ]
read an image from file - PIL doesnt close nicely
[ "read", "an", "image", "from", "file", "-", "PIL", "doesnt", "close", "nicely" ]
fcf1582dc5f884b9a4fa7c6e20e9de9d94d21d03
https://github.com/acutesoftware/AIKIF/blob/fcf1582dc5f884b9a4fa7c6e20e9de9d94d21d03/aikif/toolbox/image_tools.py#L402-L407
train
acutesoftware/AIKIF
aikif/toolbox/image_tools.py
dump_img
def dump_img(fname): """ output the image as text """ img = Image.open(fname) width, _ = img.size txt = '' pixels = list(img.getdata()) for col in range(width): txt += str(pixels[col:col+width]) return txt
python
def dump_img(fname): """ output the image as text """ img = Image.open(fname) width, _ = img.size txt = '' pixels = list(img.getdata()) for col in range(width): txt += str(pixels[col:col+width]) return txt
[ "def", "dump_img", "(", "fname", ")", ":", "img", "=", "Image", ".", "open", "(", "fname", ")", "width", ",", "_", "=", "img", ".", "size", "txt", "=", "''", "pixels", "=", "list", "(", "img", ".", "getdata", "(", ")", ")", "for", "col", "in", ...
output the image as text
[ "output", "the", "image", "as", "text" ]
fcf1582dc5f884b9a4fa7c6e20e9de9d94d21d03
https://github.com/acutesoftware/AIKIF/blob/fcf1582dc5f884b9a4fa7c6e20e9de9d94d21d03/aikif/toolbox/image_tools.py#L409-L417
train
mpg-age-bioinformatics/AGEpy
AGEpy/plots.py
NormInt
def NormInt(df,sampleA,sampleB): """ Normalizes intensities of a gene in two samples :param df: dataframe output of GetData() :param sampleA: column header of sample A :param sampleB: column header of sample B :returns: normalized intensities """ c1=df[sampleA] c2=df[sampleB] ...
python
def NormInt(df,sampleA,sampleB): """ Normalizes intensities of a gene in two samples :param df: dataframe output of GetData() :param sampleA: column header of sample A :param sampleB: column header of sample B :returns: normalized intensities """ c1=df[sampleA] c2=df[sampleB] ...
[ "def", "NormInt", "(", "df", ",", "sampleA", ",", "sampleB", ")", ":", "c1", "=", "df", "[", "sampleA", "]", "c2", "=", "df", "[", "sampleB", "]", "return", "np", ".", "log10", "(", "np", ".", "sqrt", "(", "c1", "*", "c2", ")", ")" ]
Normalizes intensities of a gene in two samples :param df: dataframe output of GetData() :param sampleA: column header of sample A :param sampleB: column header of sample B :returns: normalized intensities
[ "Normalizes", "intensities", "of", "a", "gene", "in", "two", "samples" ]
887808a7a2c1504f39ce8d8cb36c15c1721cd29f
https://github.com/mpg-age-bioinformatics/AGEpy/blob/887808a7a2c1504f39ce8d8cb36c15c1721cd29f/AGEpy/plots.py#L341-L354
train
Nachtfeuer/pipeline
examples/python/primes/demo/primes.py
is_prime
def is_prime(number): """ Testing given number to be a prime. >>> [n for n in range(100+1) if is_prime(n)] [2, 3, 5, 7, 11, 13, 17, 19, 23, 29, 31, 37, 41, 43, 47, 53, 59, 61, 67, 71, 73, 79, 83, 89, 97] """ if number < 2: return False if number % 2 == 0: return number == 2 ...
python
def is_prime(number): """ Testing given number to be a prime. >>> [n for n in range(100+1) if is_prime(n)] [2, 3, 5, 7, 11, 13, 17, 19, 23, 29, 31, 37, 41, 43, 47, 53, 59, 61, 67, 71, 73, 79, 83, 89, 97] """ if number < 2: return False if number % 2 == 0: return number == 2 ...
[ "def", "is_prime", "(", "number", ")", ":", "if", "number", "<", "2", ":", "return", "False", "if", "number", "%", "2", "==", "0", ":", "return", "number", "==", "2", "limit", "=", "int", "(", "math", ".", "sqrt", "(", "number", ")", ")", "for", ...
Testing given number to be a prime. >>> [n for n in range(100+1) if is_prime(n)] [2, 3, 5, 7, 11, 13, 17, 19, 23, 29, 31, 37, 41, 43, 47, 53, 59, 61, 67, 71, 73, 79, 83, 89, 97]
[ "Testing", "given", "number", "to", "be", "a", "prime", "." ]
04ca18c4e95e4349532bb45b768206393e1f2c13
https://github.com/Nachtfeuer/pipeline/blob/04ca18c4e95e4349532bb45b768206393e1f2c13/examples/python/primes/demo/primes.py#L32-L49
train
OpenHydrology/floodestimation
floodestimation/analysis.py
QmedAnalysis.qmed_all_methods
def qmed_all_methods(self): """ Returns a dict of QMED methods using all available methods. Available methods are defined in :attr:`qmed_methods`. The returned dict keys contain the method name, e.g. `amax_record` with value representing the corresponding QMED estimate in m³/s. ...
python
def qmed_all_methods(self): """ Returns a dict of QMED methods using all available methods. Available methods are defined in :attr:`qmed_methods`. The returned dict keys contain the method name, e.g. `amax_record` with value representing the corresponding QMED estimate in m³/s. ...
[ "def", "qmed_all_methods", "(", "self", ")", ":", "result", "=", "{", "}", "for", "method", "in", "self", ".", "methods", ":", "try", ":", "result", "[", "method", "]", "=", "getattr", "(", "self", ",", "'_qmed_from_'", "+", "method", ")", "(", ")", ...
Returns a dict of QMED methods using all available methods. Available methods are defined in :attr:`qmed_methods`. The returned dict keys contain the method name, e.g. `amax_record` with value representing the corresponding QMED estimate in m³/s. :return: dict of QMED estimates :rtype:...
[ "Returns", "a", "dict", "of", "QMED", "methods", "using", "all", "available", "methods", "." ]
782da7c5abd1348923129efe89fb70003ebb088c
https://github.com/OpenHydrology/floodestimation/blob/782da7c5abd1348923129efe89fb70003ebb088c/floodestimation/analysis.py#L169-L185
train
OpenHydrology/floodestimation
floodestimation/analysis.py
QmedAnalysis._qmed_from_amax_records
def _qmed_from_amax_records(self): """ Return QMED estimate based on annual maximum flow records. :return: QMED in m³/s :rtype: float """ valid_flows = valid_flows_array(self.catchment) n = len(valid_flows) if n < 2: raise InsufficientDataErro...
python
def _qmed_from_amax_records(self): """ Return QMED estimate based on annual maximum flow records. :return: QMED in m³/s :rtype: float """ valid_flows = valid_flows_array(self.catchment) n = len(valid_flows) if n < 2: raise InsufficientDataErro...
[ "def", "_qmed_from_amax_records", "(", "self", ")", ":", "valid_flows", "=", "valid_flows_array", "(", "self", ".", "catchment", ")", "n", "=", "len", "(", "valid_flows", ")", "if", "n", "<", "2", ":", "raise", "InsufficientDataError", "(", "\"Insufficient ann...
Return QMED estimate based on annual maximum flow records. :return: QMED in m³/s :rtype: float
[ "Return", "QMED", "estimate", "based", "on", "annual", "maximum", "flow", "records", "." ]
782da7c5abd1348923129efe89fb70003ebb088c
https://github.com/OpenHydrology/floodestimation/blob/782da7c5abd1348923129efe89fb70003ebb088c/floodestimation/analysis.py#L187-L199
train
OpenHydrology/floodestimation
floodestimation/analysis.py
QmedAnalysis._pot_month_counts
def _pot_month_counts(self, pot_dataset): """ Return a list of 12 sets. Each sets contains the years included in the POT record period. :param pot_dataset: POT dataset (records and meta data) :type pot_dataset: :class:`floodestimation.entities.PotDataset` """ periods = p...
python
def _pot_month_counts(self, pot_dataset): """ Return a list of 12 sets. Each sets contains the years included in the POT record period. :param pot_dataset: POT dataset (records and meta data) :type pot_dataset: :class:`floodestimation.entities.PotDataset` """ periods = p...
[ "def", "_pot_month_counts", "(", "self", ",", "pot_dataset", ")", ":", "periods", "=", "pot_dataset", ".", "continuous_periods", "(", ")", "result", "=", "[", "set", "(", ")", "for", "x", "in", "range", "(", "12", ")", "]", "for", "period", "in", "peri...
Return a list of 12 sets. Each sets contains the years included in the POT record period. :param pot_dataset: POT dataset (records and meta data) :type pot_dataset: :class:`floodestimation.entities.PotDataset`
[ "Return", "a", "list", "of", "12", "sets", ".", "Each", "sets", "contains", "the", "years", "included", "in", "the", "POT", "record", "period", "." ]
782da7c5abd1348923129efe89fb70003ebb088c
https://github.com/OpenHydrology/floodestimation/blob/782da7c5abd1348923129efe89fb70003ebb088c/floodestimation/analysis.py#L229-L252
train
OpenHydrology/floodestimation
floodestimation/analysis.py
QmedAnalysis._qmed_from_area
def _qmed_from_area(self): """ Return QMED estimate based on catchment area. TODO: add source of method :return: QMED in m³/s :rtype: float """ try: return 1.172 * self.catchment.descriptors.dtm_area ** self._area_exponent() # Area in km² ex...
python
def _qmed_from_area(self): """ Return QMED estimate based on catchment area. TODO: add source of method :return: QMED in m³/s :rtype: float """ try: return 1.172 * self.catchment.descriptors.dtm_area ** self._area_exponent() # Area in km² ex...
[ "def", "_qmed_from_area", "(", "self", ")", ":", "try", ":", "return", "1.172", "*", "self", ".", "catchment", ".", "descriptors", ".", "dtm_area", "**", "self", ".", "_area_exponent", "(", ")", "except", "(", "TypeError", ",", "KeyError", ")", ":", "rai...
Return QMED estimate based on catchment area. TODO: add source of method :return: QMED in m³/s :rtype: float
[ "Return", "QMED", "estimate", "based", "on", "catchment", "area", "." ]
782da7c5abd1348923129efe89fb70003ebb088c
https://github.com/OpenHydrology/floodestimation/blob/782da7c5abd1348923129efe89fb70003ebb088c/floodestimation/analysis.py#L310-L322
train
OpenHydrology/floodestimation
floodestimation/analysis.py
QmedAnalysis._qmed_from_descriptors_1999
def _qmed_from_descriptors_1999(self, as_rural=False): """ Return QMED estimation based on FEH catchment descriptors, 1999 methodology. Methodology source: FEH, Vol. 3, p. 14 :param as_rural: assume catchment is fully rural. Default: false. :type as_rural: bool :return:...
python
def _qmed_from_descriptors_1999(self, as_rural=False): """ Return QMED estimation based on FEH catchment descriptors, 1999 methodology. Methodology source: FEH, Vol. 3, p. 14 :param as_rural: assume catchment is fully rural. Default: false. :type as_rural: bool :return:...
[ "def", "_qmed_from_descriptors_1999", "(", "self", ",", "as_rural", "=", "False", ")", ":", "try", ":", "qmed_rural", "=", "1.172", "*", "self", ".", "catchment", ".", "descriptors", ".", "dtm_area", "**", "self", ".", "_area_exponent", "(", ")", "*", "(",...
Return QMED estimation based on FEH catchment descriptors, 1999 methodology. Methodology source: FEH, Vol. 3, p. 14 :param as_rural: assume catchment is fully rural. Default: false. :type as_rural: bool :return: QMED in m³/s :rtype: float
[ "Return", "QMED", "estimation", "based", "on", "FEH", "catchment", "descriptors", "1999", "methodology", "." ]
782da7c5abd1348923129efe89fb70003ebb088c
https://github.com/OpenHydrology/floodestimation/blob/782da7c5abd1348923129efe89fb70003ebb088c/floodestimation/analysis.py#L324-L346
train
OpenHydrology/floodestimation
floodestimation/analysis.py
QmedAnalysis._qmed_from_descriptors_2008
def _qmed_from_descriptors_2008(self, as_rural=False, donor_catchments=None): """ Return QMED estimation based on FEH catchment descriptors, 2008 methodology. Methodology source: Science Report SC050050, p. 36 :param as_rural: assume catchment is fully rural. Default: false. :t...
python
def _qmed_from_descriptors_2008(self, as_rural=False, donor_catchments=None): """ Return QMED estimation based on FEH catchment descriptors, 2008 methodology. Methodology source: Science Report SC050050, p. 36 :param as_rural: assume catchment is fully rural. Default: false. :t...
[ "def", "_qmed_from_descriptors_2008", "(", "self", ",", "as_rural", "=", "False", ",", "donor_catchments", "=", "None", ")", ":", "try", ":", "lnqmed_rural", "=", "2.1170", "+", "0.8510", "*", "log", "(", "self", ".", "catchment", ".", "descriptors", ".", ...
Return QMED estimation based on FEH catchment descriptors, 2008 methodology. Methodology source: Science Report SC050050, p. 36 :param as_rural: assume catchment is fully rural. Default: false. :type as rural: bool :param donor_catchments: override donor catchment to improve QMED catch...
[ "Return", "QMED", "estimation", "based", "on", "FEH", "catchment", "descriptors", "2008", "methodology", "." ]
782da7c5abd1348923129efe89fb70003ebb088c
https://github.com/OpenHydrology/floodestimation/blob/782da7c5abd1348923129efe89fb70003ebb088c/floodestimation/analysis.py#L359-L416
train
OpenHydrology/floodestimation
floodestimation/analysis.py
QmedAnalysis._pruaf
def _pruaf(self): """ Return percentage runoff urban adjustment factor. Methodology source: eqn. 6, Kjeldsen 2010 """ return 1 + 0.47 * self.catchment.descriptors.urbext(self.year) \ * self.catchment.descriptors.bfihost / (1 - self.catchment.descriptors.bfihos...
python
def _pruaf(self): """ Return percentage runoff urban adjustment factor. Methodology source: eqn. 6, Kjeldsen 2010 """ return 1 + 0.47 * self.catchment.descriptors.urbext(self.year) \ * self.catchment.descriptors.bfihost / (1 - self.catchment.descriptors.bfihos...
[ "def", "_pruaf", "(", "self", ")", ":", "return", "1", "+", "0.47", "*", "self", ".", "catchment", ".", "descriptors", ".", "urbext", "(", "self", ".", "year", ")", "*", "self", ".", "catchment", ".", "descriptors", ".", "bfihost", "/", "(", "1", "...
Return percentage runoff urban adjustment factor. Methodology source: eqn. 6, Kjeldsen 2010
[ "Return", "percentage", "runoff", "urban", "adjustment", "factor", "." ]
782da7c5abd1348923129efe89fb70003ebb088c
https://github.com/OpenHydrology/floodestimation/blob/782da7c5abd1348923129efe89fb70003ebb088c/floodestimation/analysis.py#L418-L425
train
OpenHydrology/floodestimation
floodestimation/analysis.py
QmedAnalysis._dist_corr
def _dist_corr(dist, phi1, phi2, phi3): """ Generic distance-decaying correlation function :param dist: Distance between catchment centrolds in km :type dist: float :param phi1: Decay function parameters 1 :type phi1: float :param phi2: Decay function parameters ...
python
def _dist_corr(dist, phi1, phi2, phi3): """ Generic distance-decaying correlation function :param dist: Distance between catchment centrolds in km :type dist: float :param phi1: Decay function parameters 1 :type phi1: float :param phi2: Decay function parameters ...
[ "def", "_dist_corr", "(", "dist", ",", "phi1", ",", "phi2", ",", "phi3", ")", ":", "return", "phi1", "*", "exp", "(", "-", "phi2", "*", "dist", ")", "+", "(", "1", "-", "phi1", ")", "*", "exp", "(", "-", "phi3", "*", "dist", ")" ]
Generic distance-decaying correlation function :param dist: Distance between catchment centrolds in km :type dist: float :param phi1: Decay function parameters 1 :type phi1: float :param phi2: Decay function parameters 2 :type phi2: float :param phi3: Decay funct...
[ "Generic", "distance", "-", "decaying", "correlation", "function" ]
782da7c5abd1348923129efe89fb70003ebb088c
https://github.com/OpenHydrology/floodestimation/blob/782da7c5abd1348923129efe89fb70003ebb088c/floodestimation/analysis.py#L443-L458
train
OpenHydrology/floodestimation
floodestimation/analysis.py
QmedAnalysis._vec_b
def _vec_b(self, donor_catchments): """ Return vector ``b`` of model error covariances to estimate weights Methodology source: Kjeldsen, Jones and Morris, 2009, eqs 3 and 10 :param donor_catchments: Catchments to use as donors :type donor_catchments: list of :class:`Catchment` ...
python
def _vec_b(self, donor_catchments): """ Return vector ``b`` of model error covariances to estimate weights Methodology source: Kjeldsen, Jones and Morris, 2009, eqs 3 and 10 :param donor_catchments: Catchments to use as donors :type donor_catchments: list of :class:`Catchment` ...
[ "def", "_vec_b", "(", "self", ",", "donor_catchments", ")", ":", "p", "=", "len", "(", "donor_catchments", ")", "b", "=", "0.1175", "*", "np", ".", "ones", "(", "p", ")", "for", "i", "in", "range", "(", "p", ")", ":", "b", "[", "i", "]", "*=", ...
Return vector ``b`` of model error covariances to estimate weights Methodology source: Kjeldsen, Jones and Morris, 2009, eqs 3 and 10 :param donor_catchments: Catchments to use as donors :type donor_catchments: list of :class:`Catchment` :return: Model error covariance vector :...
[ "Return", "vector", "b", "of", "model", "error", "covariances", "to", "estimate", "weights" ]
782da7c5abd1348923129efe89fb70003ebb088c
https://github.com/OpenHydrology/floodestimation/blob/782da7c5abd1348923129efe89fb70003ebb088c/floodestimation/analysis.py#L492-L507
train
OpenHydrology/floodestimation
floodestimation/analysis.py
QmedAnalysis._beta
def _beta(catchment): """ Return beta, the GLO scale parameter divided by loc parameter estimated using simple regression model Methodology source: Kjeldsen & Jones, 2009, table 2 :param catchment: Catchment to estimate beta for :type catchment: :class:`Catchment` :retu...
python
def _beta(catchment): """ Return beta, the GLO scale parameter divided by loc parameter estimated using simple regression model Methodology source: Kjeldsen & Jones, 2009, table 2 :param catchment: Catchment to estimate beta for :type catchment: :class:`Catchment` :retu...
[ "def", "_beta", "(", "catchment", ")", ":", "lnbeta", "=", "-", "1.1221", "-", "0.0816", "*", "log", "(", "catchment", ".", "descriptors", ".", "dtm_area", ")", "-", "0.4580", "*", "log", "(", "catchment", ".", "descriptors", ".", "saar", "/", "1000", ...
Return beta, the GLO scale parameter divided by loc parameter estimated using simple regression model Methodology source: Kjeldsen & Jones, 2009, table 2 :param catchment: Catchment to estimate beta for :type catchment: :class:`Catchment` :return: beta :rtype: float
[ "Return", "beta", "the", "GLO", "scale", "parameter", "divided", "by", "loc", "parameter", "estimated", "using", "simple", "regression", "model" ]
782da7c5abd1348923129efe89fb70003ebb088c
https://github.com/OpenHydrology/floodestimation/blob/782da7c5abd1348923129efe89fb70003ebb088c/floodestimation/analysis.py#L510-L525
train
OpenHydrology/floodestimation
floodestimation/analysis.py
QmedAnalysis._matrix_sigma_eta
def _matrix_sigma_eta(self, donor_catchments): """ Return model error coveriance matrix Sigma eta Methodology source: Kjelsen, Jones & Morris 2014, eqs 2 and 3 :param donor_catchments: Catchments to use as donors :type donor_catchments: list of :class:`Catchment` :retur...
python
def _matrix_sigma_eta(self, donor_catchments): """ Return model error coveriance matrix Sigma eta Methodology source: Kjelsen, Jones & Morris 2014, eqs 2 and 3 :param donor_catchments: Catchments to use as donors :type donor_catchments: list of :class:`Catchment` :retur...
[ "def", "_matrix_sigma_eta", "(", "self", ",", "donor_catchments", ")", ":", "p", "=", "len", "(", "donor_catchments", ")", "sigma", "=", "0.1175", "*", "np", ".", "ones", "(", "(", "p", ",", "p", ")", ")", "for", "i", "in", "range", "(", "p", ")", ...
Return model error coveriance matrix Sigma eta Methodology source: Kjelsen, Jones & Morris 2014, eqs 2 and 3 :param donor_catchments: Catchments to use as donors :type donor_catchments: list of :class:`Catchment` :return: 2-Dimensional, symmetric covariance matrix :rtype: :clas...
[ "Return", "model", "error", "coveriance", "matrix", "Sigma", "eta" ]
782da7c5abd1348923129efe89fb70003ebb088c
https://github.com/OpenHydrology/floodestimation/blob/782da7c5abd1348923129efe89fb70003ebb088c/floodestimation/analysis.py#L527-L544
train
OpenHydrology/floodestimation
floodestimation/analysis.py
QmedAnalysis._matrix_sigma_eps
def _matrix_sigma_eps(self, donor_catchments): """ Return sampling error coveriance matrix Sigma eta Methodology source: Kjeldsen & Jones 2009, eq 9 :param donor_catchments: Catchments to use as donors :type donor_catchments: list of :class:`Catchment` :return: 2-Dimens...
python
def _matrix_sigma_eps(self, donor_catchments): """ Return sampling error coveriance matrix Sigma eta Methodology source: Kjeldsen & Jones 2009, eq 9 :param donor_catchments: Catchments to use as donors :type donor_catchments: list of :class:`Catchment` :return: 2-Dimens...
[ "def", "_matrix_sigma_eps", "(", "self", ",", "donor_catchments", ")", ":", "p", "=", "len", "(", "donor_catchments", ")", "sigma", "=", "np", ".", "empty", "(", "(", "p", ",", "p", ")", ")", "for", "i", "in", "range", "(", "p", ")", ":", "beta_i",...
Return sampling error coveriance matrix Sigma eta Methodology source: Kjeldsen & Jones 2009, eq 9 :param donor_catchments: Catchments to use as donors :type donor_catchments: list of :class:`Catchment` :return: 2-Dimensional, symmetric covariance matrix :rtype: :class:`numpy.nd...
[ "Return", "sampling", "error", "coveriance", "matrix", "Sigma", "eta" ]
782da7c5abd1348923129efe89fb70003ebb088c
https://github.com/OpenHydrology/floodestimation/blob/782da7c5abd1348923129efe89fb70003ebb088c/floodestimation/analysis.py#L546-L569
train
OpenHydrology/floodestimation
floodestimation/analysis.py
QmedAnalysis._vec_alpha
def _vec_alpha(self, donor_catchments): """ Return vector alpha which is the weights for donor model errors Methodology source: Kjeldsen, Jones & Morris 2014, eq 10 :param donor_catchments: Catchments to use as donors :type donor_catchments: list of :class:`Catchment` :...
python
def _vec_alpha(self, donor_catchments): """ Return vector alpha which is the weights for donor model errors Methodology source: Kjeldsen, Jones & Morris 2014, eq 10 :param donor_catchments: Catchments to use as donors :type donor_catchments: list of :class:`Catchment` :...
[ "def", "_vec_alpha", "(", "self", ",", "donor_catchments", ")", ":", "return", "np", ".", "dot", "(", "linalg", ".", "inv", "(", "self", ".", "_matrix_omega", "(", "donor_catchments", ")", ")", ",", "self", ".", "_vec_b", "(", "donor_catchments", ")", ")...
Return vector alpha which is the weights for donor model errors Methodology source: Kjeldsen, Jones & Morris 2014, eq 10 :param donor_catchments: Catchments to use as donors :type donor_catchments: list of :class:`Catchment` :return: Vector of donor weights :rtype: :class:`nump...
[ "Return", "vector", "alpha", "which", "is", "the", "weights", "for", "donor", "model", "errors" ]
782da7c5abd1348923129efe89fb70003ebb088c
https://github.com/OpenHydrology/floodestimation/blob/782da7c5abd1348923129efe89fb70003ebb088c/floodestimation/analysis.py#L574-L585
train
OpenHydrology/floodestimation
floodestimation/analysis.py
QmedAnalysis.find_donor_catchments
def find_donor_catchments(self, limit=6, dist_limit=500): """ Return a suitable donor catchment to improve a QMED estimate based on catchment descriptors alone. :param limit: maximum number of catchments to return. Default: 6. Set to `None` to return all available catchmen...
python
def find_donor_catchments(self, limit=6, dist_limit=500): """ Return a suitable donor catchment to improve a QMED estimate based on catchment descriptors alone. :param limit: maximum number of catchments to return. Default: 6. Set to `None` to return all available catchmen...
[ "def", "find_donor_catchments", "(", "self", ",", "limit", "=", "6", ",", "dist_limit", "=", "500", ")", ":", "if", "self", ".", "gauged_catchments", ":", "return", "self", ".", "gauged_catchments", ".", "nearest_qmed_catchments", "(", "self", ".", "catchment"...
Return a suitable donor catchment to improve a QMED estimate based on catchment descriptors alone. :param limit: maximum number of catchments to return. Default: 6. Set to `None` to return all available catchments. :type limit: int :param dist_limit: maximum distance in km...
[ "Return", "a", "suitable", "donor", "catchment", "to", "improve", "a", "QMED", "estimate", "based", "on", "catchment", "descriptors", "alone", "." ]
782da7c5abd1348923129efe89fb70003ebb088c
https://github.com/OpenHydrology/floodestimation/blob/782da7c5abd1348923129efe89fb70003ebb088c/floodestimation/analysis.py#L616-L632
train
OpenHydrology/floodestimation
floodestimation/analysis.py
GrowthCurveAnalysis._var_and_skew
def _var_and_skew(self, catchments, as_rural=False): """ Calculate L-CV and L-SKEW from a single catchment or a pooled group of catchments. Methodology source: Science Report SC050050, para. 6.4.1-6.4.2 """ if not hasattr(catchments, '__getitem__'): # In case of a single catchm...
python
def _var_and_skew(self, catchments, as_rural=False): """ Calculate L-CV and L-SKEW from a single catchment or a pooled group of catchments. Methodology source: Science Report SC050050, para. 6.4.1-6.4.2 """ if not hasattr(catchments, '__getitem__'): # In case of a single catchm...
[ "def", "_var_and_skew", "(", "self", ",", "catchments", ",", "as_rural", "=", "False", ")", ":", "if", "not", "hasattr", "(", "catchments", ",", "'__getitem__'", ")", ":", "l_cv", ",", "l_skew", "=", "self", ".", "_l_cv_and_skew", "(", "self", ".", "catc...
Calculate L-CV and L-SKEW from a single catchment or a pooled group of catchments. Methodology source: Science Report SC050050, para. 6.4.1-6.4.2
[ "Calculate", "L", "-", "CV", "and", "L", "-", "SKEW", "from", "a", "single", "catchment", "or", "a", "pooled", "group", "of", "catchments", "." ]
782da7c5abd1348923129efe89fb70003ebb088c
https://github.com/OpenHydrology/floodestimation/blob/782da7c5abd1348923129efe89fb70003ebb088c/floodestimation/analysis.py#L713-L770
train
OpenHydrology/floodestimation
floodestimation/analysis.py
GrowthCurveAnalysis._l_cv_and_skew
def _l_cv_and_skew(self, catchment): """ Calculate L-CV and L-SKEW for a gauged catchment. Uses `lmoments3` library. Methodology source: Science Report SC050050, para. 6.7.5 """ z = self._dimensionless_flows(catchment) l1, l2, t3 = lm.lmom_ratios(z, nmom=3) retur...
python
def _l_cv_and_skew(self, catchment): """ Calculate L-CV and L-SKEW for a gauged catchment. Uses `lmoments3` library. Methodology source: Science Report SC050050, para. 6.7.5 """ z = self._dimensionless_flows(catchment) l1, l2, t3 = lm.lmom_ratios(z, nmom=3) retur...
[ "def", "_l_cv_and_skew", "(", "self", ",", "catchment", ")", ":", "z", "=", "self", ".", "_dimensionless_flows", "(", "catchment", ")", "l1", ",", "l2", ",", "t3", "=", "lm", ".", "lmom_ratios", "(", "z", ",", "nmom", "=", "3", ")", "return", "l2", ...
Calculate L-CV and L-SKEW for a gauged catchment. Uses `lmoments3` library. Methodology source: Science Report SC050050, para. 6.7.5
[ "Calculate", "L", "-", "CV", "and", "L", "-", "SKEW", "for", "a", "gauged", "catchment", ".", "Uses", "lmoments3", "library", "." ]
782da7c5abd1348923129efe89fb70003ebb088c
https://github.com/OpenHydrology/floodestimation/blob/782da7c5abd1348923129efe89fb70003ebb088c/floodestimation/analysis.py#L772-L780
train
OpenHydrology/floodestimation
floodestimation/analysis.py
GrowthCurveAnalysis._l_cv_weight
def _l_cv_weight(self, donor_catchment): """ Return L-CV weighting for a donor catchment. Methodology source: Science Report SC050050, eqn. 6.18 and 6.22a """ try: dist = donor_catchment.similarity_dist except AttributeError: dist = self._similari...
python
def _l_cv_weight(self, donor_catchment): """ Return L-CV weighting for a donor catchment. Methodology source: Science Report SC050050, eqn. 6.18 and 6.22a """ try: dist = donor_catchment.similarity_dist except AttributeError: dist = self._similari...
[ "def", "_l_cv_weight", "(", "self", ",", "donor_catchment", ")", ":", "try", ":", "dist", "=", "donor_catchment", ".", "similarity_dist", "except", "AttributeError", ":", "dist", "=", "self", ".", "_similarity_distance", "(", "self", ".", "catchment", ",", "do...
Return L-CV weighting for a donor catchment. Methodology source: Science Report SC050050, eqn. 6.18 and 6.22a
[ "Return", "L", "-", "CV", "weighting", "for", "a", "donor", "catchment", "." ]
782da7c5abd1348923129efe89fb70003ebb088c
https://github.com/OpenHydrology/floodestimation/blob/782da7c5abd1348923129efe89fb70003ebb088c/floodestimation/analysis.py#L782-L794
train
OpenHydrology/floodestimation
floodestimation/analysis.py
GrowthCurveAnalysis._l_cv_weight_factor
def _l_cv_weight_factor(self): """ Return multiplier for L-CV weightings in case of enhanced single site analysis. Methodology source: Science Report SC050050, eqn. 6.15a and 6.15b """ b = 0.0047 * sqrt(0) + 0.0023 / 2 c = 0.02609 / (self.catchment.record_length - 1) ...
python
def _l_cv_weight_factor(self): """ Return multiplier for L-CV weightings in case of enhanced single site analysis. Methodology source: Science Report SC050050, eqn. 6.15a and 6.15b """ b = 0.0047 * sqrt(0) + 0.0023 / 2 c = 0.02609 / (self.catchment.record_length - 1) ...
[ "def", "_l_cv_weight_factor", "(", "self", ")", ":", "b", "=", "0.0047", "*", "sqrt", "(", "0", ")", "+", "0.0023", "/", "2", "c", "=", "0.02609", "/", "(", "self", ".", "catchment", ".", "record_length", "-", "1", ")", "return", "c", "/", "(", "...
Return multiplier for L-CV weightings in case of enhanced single site analysis. Methodology source: Science Report SC050050, eqn. 6.15a and 6.15b
[ "Return", "multiplier", "for", "L", "-", "CV", "weightings", "in", "case", "of", "enhanced", "single", "site", "analysis", "." ]
782da7c5abd1348923129efe89fb70003ebb088c
https://github.com/OpenHydrology/floodestimation/blob/782da7c5abd1348923129efe89fb70003ebb088c/floodestimation/analysis.py#L796-L804
train
OpenHydrology/floodestimation
floodestimation/analysis.py
GrowthCurveAnalysis._l_skew_weight
def _l_skew_weight(self, donor_catchment): """ Return L-SKEW weighting for donor catchment. Methodology source: Science Report SC050050, eqn. 6.19 and 6.22b """ try: dist = donor_catchment.similarity_dist except AttributeError: dist = self._simila...
python
def _l_skew_weight(self, donor_catchment): """ Return L-SKEW weighting for donor catchment. Methodology source: Science Report SC050050, eqn. 6.19 and 6.22b """ try: dist = donor_catchment.similarity_dist except AttributeError: dist = self._simila...
[ "def", "_l_skew_weight", "(", "self", ",", "donor_catchment", ")", ":", "try", ":", "dist", "=", "donor_catchment", ".", "similarity_dist", "except", "AttributeError", ":", "dist", "=", "self", ".", "_similarity_distance", "(", "self", ".", "catchment", ",", "...
Return L-SKEW weighting for donor catchment. Methodology source: Science Report SC050050, eqn. 6.19 and 6.22b
[ "Return", "L", "-", "SKEW", "weighting", "for", "donor", "catchment", "." ]
782da7c5abd1348923129efe89fb70003ebb088c
https://github.com/OpenHydrology/floodestimation/blob/782da7c5abd1348923129efe89fb70003ebb088c/floodestimation/analysis.py#L806-L818
train
OpenHydrology/floodestimation
floodestimation/analysis.py
GrowthCurveAnalysis._growth_curve_single_site
def _growth_curve_single_site(self, distr='glo'): """ Return flood growth curve function based on `amax_records` from the subject catchment only. :return: Inverse cumulative distribution function with one parameter `aep` (annual exceedance probability) :type: :class:`.GrowthCurve` ...
python
def _growth_curve_single_site(self, distr='glo'): """ Return flood growth curve function based on `amax_records` from the subject catchment only. :return: Inverse cumulative distribution function with one parameter `aep` (annual exceedance probability) :type: :class:`.GrowthCurve` ...
[ "def", "_growth_curve_single_site", "(", "self", ",", "distr", "=", "'glo'", ")", ":", "if", "self", ".", "catchment", ".", "amax_records", ":", "self", ".", "donor_catchments", "=", "[", "]", "return", "GrowthCurve", "(", "distr", ",", "*", "self", ".", ...
Return flood growth curve function based on `amax_records` from the subject catchment only. :return: Inverse cumulative distribution function with one parameter `aep` (annual exceedance probability) :type: :class:`.GrowthCurve`
[ "Return", "flood", "growth", "curve", "function", "based", "on", "amax_records", "from", "the", "subject", "catchment", "only", "." ]
782da7c5abd1348923129efe89fb70003ebb088c
https://github.com/OpenHydrology/floodestimation/blob/782da7c5abd1348923129efe89fb70003ebb088c/floodestimation/analysis.py#L820-L831
train
OpenHydrology/floodestimation
floodestimation/analysis.py
GrowthCurveAnalysis._growth_curve_pooling_group
def _growth_curve_pooling_group(self, distr='glo', as_rural=False): """ Return flood growth curve function based on `amax_records` from a pooling group. :return: Inverse cumulative distribution function with one parameter `aep` (annual exceedance probability) :type: :class:`.GrowthCurve...
python
def _growth_curve_pooling_group(self, distr='glo', as_rural=False): """ Return flood growth curve function based on `amax_records` from a pooling group. :return: Inverse cumulative distribution function with one parameter `aep` (annual exceedance probability) :type: :class:`.GrowthCurve...
[ "def", "_growth_curve_pooling_group", "(", "self", ",", "distr", "=", "'glo'", ",", "as_rural", "=", "False", ")", ":", "if", "not", "self", ".", "donor_catchments", ":", "self", ".", "find_donor_catchments", "(", ")", "gc", "=", "GrowthCurve", "(", "distr",...
Return flood growth curve function based on `amax_records` from a pooling group. :return: Inverse cumulative distribution function with one parameter `aep` (annual exceedance probability) :type: :class:`.GrowthCurve` :param as_rural: assume catchment is fully rural. Default: false. :typ...
[ "Return", "flood", "growth", "curve", "function", "based", "on", "amax_records", "from", "a", "pooling", "group", "." ]
782da7c5abd1348923129efe89fb70003ebb088c
https://github.com/OpenHydrology/floodestimation/blob/782da7c5abd1348923129efe89fb70003ebb088c/floodestimation/analysis.py#L833-L849
train
Nachtfeuer/pipeline
spline/tools/version.py
VersionsCheck.process
def process(self, document): """Logging versions of required tools.""" content = json.dumps(document) versions = {} versions.update({'Spline': Version(VERSION)}) versions.update(self.get_version("Bash", self.BASH_VERSION)) if content.find('"docker(container)":') >= 0 or...
python
def process(self, document): """Logging versions of required tools.""" content = json.dumps(document) versions = {} versions.update({'Spline': Version(VERSION)}) versions.update(self.get_version("Bash", self.BASH_VERSION)) if content.find('"docker(container)":') >= 0 or...
[ "def", "process", "(", "self", ",", "document", ")", ":", "content", "=", "json", ".", "dumps", "(", "document", ")", "versions", "=", "{", "}", "versions", ".", "update", "(", "{", "'Spline'", ":", "Version", "(", "VERSION", ")", "}", ")", "versions...
Logging versions of required tools.
[ "Logging", "versions", "of", "required", "tools", "." ]
04ca18c4e95e4349532bb45b768206393e1f2c13
https://github.com/Nachtfeuer/pipeline/blob/04ca18c4e95e4349532bb45b768206393e1f2c13/spline/tools/version.py#L70-L85
train
Nachtfeuer/pipeline
spline/tools/version.py
VersionsCheck.get_version
def get_version(tool_name, tool_command): """ Get name and version of a tool defined by given command. Args: tool_name (str): name of the tool. tool_command (str): Bash one line command to get the version of the tool. Returns: dict: tool name and ver...
python
def get_version(tool_name, tool_command): """ Get name and version of a tool defined by given command. Args: tool_name (str): name of the tool. tool_command (str): Bash one line command to get the version of the tool. Returns: dict: tool name and ver...
[ "def", "get_version", "(", "tool_name", ",", "tool_command", ")", ":", "result", "=", "{", "}", "for", "line", "in", "Bash", "(", "ShellConfig", "(", "script", "=", "tool_command", ",", "internal", "=", "True", ")", ")", ".", "process", "(", ")", ":", ...
Get name and version of a tool defined by given command. Args: tool_name (str): name of the tool. tool_command (str): Bash one line command to get the version of the tool. Returns: dict: tool name and version or empty when no line has been found
[ "Get", "name", "and", "version", "of", "a", "tool", "defined", "by", "given", "command", "." ]
04ca18c4e95e4349532bb45b768206393e1f2c13
https://github.com/Nachtfeuer/pipeline/blob/04ca18c4e95e4349532bb45b768206393e1f2c13/spline/tools/version.py#L88-L108
train
Nachtfeuer/pipeline
spline/tools/version.py
VersionsReport.process
def process(self, versions): """Logging version sorted ascending by tool name.""" for tool_name in sorted(versions.keys()): version = versions[tool_name] self._log("Using tool '%s', %s" % (tool_name, version))
python
def process(self, versions): """Logging version sorted ascending by tool name.""" for tool_name in sorted(versions.keys()): version = versions[tool_name] self._log("Using tool '%s', %s" % (tool_name, version))
[ "def", "process", "(", "self", ",", "versions", ")", ":", "for", "tool_name", "in", "sorted", "(", "versions", ".", "keys", "(", ")", ")", ":", "version", "=", "versions", "[", "tool_name", "]", "self", ".", "_log", "(", "\"Using tool '%s', %s\"", "%", ...
Logging version sorted ascending by tool name.
[ "Logging", "version", "sorted", "ascending", "by", "tool", "name", "." ]
04ca18c4e95e4349532bb45b768206393e1f2c13
https://github.com/Nachtfeuer/pipeline/blob/04ca18c4e95e4349532bb45b768206393e1f2c13/spline/tools/version.py#L117-L121
train
nocarryr/python-dispatch
pydispatch/dispatch.py
Dispatcher.register_event
def register_event(self, *names): """Registers new events after instance creation Args: *names (str): Name or names of the events to register """ for name in names: if name in self.__events: continue self.__events[name] = Event(name)
python
def register_event(self, *names): """Registers new events after instance creation Args: *names (str): Name or names of the events to register """ for name in names: if name in self.__events: continue self.__events[name] = Event(name)
[ "def", "register_event", "(", "self", ",", "*", "names", ")", ":", "for", "name", "in", "names", ":", "if", "name", "in", "self", ".", "__events", ":", "continue", "self", ".", "__events", "[", "name", "]", "=", "Event", "(", "name", ")" ]
Registers new events after instance creation Args: *names (str): Name or names of the events to register
[ "Registers", "new", "events", "after", "instance", "creation" ]
7c5ca03835c922cbfdfd62772c9e560062c954c7
https://github.com/nocarryr/python-dispatch/blob/7c5ca03835c922cbfdfd62772c9e560062c954c7/pydispatch/dispatch.py#L125-L134
train
nocarryr/python-dispatch
pydispatch/dispatch.py
Dispatcher.emit
def emit(self, name, *args, **kwargs): """Dispatches an event to any subscribed listeners Note: If a listener returns :obj:`False`, the event will stop dispatching to other listeners. Any other return value is ignored. Args: name (str): The name of the :clas...
python
def emit(self, name, *args, **kwargs): """Dispatches an event to any subscribed listeners Note: If a listener returns :obj:`False`, the event will stop dispatching to other listeners. Any other return value is ignored. Args: name (str): The name of the :clas...
[ "def", "emit", "(", "self", ",", "name", ",", "*", "args", ",", "**", "kwargs", ")", ":", "e", "=", "self", ".", "__property_events", ".", "get", "(", "name", ")", "if", "e", "is", "None", ":", "e", "=", "self", ".", "__events", "[", "name", "]...
Dispatches an event to any subscribed listeners Note: If a listener returns :obj:`False`, the event will stop dispatching to other listeners. Any other return value is ignored. Args: name (str): The name of the :class:`Event` to dispatch *args (Optional)...
[ "Dispatches", "an", "event", "to", "any", "subscribed", "listeners" ]
7c5ca03835c922cbfdfd62772c9e560062c954c7
https://github.com/nocarryr/python-dispatch/blob/7c5ca03835c922cbfdfd62772c9e560062c954c7/pydispatch/dispatch.py#L236-L251
train
nocarryr/python-dispatch
pydispatch/dispatch.py
Dispatcher.get_dispatcher_event
def get_dispatcher_event(self, name): """Retrieves an Event object by name Args: name (str): The name of the :class:`Event` or :class:`~pydispatch.properties.Property` object to retrieve Returns: The :class:`Event` instance for the event or property defi...
python
def get_dispatcher_event(self, name): """Retrieves an Event object by name Args: name (str): The name of the :class:`Event` or :class:`~pydispatch.properties.Property` object to retrieve Returns: The :class:`Event` instance for the event or property defi...
[ "def", "get_dispatcher_event", "(", "self", ",", "name", ")", ":", "e", "=", "self", ".", "__property_events", ".", "get", "(", "name", ")", "if", "e", "is", "None", ":", "e", "=", "self", ".", "__events", "[", "name", "]", "return", "e" ]
Retrieves an Event object by name Args: name (str): The name of the :class:`Event` or :class:`~pydispatch.properties.Property` object to retrieve Returns: The :class:`Event` instance for the event or property definition .. versionadded:: 0.1.0
[ "Retrieves", "an", "Event", "object", "by", "name" ]
7c5ca03835c922cbfdfd62772c9e560062c954c7
https://github.com/nocarryr/python-dispatch/blob/7c5ca03835c922cbfdfd62772c9e560062c954c7/pydispatch/dispatch.py#L252-L267
train
nocarryr/python-dispatch
pydispatch/dispatch.py
Dispatcher.emission_lock
def emission_lock(self, name): """Holds emission of events and dispatches the last event on release The context manager returned will store the last event data called by :meth:`emit` and prevent callbacks until it exits. On exit, it will dispatch the last event captured (if any):: ...
python
def emission_lock(self, name): """Holds emission of events and dispatches the last event on release The context manager returned will store the last event data called by :meth:`emit` and prevent callbacks until it exits. On exit, it will dispatch the last event captured (if any):: ...
[ "def", "emission_lock", "(", "self", ",", "name", ")", ":", "e", "=", "self", ".", "__property_events", ".", "get", "(", "name", ")", "if", "e", "is", "None", ":", "e", "=", "self", ".", "__events", "[", "name", "]", "return", "e", ".", "emission_l...
Holds emission of events and dispatches the last event on release The context manager returned will store the last event data called by :meth:`emit` and prevent callbacks until it exits. On exit, it will dispatch the last event captured (if any):: class Foo(Dispatcher): ...
[ "Holds", "emission", "of", "events", "and", "dispatches", "the", "last", "event", "on", "release" ]
7c5ca03835c922cbfdfd62772c9e560062c954c7
https://github.com/nocarryr/python-dispatch/blob/7c5ca03835c922cbfdfd62772c9e560062c954c7/pydispatch/dispatch.py#L268-L309
train
acutesoftware/AIKIF
aikif/toolbox/image_detection_tools.py
TEST
def TEST(fname): """ Test function to step through all functions in order to try and identify all features on a map This test function should be placed in a main section later """ #fname = os.path.join(os.getcwd(), '..','..', # os.path.join(os.path.getcwd(), ' m = MapObject(fname, os.p...
python
def TEST(fname): """ Test function to step through all functions in order to try and identify all features on a map This test function should be placed in a main section later """ #fname = os.path.join(os.getcwd(), '..','..', # os.path.join(os.path.getcwd(), ' m = MapObject(fname, os.p...
[ "def", "TEST", "(", "fname", ")", ":", "m", "=", "MapObject", "(", "fname", ",", "os", ".", "path", ".", "join", "(", "os", ".", "getcwd", "(", ")", ",", "'img_prog_results'", ")", ")", "m", ".", "add_layer", "(", "ImagePathFollow", "(", "'border'", ...
Test function to step through all functions in order to try and identify all features on a map This test function should be placed in a main section later
[ "Test", "function", "to", "step", "through", "all", "functions", "in", "order", "to", "try", "and", "identify", "all", "features", "on", "a", "map", "This", "test", "function", "should", "be", "placed", "in", "a", "main", "section", "later" ]
fcf1582dc5f884b9a4fa7c6e20e9de9d94d21d03
https://github.com/acutesoftware/AIKIF/blob/fcf1582dc5f884b9a4fa7c6e20e9de9d94d21d03/aikif/toolbox/image_detection_tools.py#L41-L61
train
acutesoftware/AIKIF
aikif/dataTools/cls_datatable.py
DataTable.describe_contents
def describe_contents(self): """ describes various contents of data table """ print('======================================================================') print(self) print('Table = ', str(len(self.header)) + ' cols x ' + str(len(self.arr)) + ' rows') print('HEADER = ', self...
python
def describe_contents(self): """ describes various contents of data table """ print('======================================================================') print(self) print('Table = ', str(len(self.header)) + ' cols x ' + str(len(self.arr)) + ' rows') print('HEADER = ', self...
[ "def", "describe_contents", "(", "self", ")", ":", "print", "(", "'======================================================================'", ")", "print", "(", "self", ")", "print", "(", "'Table = '", ",", "str", "(", "len", "(", "self", ".", "header", ")", ")", ...
describes various contents of data table
[ "describes", "various", "contents", "of", "data", "table" ]
fcf1582dc5f884b9a4fa7c6e20e9de9d94d21d03
https://github.com/acutesoftware/AIKIF/blob/fcf1582dc5f884b9a4fa7c6e20e9de9d94d21d03/aikif/dataTools/cls_datatable.py#L68-L74
train
acutesoftware/AIKIF
aikif/dataTools/cls_datatable.py
DataTable.get_distinct_values_from_cols
def get_distinct_values_from_cols(self, l_col_list): """ returns the list of distinct combinations in a dataset based on the columns in the list. Note that this is currently implemented as MAX permutations of the combo so it is not guarenteed to have values in each case. ...
python
def get_distinct_values_from_cols(self, l_col_list): """ returns the list of distinct combinations in a dataset based on the columns in the list. Note that this is currently implemented as MAX permutations of the combo so it is not guarenteed to have values in each case. ...
[ "def", "get_distinct_values_from_cols", "(", "self", ",", "l_col_list", ")", ":", "uniq_vals", "=", "[", "]", "for", "l_col_name", "in", "l_col_list", ":", "uniq_vals", ".", "append", "(", "set", "(", "self", ".", "get_col_data_by_name", "(", "l_col_name", ")"...
returns the list of distinct combinations in a dataset based on the columns in the list. Note that this is currently implemented as MAX permutations of the combo so it is not guarenteed to have values in each case.
[ "returns", "the", "list", "of", "distinct", "combinations", "in", "a", "dataset", "based", "on", "the", "columns", "in", "the", "list", ".", "Note", "that", "this", "is", "currently", "implemented", "as", "MAX", "permutations", "of", "the", "combo", "so", ...
fcf1582dc5f884b9a4fa7c6e20e9de9d94d21d03
https://github.com/acutesoftware/AIKIF/blob/fcf1582dc5f884b9a4fa7c6e20e9de9d94d21d03/aikif/dataTools/cls_datatable.py#L79-L104
train
acutesoftware/AIKIF
aikif/dataTools/cls_datatable.py
DataTable.select_where
def select_where(self, where_col_list, where_value_list, col_name=''): """ selects rows from the array where col_list == val_list """ res = [] # list of rows to be returned col_ids = [] # ids of the columns to check #print('select_where : arr = ', len(self.ar...
python
def select_where(self, where_col_list, where_value_list, col_name=''): """ selects rows from the array where col_list == val_list """ res = [] # list of rows to be returned col_ids = [] # ids of the columns to check #print('select_where : arr = ', len(self.ar...
[ "def", "select_where", "(", "self", ",", "where_col_list", ",", "where_value_list", ",", "col_name", "=", "''", ")", ":", "res", "=", "[", "]", "col_ids", "=", "[", "]", "for", "col_id", ",", "col", "in", "enumerate", "(", "self", ".", "header", ")", ...
selects rows from the array where col_list == val_list
[ "selects", "rows", "from", "the", "array", "where", "col_list", "==", "val_list" ]
fcf1582dc5f884b9a4fa7c6e20e9de9d94d21d03
https://github.com/acutesoftware/AIKIF/blob/fcf1582dc5f884b9a4fa7c6e20e9de9d94d21d03/aikif/dataTools/cls_datatable.py#L117-L145
train
acutesoftware/AIKIF
aikif/dataTools/cls_datatable.py
DataTable.update_where
def update_where(self, col, value, where_col_list, where_value_list): """ updates the array to set cell = value where col_list == val_list """ if type(col) is str: col_ndx = self.get_col_by_name(col) else: col_ndx = col #print('col_ndx = ', col_nd...
python
def update_where(self, col, value, where_col_list, where_value_list): """ updates the array to set cell = value where col_list == val_list """ if type(col) is str: col_ndx = self.get_col_by_name(col) else: col_ndx = col #print('col_ndx = ', col_nd...
[ "def", "update_where", "(", "self", ",", "col", ",", "value", ",", "where_col_list", ",", "where_value_list", ")", ":", "if", "type", "(", "col", ")", "is", "str", ":", "col_ndx", "=", "self", ".", "get_col_by_name", "(", "col", ")", "else", ":", "col_...
updates the array to set cell = value where col_list == val_list
[ "updates", "the", "array", "to", "set", "cell", "=", "value", "where", "col_list", "==", "val_list" ]
fcf1582dc5f884b9a4fa7c6e20e9de9d94d21d03
https://github.com/acutesoftware/AIKIF/blob/fcf1582dc5f884b9a4fa7c6e20e9de9d94d21d03/aikif/dataTools/cls_datatable.py#L171-L184
train
acutesoftware/AIKIF
aikif/dataTools/cls_datatable.py
DataTable.percentile
def percentile(self, lst_data, percent , key=lambda x:x): """ calculates the 'num' percentile of the items in the list """ new_list = sorted(lst_data) #print('new list = ' , new_list) #n = float(len(lst_data)) k = (len(new_list)-1) * percent f = math.floor(k) c = ...
python
def percentile(self, lst_data, percent , key=lambda x:x): """ calculates the 'num' percentile of the items in the list """ new_list = sorted(lst_data) #print('new list = ' , new_list) #n = float(len(lst_data)) k = (len(new_list)-1) * percent f = math.floor(k) c = ...
[ "def", "percentile", "(", "self", ",", "lst_data", ",", "percent", ",", "key", "=", "lambda", "x", ":", "x", ")", ":", "new_list", "=", "sorted", "(", "lst_data", ")", "k", "=", "(", "len", "(", "new_list", ")", "-", "1", ")", "*", "percent", "f"...
calculates the 'num' percentile of the items in the list
[ "calculates", "the", "num", "percentile", "of", "the", "items", "in", "the", "list" ]
fcf1582dc5f884b9a4fa7c6e20e9de9d94d21d03
https://github.com/acutesoftware/AIKIF/blob/fcf1582dc5f884b9a4fa7c6e20e9de9d94d21d03/aikif/dataTools/cls_datatable.py#L206-L219
train
acutesoftware/AIKIF
aikif/dataTools/cls_datatable.py
DataTable.save
def save(self, filename, content): """ default is to save a file from list of lines """ with open(filename, "w") as f: if hasattr(content, '__iter__'): f.write('\n'.join([row for row in content])) else: print('WRINGI CONTWETESWREWR'...
python
def save(self, filename, content): """ default is to save a file from list of lines """ with open(filename, "w") as f: if hasattr(content, '__iter__'): f.write('\n'.join([row for row in content])) else: print('WRINGI CONTWETESWREWR'...
[ "def", "save", "(", "self", ",", "filename", ",", "content", ")", ":", "with", "open", "(", "filename", ",", "\"w\"", ")", "as", "f", ":", "if", "hasattr", "(", "content", ",", "'__iter__'", ")", ":", "f", ".", "write", "(", "'\\n'", ".", "join", ...
default is to save a file from list of lines
[ "default", "is", "to", "save", "a", "file", "from", "list", "of", "lines" ]
fcf1582dc5f884b9a4fa7c6e20e9de9d94d21d03
https://github.com/acutesoftware/AIKIF/blob/fcf1582dc5f884b9a4fa7c6e20e9de9d94d21d03/aikif/dataTools/cls_datatable.py#L230-L239
train
acutesoftware/AIKIF
aikif/dataTools/cls_datatable.py
DataTable.save_csv
def save_csv(self, filename, write_header_separately=True): """ save the default array as a CSV file """ txt = '' #print("SAVING arr = ", self.arr) with open(filename, "w") as f: if write_header_separately: f.write(','.join([c for c in...
python
def save_csv(self, filename, write_header_separately=True): """ save the default array as a CSV file """ txt = '' #print("SAVING arr = ", self.arr) with open(filename, "w") as f: if write_header_separately: f.write(','.join([c for c in...
[ "def", "save_csv", "(", "self", ",", "filename", ",", "write_header_separately", "=", "True", ")", ":", "txt", "=", "''", "with", "open", "(", "filename", ",", "\"w\"", ")", "as", "f", ":", "if", "write_header_separately", ":", "f", ".", "write", "(", ...
save the default array as a CSV file
[ "save", "the", "default", "array", "as", "a", "CSV", "file" ]
fcf1582dc5f884b9a4fa7c6e20e9de9d94d21d03
https://github.com/acutesoftware/AIKIF/blob/fcf1582dc5f884b9a4fa7c6e20e9de9d94d21d03/aikif/dataTools/cls_datatable.py#L241-L256
train
acutesoftware/AIKIF
aikif/dataTools/cls_datatable.py
DataTable.drop
def drop(self, fname): """ drop the table, view or delete the file """ if self.dataset_type == 'file': import os try: os.remove(fname) except Exception as ex: print('cant drop file "' + fname + '" : ' + str(ex))
python
def drop(self, fname): """ drop the table, view or delete the file """ if self.dataset_type == 'file': import os try: os.remove(fname) except Exception as ex: print('cant drop file "' + fname + '" : ' + str(ex))
[ "def", "drop", "(", "self", ",", "fname", ")", ":", "if", "self", ".", "dataset_type", "==", "'file'", ":", "import", "os", "try", ":", "os", ".", "remove", "(", "fname", ")", "except", "Exception", "as", "ex", ":", "print", "(", "'cant drop file \"'",...
drop the table, view or delete the file
[ "drop", "the", "table", "view", "or", "delete", "the", "file" ]
fcf1582dc5f884b9a4fa7c6e20e9de9d94d21d03
https://github.com/acutesoftware/AIKIF/blob/fcf1582dc5f884b9a4fa7c6e20e9de9d94d21d03/aikif/dataTools/cls_datatable.py#L258-L267
train
acutesoftware/AIKIF
aikif/dataTools/cls_datatable.py
DataTable.get_col_data_by_name
def get_col_data_by_name(self, col_name, WHERE_Clause=''): """ returns the values of col_name according to where """ #print('get_col_data_by_name: col_name = ', col_name, ' WHERE = ', WHERE_Clause) col_key = self.get_col_by_name(col_name) if col_key is None: print('get_col_da...
python
def get_col_data_by_name(self, col_name, WHERE_Clause=''): """ returns the values of col_name according to where """ #print('get_col_data_by_name: col_name = ', col_name, ' WHERE = ', WHERE_Clause) col_key = self.get_col_by_name(col_name) if col_key is None: print('get_col_da...
[ "def", "get_col_data_by_name", "(", "self", ",", "col_name", ",", "WHERE_Clause", "=", "''", ")", ":", "col_key", "=", "self", ".", "get_col_by_name", "(", "col_name", ")", "if", "col_key", "is", "None", ":", "print", "(", "'get_col_data_by_name: col_name = '", ...
returns the values of col_name according to where
[ "returns", "the", "values", "of", "col_name", "according", "to", "where" ]
fcf1582dc5f884b9a4fa7c6e20e9de9d94d21d03
https://github.com/acutesoftware/AIKIF/blob/fcf1582dc5f884b9a4fa7c6e20e9de9d94d21d03/aikif/dataTools/cls_datatable.py#L311-L323
train
acutesoftware/AIKIF
aikif/dataTools/cls_datatable.py
DataTable.format_rst
def format_rst(self): """ return table in RST format """ res = '' num_cols = len(self.header) col_width = 25 for _ in range(num_cols): res += ''.join(['=' for _ in range(col_width - 1)]) + ' ' res += '\n' for c in self.header: ...
python
def format_rst(self): """ return table in RST format """ res = '' num_cols = len(self.header) col_width = 25 for _ in range(num_cols): res += ''.join(['=' for _ in range(col_width - 1)]) + ' ' res += '\n' for c in self.header: ...
[ "def", "format_rst", "(", "self", ")", ":", "res", "=", "''", "num_cols", "=", "len", "(", "self", ".", "header", ")", "col_width", "=", "25", "for", "_", "in", "range", "(", "num_cols", ")", ":", "res", "+=", "''", ".", "join", "(", "[", "'='", ...
return table in RST format
[ "return", "table", "in", "RST", "format" ]
fcf1582dc5f884b9a4fa7c6e20e9de9d94d21d03
https://github.com/acutesoftware/AIKIF/blob/fcf1582dc5f884b9a4fa7c6e20e9de9d94d21d03/aikif/dataTools/cls_datatable.py#L325-L348
train
mpg-age-bioinformatics/AGEpy
AGEpy/homology.py
getHomoloGene
def getHomoloGene(taxfile="build_inputs/taxid_taxname",\ genefile="homologene.data",\ proteinsfile="build_inputs/all_proteins.data",\ proteinsclusterfile="build_inputs/proteins_for_clustering.data",\ baseURL="http://ftp.ncbi.nih.gov/pub/HomoloGene/...
python
def getHomoloGene(taxfile="build_inputs/taxid_taxname",\ genefile="homologene.data",\ proteinsfile="build_inputs/all_proteins.data",\ proteinsclusterfile="build_inputs/proteins_for_clustering.data",\ baseURL="http://ftp.ncbi.nih.gov/pub/HomoloGene/...
[ "def", "getHomoloGene", "(", "taxfile", "=", "\"build_inputs/taxid_taxname\"", ",", "genefile", "=", "\"homologene.data\"", ",", "proteinsfile", "=", "\"build_inputs/all_proteins.data\"", ",", "proteinsclusterfile", "=", "\"build_inputs/proteins_for_clustering.data\"", ",", "ba...
Returns NBCI's Homolog Gene tables. :param taxfile: path to local file or to baseURL/taxfile :param genefile: path to local file or to baseURL/genefile :param proteinsfile: path to local file or to baseURL/proteinsfile :param proteinsclusterfile: path to local file or to baseURL/proteinsclusterfile ...
[ "Returns", "NBCI", "s", "Homolog", "Gene", "tables", "." ]
887808a7a2c1504f39ce8d8cb36c15c1721cd29f
https://github.com/mpg-age-bioinformatics/AGEpy/blob/887808a7a2c1504f39ce8d8cb36c15c1721cd29f/AGEpy/homology.py#L6-L66
train
mpg-age-bioinformatics/AGEpy
AGEpy/fasta.py
getFasta
def getFasta(opened_file, sequence_name): """ Retrieves a sequence from an opened multifasta file :param opened_file: an opened multifasta file eg. opened_file=open("/path/to/file.fa",'r+') :param sequence_name: the name of the sequence to be retrieved eg. for '>2 dna:chromosome chromosome:GRCm38:2:1:1...
python
def getFasta(opened_file, sequence_name): """ Retrieves a sequence from an opened multifasta file :param opened_file: an opened multifasta file eg. opened_file=open("/path/to/file.fa",'r+') :param sequence_name: the name of the sequence to be retrieved eg. for '>2 dna:chromosome chromosome:GRCm38:2:1:1...
[ "def", "getFasta", "(", "opened_file", ",", "sequence_name", ")", ":", "lines", "=", "opened_file", ".", "readlines", "(", ")", "seq", "=", "str", "(", "\"\"", ")", "for", "i", "in", "range", "(", "0", ",", "len", "(", "lines", ")", ")", ":", "line...
Retrieves a sequence from an opened multifasta file :param opened_file: an opened multifasta file eg. opened_file=open("/path/to/file.fa",'r+') :param sequence_name: the name of the sequence to be retrieved eg. for '>2 dna:chromosome chromosome:GRCm38:2:1:182113224:1 REF' use: sequence_name=str(2) returns...
[ "Retrieves", "a", "sequence", "from", "an", "opened", "multifasta", "file" ]
887808a7a2c1504f39ce8d8cb36c15c1721cd29f
https://github.com/mpg-age-bioinformatics/AGEpy/blob/887808a7a2c1504f39ce8d8cb36c15c1721cd29f/AGEpy/fasta.py#L2-L34
train
mpg-age-bioinformatics/AGEpy
AGEpy/fasta.py
writeFasta
def writeFasta(sequence, sequence_name, output_file): """ Writes a fasta sequence into a file. :param sequence: a string with the sequence to be written :param sequence_name: name of the the fasta sequence :param output_file: /path/to/file.fa to be written :returns: nothing """ i=0 ...
python
def writeFasta(sequence, sequence_name, output_file): """ Writes a fasta sequence into a file. :param sequence: a string with the sequence to be written :param sequence_name: name of the the fasta sequence :param output_file: /path/to/file.fa to be written :returns: nothing """ i=0 ...
[ "def", "writeFasta", "(", "sequence", ",", "sequence_name", ",", "output_file", ")", ":", "i", "=", "0", "f", "=", "open", "(", "output_file", ",", "'w'", ")", "f", ".", "write", "(", "\">\"", "+", "str", "(", "sequence_name", ")", "+", "\"\\n\"", ")...
Writes a fasta sequence into a file. :param sequence: a string with the sequence to be written :param sequence_name: name of the the fasta sequence :param output_file: /path/to/file.fa to be written :returns: nothing
[ "Writes", "a", "fasta", "sequence", "into", "a", "file", "." ]
887808a7a2c1504f39ce8d8cb36c15c1721cd29f
https://github.com/mpg-age-bioinformatics/AGEpy/blob/887808a7a2c1504f39ce8d8cb36c15c1721cd29f/AGEpy/fasta.py#L36-L52
train
mpg-age-bioinformatics/AGEpy
AGEpy/fasta.py
rewriteFasta
def rewriteFasta(sequence, sequence_name, fasta_in, fasta_out): """ Rewrites a specific sequence in a multifasta file while keeping the sequence header. :param sequence: a string with the sequence to be written :param sequence_name: the name of the sequence to be retrieved eg. for '>2 dna:chromosome ch...
python
def rewriteFasta(sequence, sequence_name, fasta_in, fasta_out): """ Rewrites a specific sequence in a multifasta file while keeping the sequence header. :param sequence: a string with the sequence to be written :param sequence_name: the name of the sequence to be retrieved eg. for '>2 dna:chromosome ch...
[ "def", "rewriteFasta", "(", "sequence", ",", "sequence_name", ",", "fasta_in", ",", "fasta_out", ")", ":", "f", "=", "open", "(", "fasta_in", ",", "'r+'", ")", "f2", "=", "open", "(", "fasta_out", ",", "'w'", ")", "lines", "=", "f", ".", "readlines", ...
Rewrites a specific sequence in a multifasta file while keeping the sequence header. :param sequence: a string with the sequence to be written :param sequence_name: the name of the sequence to be retrieved eg. for '>2 dna:chromosome chromosome:GRCm38:2:1:182113224:1 REF' use: sequence_name=str(2) :param fa...
[ "Rewrites", "a", "specific", "sequence", "in", "a", "multifasta", "file", "while", "keeping", "the", "sequence", "header", "." ]
887808a7a2c1504f39ce8d8cb36c15c1721cd29f
https://github.com/mpg-age-bioinformatics/AGEpy/blob/887808a7a2c1504f39ce8d8cb36c15c1721cd29f/AGEpy/fasta.py#L54-L92
train
acutesoftware/AIKIF
aikif/toolbox/Toolbox.py
Toolbox._get_tool_str
def _get_tool_str(self, tool): """ get a string representation of the tool """ res = tool['file'] try: res += '.' + tool['function'] except Exception as ex: print('Warning - no function defined for tool ' + str(tool)) res += '\n' r...
python
def _get_tool_str(self, tool): """ get a string representation of the tool """ res = tool['file'] try: res += '.' + tool['function'] except Exception as ex: print('Warning - no function defined for tool ' + str(tool)) res += '\n' r...
[ "def", "_get_tool_str", "(", "self", ",", "tool", ")", ":", "res", "=", "tool", "[", "'file'", "]", "try", ":", "res", "+=", "'.'", "+", "tool", "[", "'function'", "]", "except", "Exception", "as", "ex", ":", "print", "(", "'Warning - no function defined...
get a string representation of the tool
[ "get", "a", "string", "representation", "of", "the", "tool" ]
fcf1582dc5f884b9a4fa7c6e20e9de9d94d21d03
https://github.com/acutesoftware/AIKIF/blob/fcf1582dc5f884b9a4fa7c6e20e9de9d94d21d03/aikif/toolbox/Toolbox.py#L42-L52
train
acutesoftware/AIKIF
aikif/toolbox/Toolbox.py
Toolbox.get_tool_by_name
def get_tool_by_name(self, nme): """ get the tool object by name or file """ for t in self.lstTools: if 'name' in t: if t['name'] == nme: return t if 'file' in t: if t['file'] == nme: return t...
python
def get_tool_by_name(self, nme): """ get the tool object by name or file """ for t in self.lstTools: if 'name' in t: if t['name'] == nme: return t if 'file' in t: if t['file'] == nme: return t...
[ "def", "get_tool_by_name", "(", "self", ",", "nme", ")", ":", "for", "t", "in", "self", ".", "lstTools", ":", "if", "'name'", "in", "t", ":", "if", "t", "[", "'name'", "]", "==", "nme", ":", "return", "t", "if", "'file'", "in", "t", ":", "if", ...
get the tool object by name or file
[ "get", "the", "tool", "object", "by", "name", "or", "file" ]
fcf1582dc5f884b9a4fa7c6e20e9de9d94d21d03
https://github.com/acutesoftware/AIKIF/blob/fcf1582dc5f884b9a4fa7c6e20e9de9d94d21d03/aikif/toolbox/Toolbox.py#L54-L65
train
acutesoftware/AIKIF
aikif/toolbox/Toolbox.py
Toolbox.save
def save(self, fname=''): """ Save the list of tools to AIKIF core and optionally to local file fname """ if fname != '': with open(fname, 'w') as f: for t in self.lstTools: self.verify(t) f.write(self.tool_as_string(t))
python
def save(self, fname=''): """ Save the list of tools to AIKIF core and optionally to local file fname """ if fname != '': with open(fname, 'w') as f: for t in self.lstTools: self.verify(t) f.write(self.tool_as_string(t))
[ "def", "save", "(", "self", ",", "fname", "=", "''", ")", ":", "if", "fname", "!=", "''", ":", "with", "open", "(", "fname", ",", "'w'", ")", "as", "f", ":", "for", "t", "in", "self", ".", "lstTools", ":", "self", ".", "verify", "(", "t", ")"...
Save the list of tools to AIKIF core and optionally to local file fname
[ "Save", "the", "list", "of", "tools", "to", "AIKIF", "core", "and", "optionally", "to", "local", "file", "fname" ]
fcf1582dc5f884b9a4fa7c6e20e9de9d94d21d03
https://github.com/acutesoftware/AIKIF/blob/fcf1582dc5f884b9a4fa7c6e20e9de9d94d21d03/aikif/toolbox/Toolbox.py#L88-L96
train
acutesoftware/AIKIF
aikif/toolbox/Toolbox.py
Toolbox.verify
def verify(self, tool): """ check that the tool exists """ if os.path.isfile(tool['file']): print('Toolbox: program exists = TOK :: ' + tool['file']) return True else: print('Toolbox: program exists = FAIL :: ' + tool['file']) retu...
python
def verify(self, tool): """ check that the tool exists """ if os.path.isfile(tool['file']): print('Toolbox: program exists = TOK :: ' + tool['file']) return True else: print('Toolbox: program exists = FAIL :: ' + tool['file']) retu...
[ "def", "verify", "(", "self", ",", "tool", ")", ":", "if", "os", ".", "path", ".", "isfile", "(", "tool", "[", "'file'", "]", ")", ":", "print", "(", "'Toolbox: program exists = TOK :: '", "+", "tool", "[", "'file'", "]", ")", "return", "True", "else"...
check that the tool exists
[ "check", "that", "the", "tool", "exists" ]
fcf1582dc5f884b9a4fa7c6e20e9de9d94d21d03
https://github.com/acutesoftware/AIKIF/blob/fcf1582dc5f884b9a4fa7c6e20e9de9d94d21d03/aikif/toolbox/Toolbox.py#L98-L107
train
acutesoftware/AIKIF
aikif/toolbox/Toolbox.py
Toolbox.run
def run(self, tool, args, new_import_path=''): """ import the tool and call the function, passing the args. """ if new_import_path != '': #print('APPENDING PATH = ', new_import_path) sys.path.append(new_import_path) #if silent == 'N': prin...
python
def run(self, tool, args, new_import_path=''): """ import the tool and call the function, passing the args. """ if new_import_path != '': #print('APPENDING PATH = ', new_import_path) sys.path.append(new_import_path) #if silent == 'N': prin...
[ "def", "run", "(", "self", ",", "tool", ",", "args", ",", "new_import_path", "=", "''", ")", ":", "if", "new_import_path", "!=", "''", ":", "sys", ".", "path", ".", "append", "(", "new_import_path", ")", "print", "(", "'main called '", "+", "tool", "["...
import the tool and call the function, passing the args.
[ "import", "the", "tool", "and", "call", "the", "function", "passing", "the", "args", "." ]
fcf1582dc5f884b9a4fa7c6e20e9de9d94d21d03
https://github.com/acutesoftware/AIKIF/blob/fcf1582dc5f884b9a4fa7c6e20e9de9d94d21d03/aikif/toolbox/Toolbox.py#L109-L123
train
Nachtfeuer/pipeline
spline/application.py
main
def main(**kwargs): """The Pipeline tool.""" options = ApplicationOptions(**kwargs) Event.configure(is_logging_enabled=options.event_logging) application = Application(options) application.run(options.definition)
python
def main(**kwargs): """The Pipeline tool.""" options = ApplicationOptions(**kwargs) Event.configure(is_logging_enabled=options.event_logging) application = Application(options) application.run(options.definition)
[ "def", "main", "(", "**", "kwargs", ")", ":", "options", "=", "ApplicationOptions", "(", "**", "kwargs", ")", "Event", ".", "configure", "(", "is_logging_enabled", "=", "options", ".", "event_logging", ")", "application", "=", "Application", "(", "options", ...
The Pipeline tool.
[ "The", "Pipeline", "tool", "." ]
04ca18c4e95e4349532bb45b768206393e1f2c13
https://github.com/Nachtfeuer/pipeline/blob/04ca18c4e95e4349532bb45b768206393e1f2c13/spline/application.py#L209-L214
train
Nachtfeuer/pipeline
spline/application.py
Application.setup_logging
def setup_logging(self): """Setup of application logging.""" is_custom_logging = len(self.options.logging_config) > 0 is_custom_logging = is_custom_logging and os.path.isfile(self.options.logging_config) is_custom_logging = is_custom_logging and not self.options.dry_run if is_cu...
python
def setup_logging(self): """Setup of application logging.""" is_custom_logging = len(self.options.logging_config) > 0 is_custom_logging = is_custom_logging and os.path.isfile(self.options.logging_config) is_custom_logging = is_custom_logging and not self.options.dry_run if is_cu...
[ "def", "setup_logging", "(", "self", ")", ":", "is_custom_logging", "=", "len", "(", "self", ".", "options", ".", "logging_config", ")", ">", "0", "is_custom_logging", "=", "is_custom_logging", "and", "os", ".", "path", ".", "isfile", "(", "self", ".", "op...
Setup of application logging.
[ "Setup", "of", "application", "logging", "." ]
04ca18c4e95e4349532bb45b768206393e1f2c13
https://github.com/Nachtfeuer/pipeline/blob/04ca18c4e95e4349532bb45b768206393e1f2c13/spline/application.py#L56-L68
train
Nachtfeuer/pipeline
spline/application.py
Application.validate_document
def validate_document(self, definition): """ Validate given pipeline document. The method is trying to load, parse and validate the spline document. The validator verifies the Python structure B{not} the file format. Args: definition (str): path and filename of a ya...
python
def validate_document(self, definition): """ Validate given pipeline document. The method is trying to load, parse and validate the spline document. The validator verifies the Python structure B{not} the file format. Args: definition (str): path and filename of a ya...
[ "def", "validate_document", "(", "self", ",", "definition", ")", ":", "initial_document", "=", "{", "}", "try", ":", "initial_document", "=", "Loader", ".", "load", "(", "definition", ")", "except", "RuntimeError", "as", "exception", ":", "self", ".", "logge...
Validate given pipeline document. The method is trying to load, parse and validate the spline document. The validator verifies the Python structure B{not} the file format. Args: definition (str): path and filename of a yaml file containing a valid spline definition. Return...
[ "Validate", "given", "pipeline", "document", "." ]
04ca18c4e95e4349532bb45b768206393e1f2c13
https://github.com/Nachtfeuer/pipeline/blob/04ca18c4e95e4349532bb45b768206393e1f2c13/spline/application.py#L70-L101
train
Nachtfeuer/pipeline
spline/application.py
Application.run_matrix
def run_matrix(self, matrix_definition, document): """ Running pipeline via a matrix. Args: matrix_definition (dict): one concrete matrix item. document (dict): spline document (complete) as loaded from yaml file. """ matrix = Matrix(matrix_definition, 'm...
python
def run_matrix(self, matrix_definition, document): """ Running pipeline via a matrix. Args: matrix_definition (dict): one concrete matrix item. document (dict): spline document (complete) as loaded from yaml file. """ matrix = Matrix(matrix_definition, 'm...
[ "def", "run_matrix", "(", "self", ",", "matrix_definition", ",", "document", ")", ":", "matrix", "=", "Matrix", "(", "matrix_definition", ",", "'matrix(parallel)'", "in", "document", ")", "process_data", "=", "MatrixProcessData", "(", ")", "process_data", ".", "...
Running pipeline via a matrix. Args: matrix_definition (dict): one concrete matrix item. document (dict): spline document (complete) as loaded from yaml file.
[ "Running", "pipeline", "via", "a", "matrix", "." ]
04ca18c4e95e4349532bb45b768206393e1f2c13
https://github.com/Nachtfeuer/pipeline/blob/04ca18c4e95e4349532bb45b768206393e1f2c13/spline/application.py#L103-L119
train
Nachtfeuer/pipeline
spline/application.py
Application.shutdown
def shutdown(self, collector, success): """Shutdown of the application.""" self.event.delegate(success) if collector is not None: collector.queue.put(None) collector.join() if not success: sys.exit(1)
python
def shutdown(self, collector, success): """Shutdown of the application.""" self.event.delegate(success) if collector is not None: collector.queue.put(None) collector.join() if not success: sys.exit(1)
[ "def", "shutdown", "(", "self", ",", "collector", ",", "success", ")", ":", "self", ".", "event", ".", "delegate", "(", "success", ")", "if", "collector", "is", "not", "None", ":", "collector", ".", "queue", ".", "put", "(", "None", ")", "collector", ...
Shutdown of the application.
[ "Shutdown", "of", "the", "application", "." ]
04ca18c4e95e4349532bb45b768206393e1f2c13
https://github.com/Nachtfeuer/pipeline/blob/04ca18c4e95e4349532bb45b768206393e1f2c13/spline/application.py#L121-L128
train
Nachtfeuer/pipeline
spline/application.py
Application.provide_temporary_scripts_path
def provide_temporary_scripts_path(self): """When configured trying to ensure that path does exist.""" if len(self.options.temporary_scripts_path) > 0: if os.path.isfile(self.options.temporary_scripts_path): self.logger.error("Error: configured script path seems to be a file!...
python
def provide_temporary_scripts_path(self): """When configured trying to ensure that path does exist.""" if len(self.options.temporary_scripts_path) > 0: if os.path.isfile(self.options.temporary_scripts_path): self.logger.error("Error: configured script path seems to be a file!...
[ "def", "provide_temporary_scripts_path", "(", "self", ")", ":", "if", "len", "(", "self", ".", "options", ".", "temporary_scripts_path", ")", ">", "0", ":", "if", "os", ".", "path", ".", "isfile", "(", "self", ".", "options", ".", "temporary_scripts_path", ...
When configured trying to ensure that path does exist.
[ "When", "configured", "trying", "to", "ensure", "that", "path", "does", "exist", "." ]
04ca18c4e95e4349532bb45b768206393e1f2c13
https://github.com/Nachtfeuer/pipeline/blob/04ca18c4e95e4349532bb45b768206393e1f2c13/spline/application.py#L162-L171
train
Nachtfeuer/pipeline
spline/application.py
Application.create_and_run_collector
def create_and_run_collector(document, options): """Create and run collector process for report data.""" collector = None if not options.report == 'off': collector = Collector() collector.store.configure(document) Event.configure(collector_queue=collector.queu...
python
def create_and_run_collector(document, options): """Create and run collector process for report data.""" collector = None if not options.report == 'off': collector = Collector() collector.store.configure(document) Event.configure(collector_queue=collector.queu...
[ "def", "create_and_run_collector", "(", "document", ",", "options", ")", ":", "collector", "=", "None", "if", "not", "options", ".", "report", "==", "'off'", ":", "collector", "=", "Collector", "(", ")", "collector", ".", "store", ".", "configure", "(", "d...
Create and run collector process for report data.
[ "Create", "and", "run", "collector", "process", "for", "report", "data", "." ]
04ca18c4e95e4349532bb45b768206393e1f2c13
https://github.com/Nachtfeuer/pipeline/blob/04ca18c4e95e4349532bb45b768206393e1f2c13/spline/application.py#L174-L182
train
Nachtfeuer/pipeline
spline/tools/filters.py
docker_environment
def docker_environment(env): """ Transform dictionary of environment variables into Docker -e parameters. >>> result = docker_environment({'param1': 'val1', 'param2': 'val2'}) >>> result in ['-e "param1=val1" -e "param2=val2"', '-e "param2=val2" -e "param1=val1"'] True """ return ' '.join( ...
python
def docker_environment(env): """ Transform dictionary of environment variables into Docker -e parameters. >>> result = docker_environment({'param1': 'val1', 'param2': 'val2'}) >>> result in ['-e "param1=val1" -e "param2=val2"', '-e "param2=val2" -e "param1=val1"'] True """ return ' '.join( ...
[ "def", "docker_environment", "(", "env", ")", ":", "return", "' '", ".", "join", "(", "[", "\"-e \\\"%s=%s\\\"\"", "%", "(", "key", ",", "value", ".", "replace", "(", "\"$\"", ",", "\"\\\\$\"", ")", ".", "replace", "(", "\"\\\"\"", ",", "\"\\\\\\\"\"", "...
Transform dictionary of environment variables into Docker -e parameters. >>> result = docker_environment({'param1': 'val1', 'param2': 'val2'}) >>> result in ['-e "param1=val1" -e "param2=val2"', '-e "param2=val2" -e "param1=val1"'] True
[ "Transform", "dictionary", "of", "environment", "variables", "into", "Docker", "-", "e", "parameters", "." ]
04ca18c4e95e4349532bb45b768206393e1f2c13
https://github.com/Nachtfeuer/pipeline/blob/04ca18c4e95e4349532bb45b768206393e1f2c13/spline/tools/filters.py#L60-L70
train
OpenHydrology/floodestimation
floodestimation/fehdata.py
_retrieve_download_url
def _retrieve_download_url(): """ Retrieves download location for FEH data zip file from hosted json configuration file. :return: URL for FEH data file :rtype: str """ try: # Try to obtain the url from the Open Hydrology json config file. with urlopen(config['nrfa']['oh_json_url...
python
def _retrieve_download_url(): """ Retrieves download location for FEH data zip file from hosted json configuration file. :return: URL for FEH data file :rtype: str """ try: # Try to obtain the url from the Open Hydrology json config file. with urlopen(config['nrfa']['oh_json_url...
[ "def", "_retrieve_download_url", "(", ")", ":", "try", ":", "with", "urlopen", "(", "config", "[", "'nrfa'", "]", "[", "'oh_json_url'", "]", ",", "timeout", "=", "10", ")", "as", "f", ":", "remote_config", "=", "json", ".", "loads", "(", "f", ".", "r...
Retrieves download location for FEH data zip file from hosted json configuration file. :return: URL for FEH data file :rtype: str
[ "Retrieves", "download", "location", "for", "FEH", "data", "zip", "file", "from", "hosted", "json", "configuration", "file", "." ]
782da7c5abd1348923129efe89fb70003ebb088c
https://github.com/OpenHydrology/floodestimation/blob/782da7c5abd1348923129efe89fb70003ebb088c/floodestimation/fehdata.py#L58-L79
train
OpenHydrology/floodestimation
floodestimation/fehdata.py
update_available
def update_available(after_days=1): """ Check whether updated NRFA data is available. :param after_days: Only check if not checked previously since a certain number of days ago :type after_days: float :return: `True` if update available, `False` if not, `None` if remote location cannot be reached. ...
python
def update_available(after_days=1): """ Check whether updated NRFA data is available. :param after_days: Only check if not checked previously since a certain number of days ago :type after_days: float :return: `True` if update available, `False` if not, `None` if remote location cannot be reached. ...
[ "def", "update_available", "(", "after_days", "=", "1", ")", ":", "never_downloaded", "=", "not", "bool", "(", "config", ".", "get", "(", "'nrfa'", ",", "'downloaded_on'", ",", "fallback", "=", "None", ")", "or", "None", ")", "if", "never_downloaded", ":",...
Check whether updated NRFA data is available. :param after_days: Only check if not checked previously since a certain number of days ago :type after_days: float :return: `True` if update available, `False` if not, `None` if remote location cannot be reached. :rtype: bool or None
[ "Check", "whether", "updated", "NRFA", "data", "is", "available", "." ]
782da7c5abd1348923129efe89fb70003ebb088c
https://github.com/OpenHydrology/floodestimation/blob/782da7c5abd1348923129efe89fb70003ebb088c/floodestimation/fehdata.py#L82-L109
train
OpenHydrology/floodestimation
floodestimation/fehdata.py
download_data
def download_data(): """ Downloads complete station dataset including catchment descriptors and amax records. And saves it into a cache folder. """ with urlopen(_retrieve_download_url()) as f: with open(os.path.join(CACHE_FOLDER, CACHE_ZIP), "wb") as local_file: local_file.write(...
python
def download_data(): """ Downloads complete station dataset including catchment descriptors and amax records. And saves it into a cache folder. """ with urlopen(_retrieve_download_url()) as f: with open(os.path.join(CACHE_FOLDER, CACHE_ZIP), "wb") as local_file: local_file.write(...
[ "def", "download_data", "(", ")", ":", "with", "urlopen", "(", "_retrieve_download_url", "(", ")", ")", "as", "f", ":", "with", "open", "(", "os", ".", "path", ".", "join", "(", "CACHE_FOLDER", ",", "CACHE_ZIP", ")", ",", "\"wb\"", ")", "as", "local_fi...
Downloads complete station dataset including catchment descriptors and amax records. And saves it into a cache folder.
[ "Downloads", "complete", "station", "dataset", "including", "catchment", "descriptors", "and", "amax", "records", ".", "And", "saves", "it", "into", "a", "cache", "folder", "." ]
782da7c5abd1348923129efe89fb70003ebb088c
https://github.com/OpenHydrology/floodestimation/blob/782da7c5abd1348923129efe89fb70003ebb088c/floodestimation/fehdata.py#L112-L119
train
OpenHydrology/floodestimation
floodestimation/fehdata.py
_update_nrfa_metadata
def _update_nrfa_metadata(remote_config): """ Save NRFA metadata to local config file using retrieved config data :param remote_config: Downloaded JSON data, not a ConfigParser object! """ config['nrfa']['oh_json_url'] = remote_config['nrfa_oh_json_url'] config['nrfa']['version'] = remote_confi...
python
def _update_nrfa_metadata(remote_config): """ Save NRFA metadata to local config file using retrieved config data :param remote_config: Downloaded JSON data, not a ConfigParser object! """ config['nrfa']['oh_json_url'] = remote_config['nrfa_oh_json_url'] config['nrfa']['version'] = remote_confi...
[ "def", "_update_nrfa_metadata", "(", "remote_config", ")", ":", "config", "[", "'nrfa'", "]", "[", "'oh_json_url'", "]", "=", "remote_config", "[", "'nrfa_oh_json_url'", "]", "config", "[", "'nrfa'", "]", "[", "'version'", "]", "=", "remote_config", "[", "'nrf...
Save NRFA metadata to local config file using retrieved config data :param remote_config: Downloaded JSON data, not a ConfigParser object!
[ "Save", "NRFA", "metadata", "to", "local", "config", "file", "using", "retrieved", "config", "data" ]
782da7c5abd1348923129efe89fb70003ebb088c
https://github.com/OpenHydrology/floodestimation/blob/782da7c5abd1348923129efe89fb70003ebb088c/floodestimation/fehdata.py#L122-L134
train
OpenHydrology/floodestimation
floodestimation/fehdata.py
nrfa_metadata
def nrfa_metadata(): """ Return metadata on the NRFA data. Returned metadata is a dict with the following elements: - `url`: string with NRFA data download URL - `version`: string with NRFA version number, e.g. '3.3.4' - `published_on`: datetime of data release/publication (only month and year...
python
def nrfa_metadata(): """ Return metadata on the NRFA data. Returned metadata is a dict with the following elements: - `url`: string with NRFA data download URL - `version`: string with NRFA version number, e.g. '3.3.4' - `published_on`: datetime of data release/publication (only month and year...
[ "def", "nrfa_metadata", "(", ")", ":", "result", "=", "{", "'url'", ":", "config", ".", "get", "(", "'nrfa'", ",", "'url'", ",", "fallback", "=", "None", ")", "or", "None", ",", "'version'", ":", "config", ".", "get", "(", "'nrfa'", ",", "'version'",...
Return metadata on the NRFA data. Returned metadata is a dict with the following elements: - `url`: string with NRFA data download URL - `version`: string with NRFA version number, e.g. '3.3.4' - `published_on`: datetime of data release/publication (only month and year are accurate, rest should be ign...
[ "Return", "metadata", "on", "the", "NRFA", "data", "." ]
782da7c5abd1348923129efe89fb70003ebb088c
https://github.com/OpenHydrology/floodestimation/blob/782da7c5abd1348923129efe89fb70003ebb088c/floodestimation/fehdata.py#L137-L157
train
OpenHydrology/floodestimation
floodestimation/fehdata.py
unzip_data
def unzip_data(): """ Extract all files from downloaded FEH data zip file. """ with ZipFile(os.path.join(CACHE_FOLDER, CACHE_ZIP), 'r') as zf: zf.extractall(path=CACHE_FOLDER)
python
def unzip_data(): """ Extract all files from downloaded FEH data zip file. """ with ZipFile(os.path.join(CACHE_FOLDER, CACHE_ZIP), 'r') as zf: zf.extractall(path=CACHE_FOLDER)
[ "def", "unzip_data", "(", ")", ":", "with", "ZipFile", "(", "os", ".", "path", ".", "join", "(", "CACHE_FOLDER", ",", "CACHE_ZIP", ")", ",", "'r'", ")", "as", "zf", ":", "zf", ".", "extractall", "(", "path", "=", "CACHE_FOLDER", ")" ]
Extract all files from downloaded FEH data zip file.
[ "Extract", "all", "files", "from", "downloaded", "FEH", "data", "zip", "file", "." ]
782da7c5abd1348923129efe89fb70003ebb088c
https://github.com/OpenHydrology/floodestimation/blob/782da7c5abd1348923129efe89fb70003ebb088c/floodestimation/fehdata.py#L160-L165
train
acutesoftware/AIKIF
aikif/toolbox/xml_tools.py
get_xml_stats
def get_xml_stats(fname): """ return a dictionary of statistics about an XML file including size in bytes, num lines, number of elements, count by elements """ f = mod_file.TextFile(fname) res = {} res['shortname'] = f.name res['folder'] = f.path res['filesize'] = str(f.size) + ...
python
def get_xml_stats(fname): """ return a dictionary of statistics about an XML file including size in bytes, num lines, number of elements, count by elements """ f = mod_file.TextFile(fname) res = {} res['shortname'] = f.name res['folder'] = f.path res['filesize'] = str(f.size) + ...
[ "def", "get_xml_stats", "(", "fname", ")", ":", "f", "=", "mod_file", ".", "TextFile", "(", "fname", ")", "res", "=", "{", "}", "res", "[", "'shortname'", "]", "=", "f", ".", "name", "res", "[", "'folder'", "]", "=", "f", ".", "path", "res", "[",...
return a dictionary of statistics about an XML file including size in bytes, num lines, number of elements, count by elements
[ "return", "a", "dictionary", "of", "statistics", "about", "an", "XML", "file", "including", "size", "in", "bytes", "num", "lines", "number", "of", "elements", "count", "by", "elements" ]
fcf1582dc5f884b9a4fa7c6e20e9de9d94d21d03
https://github.com/acutesoftware/AIKIF/blob/fcf1582dc5f884b9a4fa7c6e20e9de9d94d21d03/aikif/toolbox/xml_tools.py#L18-L32
train
acutesoftware/AIKIF
aikif/toolbox/xml_tools.py
make_random_xml_file
def make_random_xml_file(fname, num_elements=200, depth=3): """ makes a random xml file mainly for testing the xml_split """ with open(fname, 'w') as f: f.write('<?xml version="1.0" ?>\n<random>\n') for dep_num, _ in enumerate(range(1,depth)): f.write(' <depth>\n <content>\n...
python
def make_random_xml_file(fname, num_elements=200, depth=3): """ makes a random xml file mainly for testing the xml_split """ with open(fname, 'w') as f: f.write('<?xml version="1.0" ?>\n<random>\n') for dep_num, _ in enumerate(range(1,depth)): f.write(' <depth>\n <content>\n...
[ "def", "make_random_xml_file", "(", "fname", ",", "num_elements", "=", "200", ",", "depth", "=", "3", ")", ":", "with", "open", "(", "fname", ",", "'w'", ")", "as", "f", ":", "f", ".", "write", "(", "'<?xml version=\"1.0\" ?>\\n<random>\\n'", ")", "for", ...
makes a random xml file mainly for testing the xml_split
[ "makes", "a", "random", "xml", "file", "mainly", "for", "testing", "the", "xml_split" ]
fcf1582dc5f884b9a4fa7c6e20e9de9d94d21d03
https://github.com/acutesoftware/AIKIF/blob/fcf1582dc5f884b9a4fa7c6e20e9de9d94d21d03/aikif/toolbox/xml_tools.py#L34-L48
train
mpg-age-bioinformatics/AGEpy
AGEpy/kegg.py
organismsKEGG
def organismsKEGG(): """ Lists all organisms present in the KEGG database. :returns: a dataframe containing one organism per row. """ organisms=urlopen("http://rest.kegg.jp/list/organism").read() organisms=organisms.split("\n") #for o in organisms: # print o # sys.stdout.flus...
python
def organismsKEGG(): """ Lists all organisms present in the KEGG database. :returns: a dataframe containing one organism per row. """ organisms=urlopen("http://rest.kegg.jp/list/organism").read() organisms=organisms.split("\n") #for o in organisms: # print o # sys.stdout.flus...
[ "def", "organismsKEGG", "(", ")", ":", "organisms", "=", "urlopen", "(", "\"http://rest.kegg.jp/list/organism\"", ")", ".", "read", "(", ")", "organisms", "=", "organisms", ".", "split", "(", "\"\\n\"", ")", "organisms", "=", "[", "s", ".", "split", "(", "...
Lists all organisms present in the KEGG database. :returns: a dataframe containing one organism per row.
[ "Lists", "all", "organisms", "present", "in", "the", "KEGG", "database", "." ]
887808a7a2c1504f39ce8d8cb36c15c1721cd29f
https://github.com/mpg-age-bioinformatics/AGEpy/blob/887808a7a2c1504f39ce8d8cb36c15c1721cd29f/AGEpy/kegg.py#L19-L33
train
mpg-age-bioinformatics/AGEpy
AGEpy/kegg.py
databasesKEGG
def databasesKEGG(organism,ens_ids): """ Finds KEGG database identifiers for a respective organism given example ensembl ids. :param organism: an organism as listed in organismsKEGG() :param ens_ids: a list of ensenbl ids of the respective organism :returns: nothing if no database was found, or a...
python
def databasesKEGG(organism,ens_ids): """ Finds KEGG database identifiers for a respective organism given example ensembl ids. :param organism: an organism as listed in organismsKEGG() :param ens_ids: a list of ensenbl ids of the respective organism :returns: nothing if no database was found, or a...
[ "def", "databasesKEGG", "(", "organism", ",", "ens_ids", ")", ":", "all_genes", "=", "urlopen", "(", "\"http://rest.kegg.jp/list/\"", "+", "organism", ")", ".", "read", "(", ")", "all_genes", "=", "all_genes", ".", "split", "(", "\"\\n\"", ")", "dbs", "=", ...
Finds KEGG database identifiers for a respective organism given example ensembl ids. :param organism: an organism as listed in organismsKEGG() :param ens_ids: a list of ensenbl ids of the respective organism :returns: nothing if no database was found, or a string if a database was found
[ "Finds", "KEGG", "database", "identifiers", "for", "a", "respective", "organism", "given", "example", "ensembl", "ids", "." ]
887808a7a2c1504f39ce8d8cb36c15c1721cd29f
https://github.com/mpg-age-bioinformatics/AGEpy/blob/887808a7a2c1504f39ce8d8cb36c15c1721cd29f/AGEpy/kegg.py#L36-L80
train
mpg-age-bioinformatics/AGEpy
AGEpy/kegg.py
ensembl_to_kegg
def ensembl_to_kegg(organism,kegg_db): """ Looks up KEGG mappings of KEGG ids to ensembl ids :param organism: an organisms as listed in organismsKEGG() :param kegg_db: a matching KEGG db as reported in databasesKEGG :returns: a Pandas dataframe of with 'KEGGid' and 'ENSid'. """ print("KEG...
python
def ensembl_to_kegg(organism,kegg_db): """ Looks up KEGG mappings of KEGG ids to ensembl ids :param organism: an organisms as listed in organismsKEGG() :param kegg_db: a matching KEGG db as reported in databasesKEGG :returns: a Pandas dataframe of with 'KEGGid' and 'ENSid'. """ print("KEG...
[ "def", "ensembl_to_kegg", "(", "organism", ",", "kegg_db", ")", ":", "print", "(", "\"KEGG API: http://rest.genome.jp/link/\"", "+", "kegg_db", "+", "\"/\"", "+", "organism", ")", "sys", ".", "stdout", ".", "flush", "(", ")", "kegg_ens", "=", "urlopen", "(", ...
Looks up KEGG mappings of KEGG ids to ensembl ids :param organism: an organisms as listed in organismsKEGG() :param kegg_db: a matching KEGG db as reported in databasesKEGG :returns: a Pandas dataframe of with 'KEGGid' and 'ENSid'.
[ "Looks", "up", "KEGG", "mappings", "of", "KEGG", "ids", "to", "ensembl", "ids" ]
887808a7a2c1504f39ce8d8cb36c15c1721cd29f
https://github.com/mpg-age-bioinformatics/AGEpy/blob/887808a7a2c1504f39ce8d8cb36c15c1721cd29f/AGEpy/kegg.py#L83-L105
train
mpg-age-bioinformatics/AGEpy
AGEpy/kegg.py
ecs_idsKEGG
def ecs_idsKEGG(organism): """ Uses KEGG to retrieve all ids and respective ecs for a given KEGG organism :param organism: an organisms as listed in organismsKEGG() :returns: a Pandas dataframe of with 'ec' and 'KEGGid'. """ kegg_ec=urlopen("http://rest.kegg.jp/link/"+organism+"/enzyme").read...
python
def ecs_idsKEGG(organism): """ Uses KEGG to retrieve all ids and respective ecs for a given KEGG organism :param organism: an organisms as listed in organismsKEGG() :returns: a Pandas dataframe of with 'ec' and 'KEGGid'. """ kegg_ec=urlopen("http://rest.kegg.jp/link/"+organism+"/enzyme").read...
[ "def", "ecs_idsKEGG", "(", "organism", ")", ":", "kegg_ec", "=", "urlopen", "(", "\"http://rest.kegg.jp/link/\"", "+", "organism", "+", "\"/enzyme\"", ")", ".", "read", "(", ")", "kegg_ec", "=", "kegg_ec", ".", "split", "(", "\"\\n\"", ")", "final", "=", "...
Uses KEGG to retrieve all ids and respective ecs for a given KEGG organism :param organism: an organisms as listed in organismsKEGG() :returns: a Pandas dataframe of with 'ec' and 'KEGGid'.
[ "Uses", "KEGG", "to", "retrieve", "all", "ids", "and", "respective", "ecs", "for", "a", "given", "KEGG", "organism" ]
887808a7a2c1504f39ce8d8cb36c15c1721cd29f
https://github.com/mpg-age-bioinformatics/AGEpy/blob/887808a7a2c1504f39ce8d8cb36c15c1721cd29f/AGEpy/kegg.py#L107-L123
train
mpg-age-bioinformatics/AGEpy
AGEpy/kegg.py
idsKEGG
def idsKEGG(organism): """ Uses KEGG to retrieve all ids for a given KEGG organism :param organism: an organism as listed in organismsKEGG() :returns: a Pandas dataframe of with 'gene_name' and 'KEGGid'. """ ORG=urlopen("http://rest.kegg.jp/list/"+organism).read() ORG=ORG.split("\n") ...
python
def idsKEGG(organism): """ Uses KEGG to retrieve all ids for a given KEGG organism :param organism: an organism as listed in organismsKEGG() :returns: a Pandas dataframe of with 'gene_name' and 'KEGGid'. """ ORG=urlopen("http://rest.kegg.jp/list/"+organism).read() ORG=ORG.split("\n") ...
[ "def", "idsKEGG", "(", "organism", ")", ":", "ORG", "=", "urlopen", "(", "\"http://rest.kegg.jp/list/\"", "+", "organism", ")", ".", "read", "(", ")", "ORG", "=", "ORG", ".", "split", "(", "\"\\n\"", ")", "final", "=", "[", "]", "for", "k", "in", "OR...
Uses KEGG to retrieve all ids for a given KEGG organism :param organism: an organism as listed in organismsKEGG() :returns: a Pandas dataframe of with 'gene_name' and 'KEGGid'.
[ "Uses", "KEGG", "to", "retrieve", "all", "ids", "for", "a", "given", "KEGG", "organism" ]
887808a7a2c1504f39ce8d8cb36c15c1721cd29f
https://github.com/mpg-age-bioinformatics/AGEpy/blob/887808a7a2c1504f39ce8d8cb36c15c1721cd29f/AGEpy/kegg.py#L126-L147
train
mpg-age-bioinformatics/AGEpy
AGEpy/kegg.py
biomaRtTOkegg
def biomaRtTOkegg(df): """ Transforms a pandas dataframe with the columns 'ensembl_gene_id','kegg_enzyme' to dataframe ready for use in ... :param df: a pandas dataframe with the following columns: 'ensembl_gene_id','kegg_enzyme' :returns: a pandas dataframe with the following columns: 'ensembl_ge...
python
def biomaRtTOkegg(df): """ Transforms a pandas dataframe with the columns 'ensembl_gene_id','kegg_enzyme' to dataframe ready for use in ... :param df: a pandas dataframe with the following columns: 'ensembl_gene_id','kegg_enzyme' :returns: a pandas dataframe with the following columns: 'ensembl_ge...
[ "def", "biomaRtTOkegg", "(", "df", ")", ":", "df", "=", "df", ".", "dropna", "(", ")", "ECcols", "=", "df", ".", "columns", ".", "tolist", "(", ")", "df", ".", "reset_index", "(", "inplace", "=", "True", ",", "drop", "=", "True", ")", "field", "=...
Transforms a pandas dataframe with the columns 'ensembl_gene_id','kegg_enzyme' to dataframe ready for use in ... :param df: a pandas dataframe with the following columns: 'ensembl_gene_id','kegg_enzyme' :returns: a pandas dataframe with the following columns: 'ensembl_gene_id','kegg_enzyme'
[ "Transforms", "a", "pandas", "dataframe", "with", "the", "columns", "ensembl_gene_id", "kegg_enzyme", "to", "dataframe", "ready", "for", "use", "in", "..." ]
887808a7a2c1504f39ce8d8cb36c15c1721cd29f
https://github.com/mpg-age-bioinformatics/AGEpy/blob/887808a7a2c1504f39ce8d8cb36c15c1721cd29f/AGEpy/kegg.py#L194-L232
train
mpg-age-bioinformatics/AGEpy
AGEpy/kegg.py
expKEGG
def expKEGG(organism,names_KEGGids): """ Gets all KEGG pathways for an organism :param organism: an organism as listed in organismsKEGG() :param names_KEGGids: a Pandas dataframe with the columns 'gene_name': and 'KEGGid' as reported from idsKEGG(organism) (or a subset of it). :returns df: a Pand...
python
def expKEGG(organism,names_KEGGids): """ Gets all KEGG pathways for an organism :param organism: an organism as listed in organismsKEGG() :param names_KEGGids: a Pandas dataframe with the columns 'gene_name': and 'KEGGid' as reported from idsKEGG(organism) (or a subset of it). :returns df: a Pand...
[ "def", "expKEGG", "(", "organism", ",", "names_KEGGids", ")", ":", "kegg_paths", "=", "urlopen", "(", "\"http://rest.kegg.jp/list/pathway/\"", "+", "organism", ")", ".", "read", "(", ")", "kegg_paths", "=", "kegg_paths", ".", "split", "(", "\"\\n\"", ")", "fin...
Gets all KEGG pathways for an organism :param organism: an organism as listed in organismsKEGG() :param names_KEGGids: a Pandas dataframe with the columns 'gene_name': and 'KEGGid' as reported from idsKEGG(organism) (or a subset of it). :returns df: a Pandas dataframe with 'KEGGid','pathID(1):pathNAME(1)...
[ "Gets", "all", "KEGG", "pathways", "for", "an", "organism" ]
887808a7a2c1504f39ce8d8cb36c15c1721cd29f
https://github.com/mpg-age-bioinformatics/AGEpy/blob/887808a7a2c1504f39ce8d8cb36c15c1721cd29f/AGEpy/kegg.py#L236-L287
train
mpg-age-bioinformatics/AGEpy
AGEpy/rbiom.py
RdatabasesBM
def RdatabasesBM(host=rbiomart_host): """ Lists BioMart databases through a RPY2 connection. :param host: address of the host server, default='www.ensembl.org' :returns: nothing """ biomaRt = importr("biomaRt") print(biomaRt.listMarts(host=host))
python
def RdatabasesBM(host=rbiomart_host): """ Lists BioMart databases through a RPY2 connection. :param host: address of the host server, default='www.ensembl.org' :returns: nothing """ biomaRt = importr("biomaRt") print(biomaRt.listMarts(host=host))
[ "def", "RdatabasesBM", "(", "host", "=", "rbiomart_host", ")", ":", "biomaRt", "=", "importr", "(", "\"biomaRt\"", ")", "print", "(", "biomaRt", ".", "listMarts", "(", "host", "=", "host", ")", ")" ]
Lists BioMart databases through a RPY2 connection. :param host: address of the host server, default='www.ensembl.org' :returns: nothing
[ "Lists", "BioMart", "databases", "through", "a", "RPY2", "connection", "." ]
887808a7a2c1504f39ce8d8cb36c15c1721cd29f
https://github.com/mpg-age-bioinformatics/AGEpy/blob/887808a7a2c1504f39ce8d8cb36c15c1721cd29f/AGEpy/rbiom.py#L16-L26
train
mpg-age-bioinformatics/AGEpy
AGEpy/rbiom.py
RdatasetsBM
def RdatasetsBM(database,host=rbiomart_host): """ Lists BioMart datasets through a RPY2 connection. :param database: a database listed in RdatabasesBM() :param host: address of the host server, default='www.ensembl.org' :returns: nothing """ biomaRt = importr("biomaRt") ensemblMart=bi...
python
def RdatasetsBM(database,host=rbiomart_host): """ Lists BioMart datasets through a RPY2 connection. :param database: a database listed in RdatabasesBM() :param host: address of the host server, default='www.ensembl.org' :returns: nothing """ biomaRt = importr("biomaRt") ensemblMart=bi...
[ "def", "RdatasetsBM", "(", "database", ",", "host", "=", "rbiomart_host", ")", ":", "biomaRt", "=", "importr", "(", "\"biomaRt\"", ")", "ensemblMart", "=", "biomaRt", ".", "useMart", "(", "database", ",", "host", "=", "host", ")", "print", "(", "biomaRt", ...
Lists BioMart datasets through a RPY2 connection. :param database: a database listed in RdatabasesBM() :param host: address of the host server, default='www.ensembl.org' :returns: nothing
[ "Lists", "BioMart", "datasets", "through", "a", "RPY2", "connection", "." ]
887808a7a2c1504f39ce8d8cb36c15c1721cd29f
https://github.com/mpg-age-bioinformatics/AGEpy/blob/887808a7a2c1504f39ce8d8cb36c15c1721cd29f/AGEpy/rbiom.py#L28-L40
train
mpg-age-bioinformatics/AGEpy
AGEpy/rbiom.py
RfiltersBM
def RfiltersBM(dataset,database,host=rbiomart_host): """ Lists BioMart filters through a RPY2 connection. :param dataset: a dataset listed in RdatasetsBM() :param database: a database listed in RdatabasesBM() :param host: address of the host server, default='www.ensembl.org' :returns: nothing ...
python
def RfiltersBM(dataset,database,host=rbiomart_host): """ Lists BioMart filters through a RPY2 connection. :param dataset: a dataset listed in RdatasetsBM() :param database: a database listed in RdatabasesBM() :param host: address of the host server, default='www.ensembl.org' :returns: nothing ...
[ "def", "RfiltersBM", "(", "dataset", ",", "database", ",", "host", "=", "rbiomart_host", ")", ":", "biomaRt", "=", "importr", "(", "\"biomaRt\"", ")", "ensemblMart", "=", "biomaRt", ".", "useMart", "(", "database", ",", "host", "=", "host", ")", "ensembl",...
Lists BioMart filters through a RPY2 connection. :param dataset: a dataset listed in RdatasetsBM() :param database: a database listed in RdatabasesBM() :param host: address of the host server, default='www.ensembl.org' :returns: nothing
[ "Lists", "BioMart", "filters", "through", "a", "RPY2", "connection", "." ]
887808a7a2c1504f39ce8d8cb36c15c1721cd29f
https://github.com/mpg-age-bioinformatics/AGEpy/blob/887808a7a2c1504f39ce8d8cb36c15c1721cd29f/AGEpy/rbiom.py#L42-L56
train
mpg-age-bioinformatics/AGEpy
AGEpy/rbiom.py
RattributesBM
def RattributesBM(dataset,database,host=rbiomart_host): """ Lists BioMart attributes through a RPY2 connection. :param dataset: a dataset listed in RdatasetsBM() :param database: a database listed in RdatabasesBM() :param host: address of the host server, default='www.ensembl.org' :returns: no...
python
def RattributesBM(dataset,database,host=rbiomart_host): """ Lists BioMart attributes through a RPY2 connection. :param dataset: a dataset listed in RdatasetsBM() :param database: a database listed in RdatabasesBM() :param host: address of the host server, default='www.ensembl.org' :returns: no...
[ "def", "RattributesBM", "(", "dataset", ",", "database", ",", "host", "=", "rbiomart_host", ")", ":", "biomaRt", "=", "importr", "(", "\"biomaRt\"", ")", "ensemblMart", "=", "biomaRt", ".", "useMart", "(", "database", ",", "host", "=", "rbiomart_host", ")", ...
Lists BioMart attributes through a RPY2 connection. :param dataset: a dataset listed in RdatasetsBM() :param database: a database listed in RdatabasesBM() :param host: address of the host server, default='www.ensembl.org' :returns: nothing
[ "Lists", "BioMart", "attributes", "through", "a", "RPY2", "connection", "." ]
887808a7a2c1504f39ce8d8cb36c15c1721cd29f
https://github.com/mpg-age-bioinformatics/AGEpy/blob/887808a7a2c1504f39ce8d8cb36c15c1721cd29f/AGEpy/rbiom.py#L58-L72
train
acutesoftware/AIKIF
scripts/examples/document_AIKIF.py
get_list_of_applications
def get_list_of_applications(): """ Get list of applications """ apps = mod_prg.Programs('Applications', 'C:\\apps') fl = mod_fl.FileList(['C:\\apps'], ['*.exe'], ["\\bk\\"]) for f in fl.get_list(): apps.add(f, 'autogenerated list') apps.list() apps.save()
python
def get_list_of_applications(): """ Get list of applications """ apps = mod_prg.Programs('Applications', 'C:\\apps') fl = mod_fl.FileList(['C:\\apps'], ['*.exe'], ["\\bk\\"]) for f in fl.get_list(): apps.add(f, 'autogenerated list') apps.list() apps.save()
[ "def", "get_list_of_applications", "(", ")", ":", "apps", "=", "mod_prg", ".", "Programs", "(", "'Applications'", ",", "'C:\\\\apps'", ")", "fl", "=", "mod_fl", ".", "FileList", "(", "[", "'C:\\\\apps'", "]", ",", "[", "'*.exe'", "]", ",", "[", "\"\\\\bk\\...
Get list of applications
[ "Get", "list", "of", "applications" ]
fcf1582dc5f884b9a4fa7c6e20e9de9d94d21d03
https://github.com/acutesoftware/AIKIF/blob/fcf1582dc5f884b9a4fa7c6e20e9de9d94d21d03/scripts/examples/document_AIKIF.py#L147-L156
train
Timusan/wtforms-dynamic-fields
wtforms_dynamic_fields/wtforms_dynamic_fields.py
WTFormsDynamicFields.add_field
def add_field(self, name, label, field_type, *args, **kwargs): """ Add the field to the internal configuration dictionary. """ if name in self._dyn_fields: raise AttributeError('Field already added to the form.') else: self._dyn_fields[name] = {'label': label, 'type': fie...
python
def add_field(self, name, label, field_type, *args, **kwargs): """ Add the field to the internal configuration dictionary. """ if name in self._dyn_fields: raise AttributeError('Field already added to the form.') else: self._dyn_fields[name] = {'label': label, 'type': fie...
[ "def", "add_field", "(", "self", ",", "name", ",", "label", ",", "field_type", ",", "*", "args", ",", "**", "kwargs", ")", ":", "if", "name", "in", "self", ".", "_dyn_fields", ":", "raise", "AttributeError", "(", "'Field already added to the form.'", ")", ...
Add the field to the internal configuration dictionary.
[ "Add", "the", "field", "to", "the", "internal", "configuration", "dictionary", "." ]
d984a646075219a6f8a0e931c96035ca3e44be56
https://github.com/Timusan/wtforms-dynamic-fields/blob/d984a646075219a6f8a0e931c96035ca3e44be56/wtforms_dynamic_fields/wtforms_dynamic_fields.py#L44-L50
train
Timusan/wtforms-dynamic-fields
wtforms_dynamic_fields/wtforms_dynamic_fields.py
WTFormsDynamicFields.add_validator
def add_validator(self, name, validator, *args, **kwargs): """ Add the validator to the internal configuration dictionary. :param name: The field machine name to apply the validator on :param validator: The WTForms validator object The rest are optional arguments...
python
def add_validator(self, name, validator, *args, **kwargs): """ Add the validator to the internal configuration dictionary. :param name: The field machine name to apply the validator on :param validator: The WTForms validator object The rest are optional arguments...
[ "def", "add_validator", "(", "self", ",", "name", ",", "validator", ",", "*", "args", ",", "**", "kwargs", ")", ":", "if", "name", "in", "self", ".", "_dyn_fields", ":", "if", "'validators'", "in", "self", ".", "_dyn_fields", "[", "name", "]", ":", "...
Add the validator to the internal configuration dictionary. :param name: The field machine name to apply the validator on :param validator: The WTForms validator object The rest are optional arguments and keyword arguments that belong to the validator. We let the...
[ "Add", "the", "validator", "to", "the", "internal", "configuration", "dictionary", "." ]
d984a646075219a6f8a0e931c96035ca3e44be56
https://github.com/Timusan/wtforms-dynamic-fields/blob/d984a646075219a6f8a0e931c96035ca3e44be56/wtforms_dynamic_fields/wtforms_dynamic_fields.py#L52-L76
train
Timusan/wtforms-dynamic-fields
wtforms_dynamic_fields/wtforms_dynamic_fields.py
WTFormsDynamicFields.process
def process(self, form, post): """ Process the given WTForm Form object. Itterate over the POST values and check each field against the configuration that was made. For each field that is valid, check all the validator parameters for possible %field% replacement, then bind ...
python
def process(self, form, post): """ Process the given WTForm Form object. Itterate over the POST values and check each field against the configuration that was made. For each field that is valid, check all the validator parameters for possible %field% replacement, then bind ...
[ "def", "process", "(", "self", ",", "form", ",", "post", ")", ":", "if", "not", "isinstance", "(", "form", ",", "FormMeta", ")", ":", "raise", "TypeError", "(", "'Given form is not a valid WTForm.'", ")", "re_field_name", "=", "re", ".", "compile", "(", "r...
Process the given WTForm Form object. Itterate over the POST values and check each field against the configuration that was made. For each field that is valid, check all the validator parameters for possible %field% replacement, then bind these parameters to their validator. ...
[ "Process", "the", "given", "WTForm", "Form", "object", "." ]
d984a646075219a6f8a0e931c96035ca3e44be56
https://github.com/Timusan/wtforms-dynamic-fields/blob/d984a646075219a6f8a0e931c96035ca3e44be56/wtforms_dynamic_fields/wtforms_dynamic_fields.py#L90-L214
train
mpg-age-bioinformatics/AGEpy
AGEpy/bed.py
GetBEDnarrowPeakgz
def GetBEDnarrowPeakgz(URL_or_PATH_TO_file): """ Reads a gz compressed BED narrow peak file from a web address or local file :param URL_or_PATH_TO_file: web address of path to local file :returns: a Pandas dataframe """ if os.path.isfile(URL_or_PATH_TO_file): response=open(URL_or_PATH...
python
def GetBEDnarrowPeakgz(URL_or_PATH_TO_file): """ Reads a gz compressed BED narrow peak file from a web address or local file :param URL_or_PATH_TO_file: web address of path to local file :returns: a Pandas dataframe """ if os.path.isfile(URL_or_PATH_TO_file): response=open(URL_or_PATH...
[ "def", "GetBEDnarrowPeakgz", "(", "URL_or_PATH_TO_file", ")", ":", "if", "os", ".", "path", ".", "isfile", "(", "URL_or_PATH_TO_file", ")", ":", "response", "=", "open", "(", "URL_or_PATH_TO_file", ",", "\"r\"", ")", "compressedFile", "=", "StringIO", ".", "St...
Reads a gz compressed BED narrow peak file from a web address or local file :param URL_or_PATH_TO_file: web address of path to local file :returns: a Pandas dataframe
[ "Reads", "a", "gz", "compressed", "BED", "narrow", "peak", "file", "from", "a", "web", "address", "or", "local", "file" ]
887808a7a2c1504f39ce8d8cb36c15c1721cd29f
https://github.com/mpg-age-bioinformatics/AGEpy/blob/887808a7a2c1504f39ce8d8cb36c15c1721cd29f/AGEpy/bed.py#L30-L53
train
mpg-age-bioinformatics/AGEpy
AGEpy/bed.py
dfTObedtool
def dfTObedtool(df): """ Transforms a pandas dataframe into a bedtool :param df: Pandas dataframe :returns: a bedtool """ df=df.astype(str) df=df.drop_duplicates() df=df.values.tolist() df=["\t".join(s) for s in df ] df="\n".join(df) df=BedTool(df, from_string=True) re...
python
def dfTObedtool(df): """ Transforms a pandas dataframe into a bedtool :param df: Pandas dataframe :returns: a bedtool """ df=df.astype(str) df=df.drop_duplicates() df=df.values.tolist() df=["\t".join(s) for s in df ] df="\n".join(df) df=BedTool(df, from_string=True) re...
[ "def", "dfTObedtool", "(", "df", ")", ":", "df", "=", "df", ".", "astype", "(", "str", ")", "df", "=", "df", ".", "drop_duplicates", "(", ")", "df", "=", "df", ".", "values", ".", "tolist", "(", ")", "df", "=", "[", "\"\\t\"", ".", "join", "(",...
Transforms a pandas dataframe into a bedtool :param df: Pandas dataframe :returns: a bedtool
[ "Transforms", "a", "pandas", "dataframe", "into", "a", "bedtool" ]
887808a7a2c1504f39ce8d8cb36c15c1721cd29f
https://github.com/mpg-age-bioinformatics/AGEpy/blob/887808a7a2c1504f39ce8d8cb36c15c1721cd29f/AGEpy/bed.py#L55-L70
train
Nachtfeuer/pipeline
spline/tools/event.py
Event.configure
def configure(**kwargs): """Global configuration for event handling.""" for key in kwargs: if key == 'is_logging_enabled': Event.is_logging_enabled = kwargs[key] elif key == 'collector_queue': Event.collector_queue = kwargs[key] else: ...
python
def configure(**kwargs): """Global configuration for event handling.""" for key in kwargs: if key == 'is_logging_enabled': Event.is_logging_enabled = kwargs[key] elif key == 'collector_queue': Event.collector_queue = kwargs[key] else: ...
[ "def", "configure", "(", "**", "kwargs", ")", ":", "for", "key", "in", "kwargs", ":", "if", "key", "==", "'is_logging_enabled'", ":", "Event", ".", "is_logging_enabled", "=", "kwargs", "[", "key", "]", "elif", "key", "==", "'collector_queue'", ":", "Event"...
Global configuration for event handling.
[ "Global", "configuration", "for", "event", "handling", "." ]
04ca18c4e95e4349532bb45b768206393e1f2c13
https://github.com/Nachtfeuer/pipeline/blob/04ca18c4e95e4349532bb45b768206393e1f2c13/spline/tools/event.py#L46-L55
train
Nachtfeuer/pipeline
spline/tools/event.py
Event.failed
def failed(self, **kwargs): """Finish event as failed with optional additional information.""" self.finished = datetime.now() self.status = 'failed' self.information.update(kwargs) self.logger.info("Failed - took %f seconds.", self.duration()) self.update_report_collector...
python
def failed(self, **kwargs): """Finish event as failed with optional additional information.""" self.finished = datetime.now() self.status = 'failed' self.information.update(kwargs) self.logger.info("Failed - took %f seconds.", self.duration()) self.update_report_collector...
[ "def", "failed", "(", "self", ",", "**", "kwargs", ")", ":", "self", ".", "finished", "=", "datetime", ".", "now", "(", ")", "self", ".", "status", "=", "'failed'", "self", ".", "information", ".", "update", "(", "kwargs", ")", "self", ".", "logger",...
Finish event as failed with optional additional information.
[ "Finish", "event", "as", "failed", "with", "optional", "additional", "information", "." ]
04ca18c4e95e4349532bb45b768206393e1f2c13
https://github.com/Nachtfeuer/pipeline/blob/04ca18c4e95e4349532bb45b768206393e1f2c13/spline/tools/event.py#L69-L75
train
Nachtfeuer/pipeline
spline/tools/event.py
Event.update_report_collector
def update_report_collector(self, timestamp): """Updating report collector for pipeline details.""" report_enabled = 'report' in self.information and self.information['report'] == 'html' report_enabled = report_enabled and 'stage' in self.information report_enabled = report_enabled and E...
python
def update_report_collector(self, timestamp): """Updating report collector for pipeline details.""" report_enabled = 'report' in self.information and self.information['report'] == 'html' report_enabled = report_enabled and 'stage' in self.information report_enabled = report_enabled and E...
[ "def", "update_report_collector", "(", "self", ",", "timestamp", ")", ":", "report_enabled", "=", "'report'", "in", "self", ".", "information", "and", "self", ".", "information", "[", "'report'", "]", "==", "'html'", "report_enabled", "=", "report_enabled", "and...
Updating report collector for pipeline details.
[ "Updating", "report", "collector", "for", "pipeline", "details", "." ]
04ca18c4e95e4349532bb45b768206393e1f2c13
https://github.com/Nachtfeuer/pipeline/blob/04ca18c4e95e4349532bb45b768206393e1f2c13/spline/tools/event.py#L89-L102
train
acutesoftware/AIKIF
aikif/toolbox/sql_tools.py
count_lines_in_file
def count_lines_in_file(src_file ): """ test function. """ tot = 0 res = '' try: with open(src_file, 'r') as f: for line in f: tot += 1 res = str(tot) + ' recs read' except: res = 'ERROR -couldnt open file' return res
python
def count_lines_in_file(src_file ): """ test function. """ tot = 0 res = '' try: with open(src_file, 'r') as f: for line in f: tot += 1 res = str(tot) + ' recs read' except: res = 'ERROR -couldnt open file' return res
[ "def", "count_lines_in_file", "(", "src_file", ")", ":", "tot", "=", "0", "res", "=", "''", "try", ":", "with", "open", "(", "src_file", ",", "'r'", ")", "as", "f", ":", "for", "line", "in", "f", ":", "tot", "+=", "1", "res", "=", "str", "(", "...
test function.
[ "test", "function", "." ]
fcf1582dc5f884b9a4fa7c6e20e9de9d94d21d03
https://github.com/acutesoftware/AIKIF/blob/fcf1582dc5f884b9a4fa7c6e20e9de9d94d21d03/aikif/toolbox/sql_tools.py#L10-L23
train
acutesoftware/AIKIF
aikif/toolbox/sql_tools.py
load_txt_to_sql
def load_txt_to_sql(tbl_name, src_file_and_path, src_file, op_folder): """ creates a SQL loader script to load a text file into a database and then executes it. Note that src_file is """ if op_folder == '': pth = '' else: pth = op_folder + os.sep fname_create_script...
python
def load_txt_to_sql(tbl_name, src_file_and_path, src_file, op_folder): """ creates a SQL loader script to load a text file into a database and then executes it. Note that src_file is """ if op_folder == '': pth = '' else: pth = op_folder + os.sep fname_create_script...
[ "def", "load_txt_to_sql", "(", "tbl_name", ",", "src_file_and_path", ",", "src_file", ",", "op_folder", ")", ":", "if", "op_folder", "==", "''", ":", "pth", "=", "''", "else", ":", "pth", "=", "op_folder", "+", "os", ".", "sep", "fname_create_script", "=",...
creates a SQL loader script to load a text file into a database and then executes it. Note that src_file is
[ "creates", "a", "SQL", "loader", "script", "to", "load", "a", "text", "file", "into", "a", "database", "and", "then", "executes", "it", ".", "Note", "that", "src_file", "is" ]
fcf1582dc5f884b9a4fa7c6e20e9de9d94d21d03
https://github.com/acutesoftware/AIKIF/blob/fcf1582dc5f884b9a4fa7c6e20e9de9d94d21d03/aikif/toolbox/sql_tools.py#L26-L44
train
asyncdef/aitertools
aitertools/__init__.py
anext
async def anext(*args): """Return the next item from an async iterator. Args: iterable: An async iterable. default: An optional default value to return if the iterable is empty. Return: The next value of the iterable. Raises: TypeError: The iterable given is not async....
python
async def anext(*args): """Return the next item from an async iterator. Args: iterable: An async iterable. default: An optional default value to return if the iterable is empty. Return: The next value of the iterable. Raises: TypeError: The iterable given is not async....
[ "async", "def", "anext", "(", "*", "args", ")", ":", "if", "not", "args", ":", "raise", "TypeError", "(", "'anext() expected at least 1 arguments, got 0'", ")", "if", "len", "(", "args", ")", ">", "2", ":", "raise", "TypeError", "(", "'anext() expected at most...
Return the next item from an async iterator. Args: iterable: An async iterable. default: An optional default value to return if the iterable is empty. Return: The next value of the iterable. Raises: TypeError: The iterable given is not async. This function will return...
[ "Return", "the", "next", "item", "from", "an", "async", "iterator", "." ]
26a6c7e71e87dd1ddc4acb755d70ca30894f7928
https://github.com/asyncdef/aitertools/blob/26a6c7e71e87dd1ddc4acb755d70ca30894f7928/aitertools/__init__.py#L102-L146
train
asyncdef/aitertools
aitertools/__init__.py
repeat
def repeat(obj, times=None): """Make an iterator that returns object over and over again.""" if times is None: return AsyncIterWrapper(sync_itertools.repeat(obj)) return AsyncIterWrapper(sync_itertools.repeat(obj, times))
python
def repeat(obj, times=None): """Make an iterator that returns object over and over again.""" if times is None: return AsyncIterWrapper(sync_itertools.repeat(obj)) return AsyncIterWrapper(sync_itertools.repeat(obj, times))
[ "def", "repeat", "(", "obj", ",", "times", "=", "None", ")", ":", "if", "times", "is", "None", ":", "return", "AsyncIterWrapper", "(", "sync_itertools", ".", "repeat", "(", "obj", ")", ")", "return", "AsyncIterWrapper", "(", "sync_itertools", ".", "repeat"...
Make an iterator that returns object over and over again.
[ "Make", "an", "iterator", "that", "returns", "object", "over", "and", "over", "again", "." ]
26a6c7e71e87dd1ddc4acb755d70ca30894f7928
https://github.com/asyncdef/aitertools/blob/26a6c7e71e87dd1ddc4acb755d70ca30894f7928/aitertools/__init__.py#L240-L246
train
asyncdef/aitertools
aitertools/__init__.py
_async_callable
def _async_callable(func): """Ensure the callable is an async def.""" if isinstance(func, types.CoroutineType): return func @functools.wraps(func) async def _async_def_wrapper(*args, **kwargs): """Wrap a a sync callable in an async def.""" return func(*args, **kwargs) retu...
python
def _async_callable(func): """Ensure the callable is an async def.""" if isinstance(func, types.CoroutineType): return func @functools.wraps(func) async def _async_def_wrapper(*args, **kwargs): """Wrap a a sync callable in an async def.""" return func(*args, **kwargs) retu...
[ "def", "_async_callable", "(", "func", ")", ":", "if", "isinstance", "(", "func", ",", "types", ".", "CoroutineType", ")", ":", "return", "func", "@", "functools", ".", "wraps", "(", "func", ")", "async", "def", "_async_def_wrapper", "(", "*", "args", ",...
Ensure the callable is an async def.
[ "Ensure", "the", "callable", "is", "an", "async", "def", "." ]
26a6c7e71e87dd1ddc4acb755d70ca30894f7928
https://github.com/asyncdef/aitertools/blob/26a6c7e71e87dd1ddc4acb755d70ca30894f7928/aitertools/__init__.py#L249-L260
train
asyncdef/aitertools
aitertools/__init__.py
tee
def tee(iterable, n=2): """Return n independent iterators from a single iterable. Once tee() has made a split, the original iterable should not be used anywhere else; otherwise, the iterable could get advanced without the tee objects being informed. This itertool may require significant auxiliary ...
python
def tee(iterable, n=2): """Return n independent iterators from a single iterable. Once tee() has made a split, the original iterable should not be used anywhere else; otherwise, the iterable could get advanced without the tee objects being informed. This itertool may require significant auxiliary ...
[ "def", "tee", "(", "iterable", ",", "n", "=", "2", ")", ":", "tees", "=", "tuple", "(", "AsyncTeeIterable", "(", "iterable", ")", "for", "_", "in", "range", "(", "n", ")", ")", "for", "tee", "in", "tees", ":", "tee", ".", "_siblings", "=", "tees"...
Return n independent iterators from a single iterable. Once tee() has made a split, the original iterable should not be used anywhere else; otherwise, the iterable could get advanced without the tee objects being informed. This itertool may require significant auxiliary storage (depending on how m...
[ "Return", "n", "independent", "iterators", "from", "a", "single", "iterable", "." ]
26a6c7e71e87dd1ddc4acb755d70ca30894f7928
https://github.com/asyncdef/aitertools/blob/26a6c7e71e87dd1ddc4acb755d70ca30894f7928/aitertools/__init__.py#L890-L907
train
nocarryr/python-dispatch
pydispatch/properties.py
Property._on_change
def _on_change(self, obj, old, value, **kwargs): """Called internally to emit changes from the instance object The keyword arguments here will be passed to callbacks through the instance object's :meth:`~pydispatch.dispatch.Dispatcher.emit` method. Keyword Args: property: T...
python
def _on_change(self, obj, old, value, **kwargs): """Called internally to emit changes from the instance object The keyword arguments here will be passed to callbacks through the instance object's :meth:`~pydispatch.dispatch.Dispatcher.emit` method. Keyword Args: property: T...
[ "def", "_on_change", "(", "self", ",", "obj", ",", "old", ",", "value", ",", "**", "kwargs", ")", ":", "kwargs", "[", "'property'", "]", "=", "self", "obj", ".", "emit", "(", "self", ".", "name", ",", "obj", ",", "value", ",", "old", "=", "old", ...
Called internally to emit changes from the instance object The keyword arguments here will be passed to callbacks through the instance object's :meth:`~pydispatch.dispatch.Dispatcher.emit` method. Keyword Args: property: The :class:`Property` instance. This is useful if multiple ...
[ "Called", "internally", "to", "emit", "changes", "from", "the", "instance", "object" ]
7c5ca03835c922cbfdfd62772c9e560062c954c7
https://github.com/nocarryr/python-dispatch/blob/7c5ca03835c922cbfdfd62772c9e560062c954c7/pydispatch/properties.py#L96-L114
train
OpenHydrology/floodestimation
floodestimation/parsers.py
FehFileParser.parse_str
def parse_str(self, s): """ Parse string and return relevant object :param s: string to parse :type s: str :return: Parsed object """ self.object = self.parsed_class() in_section = None # Holds name of FEH file section while traversing through file. ...
python
def parse_str(self, s): """ Parse string and return relevant object :param s: string to parse :type s: str :return: Parsed object """ self.object = self.parsed_class() in_section = None # Holds name of FEH file section while traversing through file. ...
[ "def", "parse_str", "(", "self", ",", "s", ")", ":", "self", ".", "object", "=", "self", ".", "parsed_class", "(", ")", "in_section", "=", "None", "for", "line", "in", "s", ".", "split", "(", "'\\n'", ")", ":", "if", "line", ".", "lower", "(", ")...
Parse string and return relevant object :param s: string to parse :type s: str :return: Parsed object
[ "Parse", "string", "and", "return", "relevant", "object" ]
782da7c5abd1348923129efe89fb70003ebb088c
https://github.com/OpenHydrology/floodestimation/blob/782da7c5abd1348923129efe89fb70003ebb088c/floodestimation/parsers.py#L70-L93
train