source_code stringlengths 45 7.24k | function_name stringlengths 1 111 | messages list |
|---|---|---|
def x():
parser = ArgumentParserNoArgHelp(formatter_class=
ArgumentDefaultsRawDescriptionFormatter, description='\n'.join([
"""Prepare alignment of protein-coding DNA sequences.
""", 'Steps:',
" * Any sequences specified by '--purgeseqs' are removed.",
' * Sequences not of length divisible by 3 are removed.',
' * Sequences with ambiguous nucleotides are removed.',
' * Sequences with non-terminal stop codons are removed;',
' terminal stop codons are trimmed.',
' * Sequences that do not encode unique proteins are removed',
" unless they are specified for retention by '--keepseqs'.",
' * A multiple sequence alignment is built using MAFFT.',
" This step is skipped if you specify '--prealigned'.",
' * Sites gapped in reference sequence are stripped.',
' * Sequences with too little protein identity to reference',
' sequence are removed, counting both mismatches and unstripped',
" gaps as differences. Identity cutoff set by '--minidentity'.",
' * Sequences too similar to other sequences are removed. An',
' effort is made to keep one representative of sequences found',
' many times in input set. Uniqueness threshold set ',
" by '--minuniqueness'. You can specify sequences to not",
" remove via '--keepseqs'.",
' * Problematic characters in header names are replaced by',
' underscores. This is any space, comma, colon, semicolon',
' parenthesis, bracket, single quote, or double quote.',
' * An alignment is written, as well as a plot with same root',
" but extension '.pdf' that shows divergence from reference",
' of all sequences retained and purged due to identity or',
' uniqueness.\n', phydmslib.__acknowledgments__, 'Version {0}'.
format(phydmslib.__version__), 'Full documentation at {0}'.format(
phydmslib.__url__)]))
parser.add_argument('inseqs', type=ExistingFile, help=
'FASTA file giving input coding sequences.')
parser.add_argument('alignment', help=
"Name of created output FASTA alignment. PDF plot has same root, but extension '.pdf'."
)
parser.add_argument('refseq', help=
"Reference sequence in 'inseqs': specify substring found ONLY in header for that sequence."
)
parser.set_defaults(prealigned=False)
parser.add_argument('--prealigned', action='store_true', dest=
'prealigned', help=
"Sequences in 'inseqs' are already aligned, do NOT re-align.")
parser.add_argument('--mafft', help=
'Path to MAFFT (http://mafft.cbrc.jp/alignment/software/).',
default='mafft')
parser.add_argument('--minidentity', type=FloatBetweenZeroAndOne, help=
"Purge sequences with <= this protein identity to 'refseq'.",
default=0.7)
parser.add_argument('--minuniqueness', type=IntGreaterThanZero, default
=2, help=
'Require each sequence to have >= this many protein differences relative to other sequences.'
)
parser.add_argument('--purgeseqs', nargs='*', help=
'Specify sequences to always purge. Any sequences with any of the substrings specified here are always removed. The substrings can either be passed as repeated arguments here, or as the name of an existing file which has one substring per line.'
)
parser.add_argument('--keepseqs', nargs='*', help=
"Do not purge any of these sequences for lack of identity or uniqueness. Specified in the same fashion as for '--purgeseqs'."
)
parser.add_argument('-v', '--version', action='version', version=
'%(prog)s {version}'.format(version=phydmslib.__version__))
return parser
| PhyDMSPrepAlignmentParser | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x():\n parser = ArgumentParserNoArgHelp(formatter_class=\n ArgumentDefaultsRawDescriptionFormatter, description='\\n'.join([\n \"\"\"Prepare alignment of protein-coding DNA sequences.\n\"\"\", 'Steps:',\n \" * Any sequences specified by '--purgeseqs' are removed.\",\n ' * Sequences not of length divisible by 3 are removed.',\n ' * Sequences with ambiguous nucleotides are removed.',\n ' * Sequences with non-terminal stop codons are removed;',\n ' terminal stop codons are trimmed.',\n ' * Sequences that do not encode unique proteins are removed',\n \" unless they are specified for retention by '--keepseqs'.\",\n ' * A multiple sequence alignment is built using MAFFT.',\n \" This step is skipped if you specify '--prealigned'.\",\n ' * Sites gapped in reference sequence are stripped.',\n ' * Sequences with too little protein identity to reference',\n ' sequence are removed, counting both mismatches and unstripped',\n \" gaps as differences. Identity cutoff set by '--minidentity'.\",\n ' * Sequences too similar to other sequences are removed. An',\n ' effort is made to keep one representative of sequences found',\n ' many times in input set. Uniqueness threshold set ',\n \" by '--minuniqueness'. You can specify sequences to not\",\n \" remove via '--keepseqs'.\",\n ' * Problematic characters in header names are replaced by',\n ' underscores. This is any space, comma, colon, semicolon',\n ' parenthesis, bracket, single quote, or double quote.',\n ' * An alignment is written, as well as a plot with same root',\n \" but extension '.pdf' that shows divergence from reference\",\n ' of all sequences retained and purged due to identity or',\n ' uniqueness.\\n', phydmslib.__acknowledgments__, 'Version {0}'.\n format(phydmslib.__version__), 'Full documentation at {0}'.format(\n phydmslib.__url__)]))\n parser.add_argument('inseqs', type=ExistingFile, help=\n 'FASTA file giving input coding sequences.')\n parser.add_argument('alignment', help=\n \"Name of created output FASTA alignment. PDF plot has same root, but extension '.pdf'.\"\n )\n parser.add_argument('refseq', help=\n \"Reference sequence in 'inseqs': specify substring found ONLY in header for that sequence.\"\n )\n parser.set_defaults(prealigned=False)\n parser.add_argument('--prealigned', action='store_true', dest=\n 'prealigned', help=\n \"Sequences in 'inseqs' are already aligned, do NOT re-align.\")\n parser.add_argument('--mafft', help=\n 'Path to MAFFT (http://mafft.cbrc.jp/alignment/software/).',\n default='mafft')\n parser.add_argument('--minidentity', type=FloatBetweenZeroAndOne, help=\n \"Purge sequences with <= this protein identity to 'refseq'.\",\n default=0.7)\n parser.add_argument('--minuniqueness', type=IntGreaterThanZero, default\n =2, help=\n 'Require each sequence to have >= this many protein differences relative to other sequences.'\n )\n parser.add_argument('--purgeseqs', nargs='*', help=\n 'Specify sequences to always purge. Any sequences with any of the substrings specified here are always removed. The substrings can either be passed as repeated arguments here, or as the name of an existing file which has one substring per line.'\n )\n parser.add_argument('--keepseqs', nargs='*', help=\n \"Do not purge any of these sequences for lack of identity or uniqueness. Specified in the same fashion as for '--purgeseqs'.\"\n )\n parser.add_argument('-v', '--version', action='version', version=\n '%(prog)s {version}'.format(version=phydmslib.__version__))\n return parser\n",
"role": "user"
},
{
"content": "PhyDMSPrepAlignmentParser",
"role": "assistant"
}
] |
def x():
parser = ArgumentParserNoArgHelp(description=
'Make logo plot of preferences or differential preferences. Uses weblogo (http://weblogo.threeplusone.com/). {0} Version {1}. Full documentation at {2}'
.format(phydmslib.__acknowledgments__, phydmslib.__version__,
phydmslib.__url__), formatter_class=argparse.
ArgumentDefaultsHelpFormatter)
group = parser.add_mutually_exclusive_group(required=True)
group.add_argument('--prefs', type=ExistingFile, help=
"File with amino-acid preferences; same format as input to 'phydms'.")
group.add_argument('--diffprefs', type=ExistingFile, help=
"File with differential preferences; in format output by 'phydms'.")
parser.add_argument('outfile', help='Name of created PDF logo plot.')
parser.add_argument('--stringency', type=FloatGreaterThanEqualToZero,
default=1, help='Stringency parameter to re-scale prefs.')
parser.add_argument('--nperline', type=IntGreaterThanZero, default=70,
help='Number of sites per line.')
parser.add_argument('--numberevery', type=IntGreaterThanZero, default=
10, help='Number sites at this interval.')
parser.add_argument('--mapmetric', default='functionalgroup', choices=[
'kd', 'mw', 'charge', 'functionalgroup'], help=
'Metric used to color amino-acid letters. kd = Kyte-Doolittle hydrophobicity; mw = molecular weight; functionalgroup = divide in 7 groups; charge = charge at neutral pH.'
)
parser.add_argument('--colormap', type=str, default='jet', help=
"Name of `matplotlib` color map for amino acids when `--mapmetric` is 'kd' or 'mw'."
)
parser.add_argument('--diffprefheight', type=FloatGreaterThanZero,
default=1.0, help='Height of diffpref logo in each direction.')
parser.add_argument('--omegabysite', help=
"Overlay omega on logo plot. Specify '*_omegabysite.txt' file from 'phydms'."
, type=ExistingFileOrNone)
parser.add_argument('--minP', type=FloatGreaterThanZero, default=0.0001,
help="Min plotted P-value for '--omegabysite' overlay.")
parser.add_argument('-v', '--version', action='version', version=
'%(prog)s {version}'.format(version=phydmslib.__version__))
return parser
| PhyDMSLogoPlotParser | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x():\n parser = ArgumentParserNoArgHelp(description=\n 'Make logo plot of preferences or differential preferences. Uses weblogo (http://weblogo.threeplusone.com/). {0} Version {1}. Full documentation at {2}'\n .format(phydmslib.__acknowledgments__, phydmslib.__version__,\n phydmslib.__url__), formatter_class=argparse.\n ArgumentDefaultsHelpFormatter)\n group = parser.add_mutually_exclusive_group(required=True)\n group.add_argument('--prefs', type=ExistingFile, help=\n \"File with amino-acid preferences; same format as input to 'phydms'.\")\n group.add_argument('--diffprefs', type=ExistingFile, help=\n \"File with differential preferences; in format output by 'phydms'.\")\n parser.add_argument('outfile', help='Name of created PDF logo plot.')\n parser.add_argument('--stringency', type=FloatGreaterThanEqualToZero,\n default=1, help='Stringency parameter to re-scale prefs.')\n parser.add_argument('--nperline', type=IntGreaterThanZero, default=70,\n help='Number of sites per line.')\n parser.add_argument('--numberevery', type=IntGreaterThanZero, default=\n 10, help='Number sites at this interval.')\n parser.add_argument('--mapmetric', default='functionalgroup', choices=[\n 'kd', 'mw', 'charge', 'functionalgroup'], help=\n 'Metric used to color amino-acid letters. kd = Kyte-Doolittle hydrophobicity; mw = molecular weight; functionalgroup = divide in 7 groups; charge = charge at neutral pH.'\n )\n parser.add_argument('--colormap', type=str, default='jet', help=\n \"Name of `matplotlib` color map for amino acids when `--mapmetric` is 'kd' or 'mw'.\"\n )\n parser.add_argument('--diffprefheight', type=FloatGreaterThanZero,\n default=1.0, help='Height of diffpref logo in each direction.')\n parser.add_argument('--omegabysite', help=\n \"Overlay omega on logo plot. Specify '*_omegabysite.txt' file from 'phydms'.\"\n , type=ExistingFileOrNone)\n parser.add_argument('--minP', type=FloatGreaterThanZero, default=0.0001,\n help=\"Min plotted P-value for '--omegabysite' overlay.\")\n parser.add_argument('-v', '--version', action='version', version=\n '%(prog)s {version}'.format(version=phydmslib.__version__))\n return parser\n",
"role": "user"
},
{
"content": "PhyDMSLogoPlotParser",
"role": "assistant"
}
] |
def x():
parser = ArgumentParserNoArgHelp(description=
"Comprehensive phylogenetic model comparison and detection of selection informed by deep mutational scanning data. This program runs 'phydms' repeatedly to compare substitution models and detect selection. {0} Version {1}. Full documentation at {2}"
.format(phydmslib.__acknowledgments__, phydmslib.__version__,
phydmslib.__url__), formatter_class=argparse.
ArgumentDefaultsHelpFormatter)
parser.add_argument('outprefix', help='Output file prefix.', type=str)
parser.add_argument('alignment', help=
'Existing FASTA file with aligned codon sequences.', type=ExistingFile)
parser.add_argument('prefsfiles', help=
'Existing files with site-specific amino-acid preferences.', type=
ExistingFile, nargs='+')
group = parser.add_mutually_exclusive_group(required=True)
group.add_argument('--raxml', help="Path to RAxML (e.g., 'raxml')")
group.add_argument('--tree', type=ExistingFile, help=
'Existing Newick file giving input tree.')
parser.add_argument('--ncpus', default=-1, help=
'Use this many CPUs; -1 means all available.', type=int)
parser.add_argument('--brlen', choices=['scale', 'optimize'], default=
'optimize', help=
'How to handle branch lengths: scale by single parameter or optimize each one'
)
parser.set_defaults(omegabysite=False)
parser.add_argument('--omegabysite', dest='omegabysite', action=
'store_true', help='Fit omega (dN/dS) for each site.')
parser.set_defaults(diffprefsbysite=False)
parser.add_argument('--diffprefsbysite', dest='diffprefsbysite', action
='store_true', help='Fit differential preferences for each site.')
parser.set_defaults(gammaomega=False)
parser.add_argument('--gammaomega', dest='gammaomega', action=
'store_true', help='Fit ExpCM with gamma distributed omega.')
parser.set_defaults(gammabeta=False)
parser.add_argument('--gammabeta', dest='gammabeta', action=
'store_true', help='Fit ExpCM with gamma distributed beta.')
parser.set_defaults(noavgprefs=False)
parser.add_argument('--no-avgprefs', dest='noavgprefs', action=
'store_true', help=
'No fitting of models with preferences averaged across sites for ExpCM.'
)
parser.set_defaults(randprefs=False)
parser.add_argument('--randprefs', dest='randprefs', action=
'store_true', help='Include ExpCM models with randomized preferences.')
parser.add_argument('-v', '--version', action='version', version=
'%(prog)s {version}'.format(version=phydmslib.__version__))
return parser
| PhyDMSComprehensiveParser | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x():\n parser = ArgumentParserNoArgHelp(description=\n \"Comprehensive phylogenetic model comparison and detection of selection informed by deep mutational scanning data. This program runs 'phydms' repeatedly to compare substitution models and detect selection. {0} Version {1}. Full documentation at {2}\"\n .format(phydmslib.__acknowledgments__, phydmslib.__version__,\n phydmslib.__url__), formatter_class=argparse.\n ArgumentDefaultsHelpFormatter)\n parser.add_argument('outprefix', help='Output file prefix.', type=str)\n parser.add_argument('alignment', help=\n 'Existing FASTA file with aligned codon sequences.', type=ExistingFile)\n parser.add_argument('prefsfiles', help=\n 'Existing files with site-specific amino-acid preferences.', type=\n ExistingFile, nargs='+')\n group = parser.add_mutually_exclusive_group(required=True)\n group.add_argument('--raxml', help=\"Path to RAxML (e.g., 'raxml')\")\n group.add_argument('--tree', type=ExistingFile, help=\n 'Existing Newick file giving input tree.')\n parser.add_argument('--ncpus', default=-1, help=\n 'Use this many CPUs; -1 means all available.', type=int)\n parser.add_argument('--brlen', choices=['scale', 'optimize'], default=\n 'optimize', help=\n 'How to handle branch lengths: scale by single parameter or optimize each one'\n )\n parser.set_defaults(omegabysite=False)\n parser.add_argument('--omegabysite', dest='omegabysite', action=\n 'store_true', help='Fit omega (dN/dS) for each site.')\n parser.set_defaults(diffprefsbysite=False)\n parser.add_argument('--diffprefsbysite', dest='diffprefsbysite', action\n ='store_true', help='Fit differential preferences for each site.')\n parser.set_defaults(gammaomega=False)\n parser.add_argument('--gammaomega', dest='gammaomega', action=\n 'store_true', help='Fit ExpCM with gamma distributed omega.')\n parser.set_defaults(gammabeta=False)\n parser.add_argument('--gammabeta', dest='gammabeta', action=\n 'store_true', help='Fit ExpCM with gamma distributed beta.')\n parser.set_defaults(noavgprefs=False)\n parser.add_argument('--no-avgprefs', dest='noavgprefs', action=\n 'store_true', help=\n 'No fitting of models with preferences averaged across sites for ExpCM.'\n )\n parser.set_defaults(randprefs=False)\n parser.add_argument('--randprefs', dest='randprefs', action=\n 'store_true', help='Include ExpCM models with randomized preferences.')\n parser.add_argument('-v', '--version', action='version', version=\n '%(prog)s {version}'.format(version=phydmslib.__version__))\n return parser\n",
"role": "user"
},
{
"content": "PhyDMSComprehensiveParser",
"role": "assistant"
}
] |
def x():
parser = ArgumentParserNoArgHelp(description=
'Phylogenetic analysis informed by deep mutational scanning data. {0} Version {1}. Full documentation at {2}'
.format(phydmslib.__acknowledgments__, phydmslib.__version__,
phydmslib.__url__), formatter_class=argparse.
ArgumentDefaultsHelpFormatter)
parser.add_argument('alignment', type=ExistingFile, help=
'Existing FASTA file with aligned codon sequences.')
parser.add_argument('tree', type=ExistingFile, help=
'Existing Newick file giving input tree.')
parser.add_argument('model', type=ModelOption, help=
"Substitution model: ExpCM_<prefsfile> or YNGKP_<m> (where <m> is {0}). For ExpCM, <prefsfile> has first column labeled 'site' and others labeled by 1-letter amino-acid code."
.format(', '.join(yngkp_modelvariants)))
parser.add_argument('outprefix', help='Output file prefix.', type=str)
parser.add_argument('--brlen', choices=['scale', 'optimize'], default=
'optimize', help=
'How to handle branch lengths: scale by single parameter or optimize each one'
)
parser.set_defaults(gammaomega=False)
parser.add_argument('--gammaomega', action='store_true', dest=
'gammaomega', help=
"Omega for ExpCM from gamma distribution rather than single value. To achieve same for YNGKP, use 'model' of YNGKP_M5."
)
parser.set_defaults(gammabeta=False)
parser.add_argument('--gammabeta', action='store_true', dest=
'gammabeta', help=
'Beta for ExpCM from gamma distribution rather than single value.')
parser.set_defaults(omegabysite=False)
parser.add_argument('--omegabysite', dest='omegabysite', action=
'store_true', help='Fit omega (dN/dS) for each site.')
parser.set_defaults(omegabysite_fixsyn=False)
parser.add_argument('--omegabysite_fixsyn', dest='omegabysite_fixsyn',
action='store_true', help=
"For '--omegabysite', assign all sites same dS rather than fit for each site."
)
parser.set_defaults(diffprefsbysite=False)
parser.add_argument('--diffprefsbysite', dest='diffprefsbysite', action
='store_true', help='Fit differential preferences for each site.')
parser.add_argument('--diffprefsprior', default='invquadratic,150,0.5',
type=diffPrefsPrior, help=
"Regularizing prior for '--diffprefsbysite': 'invquadratic,C1,C2' is prior in Bloom, Biology Direct, 12:1."
)
parser.set_defaults(fitphi=False)
parser.add_argument('--fitphi', action='store_true', dest='fitphi',
help=
'Fit ExpCM phi rather than setting so stationary state matches alignment frequencies.'
)
parser.set_defaults(randprefs=False)
parser.add_argument('--randprefs', dest='randprefs', action=
'store_true', help='Randomize preferences among sites for ExpCM.')
parser.set_defaults(avgprefs=False)
parser.add_argument('--avgprefs', dest='avgprefs', action='store_true',
help='Average preferences across sites for ExpCM.')
parser.add_argument('--divpressure', type=ExistingFileOrNone, help=
'Known diversifying pressure at sites: file with column 1 = position, column 2 = diversification pressure; columns space-, tab-, or comma-delimited.'
)
parser.add_argument('--ncpus', default=1, type=int, help=
'Use this many CPUs; -1 means all available.')
parser.add_argument('--fitprefsmethod', choices=[1, 2], default=2, help
='Implementation to we use when fitting prefs.', type=int)
parser.add_argument('--ncats', default=4, type=IntGreaterThanOne, help=
'Number of categories for gamma-distribution.')
parser.add_argument('--minbrlen', type=FloatGreaterThanZero, default=
phydmslib.constants.ALMOST_ZERO, help=
"Adjust all branch lengths in starting 'tree' to >= this.")
parser.add_argument('--minpref', default=0.002, type=
FloatGreaterThanZero, help=
"Adjust all preferences in ExpCM 'prefsfile' to >= this.")
parser.add_argument('--seed', type=int, default=1, help=
'Random number seed.')
parser.add_argument('--initparams', type=ExistingFile, help=
"Initialize model params from this file, which should be format of '*_modelparams.txt' file created by 'phydms' with this model."
)
parser.set_defaults(profile=False)
parser.add_argument('--profile', dest='profile', action='store_true',
help=
'Profile likelihood maximization, write pstats files. For code-development purposes.'
)
parser.set_defaults(opt_details=False)
parser.add_argument('--opt_details', dest='opt_details', action=
'store_true', help='Print details about optimization')
parser.set_defaults(nograd=False)
parser.add_argument('--nograd', dest='nograd', action='store_true',
help='Do not use gradients for likelihood maximization.')
parser.add_argument('-v', '--version', action='version', version=
'%(prog)s {version}'.format(version=phydmslib.__version__))
return parser
| PhyDMSParser | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x():\n parser = ArgumentParserNoArgHelp(description=\n 'Phylogenetic analysis informed by deep mutational scanning data. {0} Version {1}. Full documentation at {2}'\n .format(phydmslib.__acknowledgments__, phydmslib.__version__,\n phydmslib.__url__), formatter_class=argparse.\n ArgumentDefaultsHelpFormatter)\n parser.add_argument('alignment', type=ExistingFile, help=\n 'Existing FASTA file with aligned codon sequences.')\n parser.add_argument('tree', type=ExistingFile, help=\n 'Existing Newick file giving input tree.')\n parser.add_argument('model', type=ModelOption, help=\n \"Substitution model: ExpCM_<prefsfile> or YNGKP_<m> (where <m> is {0}). For ExpCM, <prefsfile> has first column labeled 'site' and others labeled by 1-letter amino-acid code.\"\n .format(', '.join(yngkp_modelvariants)))\n parser.add_argument('outprefix', help='Output file prefix.', type=str)\n parser.add_argument('--brlen', choices=['scale', 'optimize'], default=\n 'optimize', help=\n 'How to handle branch lengths: scale by single parameter or optimize each one'\n )\n parser.set_defaults(gammaomega=False)\n parser.add_argument('--gammaomega', action='store_true', dest=\n 'gammaomega', help=\n \"Omega for ExpCM from gamma distribution rather than single value. To achieve same for YNGKP, use 'model' of YNGKP_M5.\"\n )\n parser.set_defaults(gammabeta=False)\n parser.add_argument('--gammabeta', action='store_true', dest=\n 'gammabeta', help=\n 'Beta for ExpCM from gamma distribution rather than single value.')\n parser.set_defaults(omegabysite=False)\n parser.add_argument('--omegabysite', dest='omegabysite', action=\n 'store_true', help='Fit omega (dN/dS) for each site.')\n parser.set_defaults(omegabysite_fixsyn=False)\n parser.add_argument('--omegabysite_fixsyn', dest='omegabysite_fixsyn',\n action='store_true', help=\n \"For '--omegabysite', assign all sites same dS rather than fit for each site.\"\n )\n parser.set_defaults(diffprefsbysite=False)\n parser.add_argument('--diffprefsbysite', dest='diffprefsbysite', action\n ='store_true', help='Fit differential preferences for each site.')\n parser.add_argument('--diffprefsprior', default='invquadratic,150,0.5',\n type=diffPrefsPrior, help=\n \"Regularizing prior for '--diffprefsbysite': 'invquadratic,C1,C2' is prior in Bloom, Biology Direct, 12:1.\"\n )\n parser.set_defaults(fitphi=False)\n parser.add_argument('--fitphi', action='store_true', dest='fitphi',\n help=\n 'Fit ExpCM phi rather than setting so stationary state matches alignment frequencies.'\n )\n parser.set_defaults(randprefs=False)\n parser.add_argument('--randprefs', dest='randprefs', action=\n 'store_true', help='Randomize preferences among sites for ExpCM.')\n parser.set_defaults(avgprefs=False)\n parser.add_argument('--avgprefs', dest='avgprefs', action='store_true',\n help='Average preferences across sites for ExpCM.')\n parser.add_argument('--divpressure', type=ExistingFileOrNone, help=\n 'Known diversifying pressure at sites: file with column 1 = position, column 2 = diversification pressure; columns space-, tab-, or comma-delimited.'\n )\n parser.add_argument('--ncpus', default=1, type=int, help=\n 'Use this many CPUs; -1 means all available.')\n parser.add_argument('--fitprefsmethod', choices=[1, 2], default=2, help\n ='Implementation to we use when fitting prefs.', type=int)\n parser.add_argument('--ncats', default=4, type=IntGreaterThanOne, help=\n 'Number of categories for gamma-distribution.')\n parser.add_argument('--minbrlen', type=FloatGreaterThanZero, default=\n phydmslib.constants.ALMOST_ZERO, help=\n \"Adjust all branch lengths in starting 'tree' to >= this.\")\n parser.add_argument('--minpref', default=0.002, type=\n FloatGreaterThanZero, help=\n \"Adjust all preferences in ExpCM 'prefsfile' to >= this.\")\n parser.add_argument('--seed', type=int, default=1, help=\n 'Random number seed.')\n parser.add_argument('--initparams', type=ExistingFile, help=\n \"Initialize model params from this file, which should be format of '*_modelparams.txt' file created by 'phydms' with this model.\"\n )\n parser.set_defaults(profile=False)\n parser.add_argument('--profile', dest='profile', action='store_true',\n help=\n 'Profile likelihood maximization, write pstats files. For code-development purposes.'\n )\n parser.set_defaults(opt_details=False)\n parser.add_argument('--opt_details', dest='opt_details', action=\n 'store_true', help='Print details about optimization')\n parser.set_defaults(nograd=False)\n parser.add_argument('--nograd', dest='nograd', action='store_true',\n help='Do not use gradients for likelihood maximization.')\n parser.add_argument('-v', '--version', action='version', version=\n '%(prog)s {version}'.format(version=phydmslib.__version__))\n return parser\n",
"role": "user"
},
{
"content": "PhyDMSParser",
"role": "assistant"
}
] |
def x(self, message):
sys.stderr.write('error: %s\n\n' % message)
self.print_help()
sys.exit(2)
| error | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, message):\n sys.stderr.write('error: %s\\n\\n' % message)\n self.print_help()\n sys.exit(2)\n",
"role": "user"
},
{
"content": "error",
"role": "assistant"
}
] |
def x(self, method=None, copy=False):
if copy:
return SymmEdgePairGraph(self._pairs.copy(), num_vertices=self.
_num_vertices)
shape = self._num_vertices, self._num_vertices
flat_inds = np.union1d(np.ravel_multi_index(self._pairs.T, shape), np.
ravel_multi_index(self._pairs.T[::-1], shape))
self._pairs = np.transpose(np.unravel_index(flat_inds, shape))
return self
| symmetrize | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, method=None, copy=False):\n if copy:\n return SymmEdgePairGraph(self._pairs.copy(), num_vertices=self.\n _num_vertices)\n shape = self._num_vertices, self._num_vertices\n flat_inds = np.union1d(np.ravel_multi_index(self._pairs.T, shape), np.\n ravel_multi_index(self._pairs.T[::-1], shape))\n self._pairs = np.transpose(np.unravel_index(flat_inds, shape))\n return self\n",
"role": "user"
},
{
"content": "symmetrize",
"role": "assistant"
}
] |
def x(self, from_idx, to_idx, symmetric=False, copy=False):
flat_inds = self._pairs.dot((self._num_vertices, 1))
to_remove = np.minimum(from_idx, to_idx) * self._num_vertices + np.maximum(
from_idx, to_idx)
mask = np.in1d(flat_inds, to_remove, invert=True)
res = self.copy() if copy else self
res._pairs = res._pairs[mask]
res._offdiag_mask = res._offdiag_mask[mask]
return res
| remove_edges | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, from_idx, to_idx, symmetric=False, copy=False):\n flat_inds = self._pairs.dot((self._num_vertices, 1))\n to_remove = np.minimum(from_idx, to_idx) * self._num_vertices + np.maximum(\n from_idx, to_idx)\n mask = np.in1d(flat_inds, to_remove, invert=True)\n res = self.copy() if copy else self\n res._pairs = res._pairs[mask]\n res._offdiag_mask = res._offdiag_mask[mask]\n return res\n",
"role": "user"
},
{
"content": "remove_edges",
"role": "assistant"
}
] |
def x(self, from_idx, to_idx, weight=1, symmetric=False, copy=False):
raise NotImplementedError()
| add_edges | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, from_idx, to_idx, weight=1, symmetric=False, copy=False):\n raise NotImplementedError()\n",
"role": "user"
},
{
"content": "add_edges",
"role": "assistant"
}
] |
def x(self, weight=None, copy=False):
ii = np.arange(self.num_vertices())
return self.add_edges(ii, ii, weight=weight, symmetric=False, copy=copy)
| add_self_edges | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, weight=None, copy=False):\n ii = np.arange(self.num_vertices())\n return self.add_edges(ii, ii, weight=weight, symmetric=False, copy=copy)\n",
"role": "user"
},
{
"content": "add_self_edges",
"role": "assistant"
}
] |
def x(self, weight, edges=None, copy=False):
if not self.is_weighted():
warnings.warn(
'Cannot supply weights for unweighted graph; ignoring call to reweight'
)
return self
if edges is None:
return self._update_edges(weight, copy=copy)
ii, jj = self.pairs()[edges].T
return self.add_edges(ii, jj, weight=weight, symmetric=False, copy=copy)
| reweight | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, weight, edges=None, copy=False):\n if not self.is_weighted():\n warnings.warn(\n 'Cannot supply weights for unweighted graph; ignoring call to reweight'\n )\n return self\n if edges is None:\n return self._update_edges(weight, copy=copy)\n ii, jj = self.pairs()[edges].T\n return self.add_edges(ii, jj, weight=weight, symmetric=False, copy=copy)\n",
"role": "user"
},
{
"content": "reweight",
"role": "assistant"
}
] |
def x(self, coords, metric='l2', copy=False):
if not self.is_weighted():
warnings.warn(
'Cannot supply weights for unweighted graph; ignoring call to reweight_by_distance'
)
return self
ii, jj = self.pairs().T
if metric == 'precomputed':
assert coords.ndim == 2 and coords.shape[0] == coords.shape[1]
d = coords[ii, jj]
else:
d = paired_distances(coords[ii], coords[jj], metric=metric)
return self._update_edges(d, copy=copy)
| reweight_by_distance | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, coords, metric='l2', copy=False):\n if not self.is_weighted():\n warnings.warn(\n 'Cannot supply weights for unweighted graph; ignoring call to reweight_by_distance'\n )\n return self\n ii, jj = self.pairs().T\n if metric == 'precomputed':\n assert coords.ndim == 2 and coords.shape[0] == coords.shape[1]\n d = coords[ii, jj]\n else:\n d = paired_distances(coords[ii], coords[jj], metric=metric)\n return self._update_edges(d, copy=copy)\n",
"role": "user"
},
{
"content": "reweight_by_distance",
"role": "assistant"
}
] |
def x(self, kind='out', weighted=True):
if kind == 'out':
axis = 1
adj = self.matrix('dense', 'csc')
else:
axis = 0
adj = self.matrix('dense', 'csr')
if not weighted and self.is_weighted():
d = (adj != 0).sum(axis=axis)
else:
d = adj.sum(axis=axis)
return np.asarray(d).ravel()
| degree | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, kind='out', weighted=True):\n if kind == 'out':\n axis = 1\n adj = self.matrix('dense', 'csc')\n else:\n axis = 0\n adj = self.matrix('dense', 'csr')\n if not weighted and self.is_weighted():\n d = (adj != 0).sum(axis=axis)\n else:\n d = adj.sum(axis=axis)\n return np.asarray(d).ravel()\n",
"role": "user"
},
{
"content": "degree",
"role": "assistant"
}
] |
def x(self, weighted=None):
import igraph
ig = igraph.Graph(n=self.num_vertices(), edges=self.pairs().tolist(),
directed=self.is_directed())
if weighted is not False and self.is_weighted():
ig.es['weight'] = self.edge_weights()
return ig
| to_igraph | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, weighted=None):\n import igraph\n ig = igraph.Graph(n=self.num_vertices(), edges=self.pairs().tolist(),\n directed=self.is_directed())\n if weighted is not False and self.is_weighted():\n ig.es['weight'] = self.edge_weights()\n return ig\n",
"role": "user"
},
{
"content": "to_igraph",
"role": "assistant"
}
] |
def x(self):
import graph_tool
gt = graph_tool.Graph(directed=self.is_directed())
gt.add_edge_list(self.pairs())
if self.is_weighted():
weights = gt.new_edge_property('double')
for e, w in zip(gt.edges(), self.edge_weights()):
weights[e] = w
gt.edge_properties['weight'] = weights
return gt
| to_graph_tool | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n import graph_tool\n gt = graph_tool.Graph(directed=self.is_directed())\n gt.add_edge_list(self.pairs())\n if self.is_weighted():\n weights = gt.new_edge_property('double')\n for e, w in zip(gt.edges(), self.edge_weights()):\n weights[e] = w\n gt.edge_properties['weight'] = weights\n return gt\n",
"role": "user"
},
{
"content": "to_graph_tool",
"role": "assistant"
}
] |
def x(self, directed=None):
import networkx as nx
directed = directed if directed is not None else self.is_directed()
cls = nx.DiGraph if directed else nx.Graph
adj = self.matrix()
if ss.issparse(adj):
return nx.from_scipy_sparse_matrix(adj, create_using=cls())
return nx.from_numpy_matrix(adj, create_using=cls())
| to_networkx | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, directed=None):\n import networkx as nx\n directed = directed if directed is not None else self.is_directed()\n cls = nx.DiGraph if directed else nx.Graph\n adj = self.matrix()\n if ss.issparse(adj):\n return nx.from_scipy_sparse_matrix(adj, create_using=cls())\n return nx.from_numpy_matrix(adj, create_using=cls())\n",
"role": "user"
},
{
"content": "to_networkx",
"role": "assistant"
}
] |
def x(z, m):
try:
nz = len(z)
z = np.array(z)
except TypeError:
z = np.array([z])
nz = len(z)
try:
nm = len(m)
m = np.array(m)
except TypeError:
m = np.array([m])
nm = len(m)
if (z < 0).any() or (m < 0).any():
raise ValueError('z and m must be positive')
if nz != nm and nz > 1 and nm > 1:
raise ValueError(
'z and m arrays must be either equal in length, OR of different length with one of length 1.'
)
else:
if type(z) != np.ndarray:
z = np.array(z)
if type(m) != np.ndarray:
m = np.array(m)
return z, m
| check_inputs | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(z, m):\n try:\n nz = len(z)\n z = np.array(z)\n except TypeError:\n z = np.array([z])\n nz = len(z)\n try:\n nm = len(m)\n m = np.array(m)\n except TypeError:\n m = np.array([m])\n nm = len(m)\n if (z < 0).any() or (m < 0).any():\n raise ValueError('z and m must be positive')\n if nz != nm and nz > 1 and nm > 1:\n raise ValueError(\n 'z and m arrays must be either equal in length, OR of different length with one of length 1.'\n )\n else:\n if type(z) != np.ndarray:\n z = np.array(z)\n if type(m) != np.ndarray:\n m = np.array(m)\n return z, m\n",
"role": "user"
},
{
"content": "check_inputs",
"role": "assistant"
}
] |
def x(z, m, h=h, Om_M=Om_M, Om_L=Om_L):
z, m = _check_inputs(z, m)
x = 1.0 / (1.0 + z) * (Om_L / Om_M) ** (1.0 / 3.0)
intEQ12 = np.zeros(len(x))
for i in range(len(x)):
temp = integrate.quad(lambda v: (v / (1 + v ** 3.0)) ** 1.5, 0, x[i])
intEQ12[i] = temp[0]
Da = 2.5 * (Om_M / Om_L) ** (1.0 / 3.0) * (np.sqrt(1.0 + x ** 3.0) / x **
1.5) * intEQ12
y = 1000000000000.0 / (h * m)
sigma = Da * (16.9 * y ** 0.41) / (1.0 + 1.102 * y ** 0.2 + 6.22 * y **
0.333)
c0 = 3.681
c1 = 5.033
alpha = 6.948
x0 = 0.424
s0 = 1.047
s1 = 1.646
beta = 7.386
x1 = 0.526
cmin = c0 + (c1 - c0) * (1.0 / np.pi * np.arctan(alpha * (x - x0)) + 0.5)
smin = s0 + (s1 - s0) * (1.0 / np.pi * np.arctan(beta * (x - x1)) + 0.5)
cmin1393 = c0 + (c1 - c0) * (1.0 / np.pi * np.arctan(alpha * (1.393 -
x0)) + 0.5)
smin1393 = s0 + (s1 - s0) * (1.0 / np.pi * np.arctan(beta * (1.393 - x1
)) + 0.5)
B0 = cmin / cmin1393
B1 = smin / smin1393
sigma_prime = B1 * sigma
A = 2.881
b = 1.257
c = 1.022
d = 0.06
Cs = A * ((sigma_prime / b) ** c + 1.0) * np.exp(d / sigma_prime ** 2.0)
concentration = B0 * Cs
return concentration
| c_Prada | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(z, m, h=h, Om_M=Om_M, Om_L=Om_L):\n z, m = _check_inputs(z, m)\n x = 1.0 / (1.0 + z) * (Om_L / Om_M) ** (1.0 / 3.0)\n intEQ12 = np.zeros(len(x))\n for i in range(len(x)):\n temp = integrate.quad(lambda v: (v / (1 + v ** 3.0)) ** 1.5, 0, x[i])\n intEQ12[i] = temp[0]\n Da = 2.5 * (Om_M / Om_L) ** (1.0 / 3.0) * (np.sqrt(1.0 + x ** 3.0) / x **\n 1.5) * intEQ12\n y = 1000000000000.0 / (h * m)\n sigma = Da * (16.9 * y ** 0.41) / (1.0 + 1.102 * y ** 0.2 + 6.22 * y **\n 0.333)\n c0 = 3.681\n c1 = 5.033\n alpha = 6.948\n x0 = 0.424\n s0 = 1.047\n s1 = 1.646\n beta = 7.386\n x1 = 0.526\n cmin = c0 + (c1 - c0) * (1.0 / np.pi * np.arctan(alpha * (x - x0)) + 0.5)\n smin = s0 + (s1 - s0) * (1.0 / np.pi * np.arctan(beta * (x - x1)) + 0.5)\n cmin1393 = c0 + (c1 - c0) * (1.0 / np.pi * np.arctan(alpha * (1.393 -\n x0)) + 0.5)\n smin1393 = s0 + (s1 - s0) * (1.0 / np.pi * np.arctan(beta * (1.393 - x1\n )) + 0.5)\n B0 = cmin / cmin1393\n B1 = smin / smin1393\n sigma_prime = B1 * sigma\n A = 2.881\n b = 1.257\n c = 1.022\n d = 0.06\n Cs = A * ((sigma_prime / b) ** c + 1.0) * np.exp(d / sigma_prime ** 2.0)\n concentration = B0 * Cs\n return concentration\n",
"role": "user"
},
{
"content": "c_Prada",
"role": "assistant"
}
] |
def x(z, m, h=h):
z, m = _check_inputs(z, m)
a = 0.52 + 0.385 * np.exp(-0.617 * z ** 1.21)
b = -0.101 + 0.026 * z
logc200 = a + b * np.log10(m * h / 10.0 ** 12)
concentration = 10.0 ** logc200
return concentration
| c_DuttonMaccio | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(z, m, h=h):\n z, m = _check_inputs(z, m)\n a = 0.52 + 0.385 * np.exp(-0.617 * z ** 1.21)\n b = -0.101 + 0.026 * z\n logc200 = a + b * np.log10(m * h / 10.0 ** 12)\n concentration = 10.0 ** logc200\n return concentration\n",
"role": "user"
},
{
"content": "c_DuttonMaccio",
"role": "assistant"
}
] |
def x(z, m, h=h):
z, m = _check_inputs(z, m)
M_pivot = 2000000000000.0 / h
A = 5.71
B = -0.084
C = -0.47
concentration = A * (m / M_pivot) ** B * (1 + z) ** C
return concentration
| c_Duffy | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(z, m, h=h):\n z, m = _check_inputs(z, m)\n M_pivot = 2000000000000.0 / h\n A = 5.71\n B = -0.084\n C = -0.47\n concentration = A * (m / M_pivot) ** B * (1 + z) ** C\n return concentration\n",
"role": "user"
},
{
"content": "c_Duffy",
"role": "assistant"
}
] |
def x(self, which, label, **kwargs):
prop_default = {'fontsize': 18}
for prop, default in prop_default.items():
kwargs[prop] = kwargs.get(prop, default)
setattr(self.label, which, label)
setattr(self.label, which + '_kwargs', kwargs)
return
| set_label | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, which, label, **kwargs):\n prop_default = {'fontsize': 18}\n for prop, default in prop_default.items():\n kwargs[prop] = kwargs.get(prop, default)\n setattr(self.label, which, label)\n setattr(self.label, which + '_kwargs', kwargs)\n return\n",
"role": "user"
},
{
"content": "set_label",
"role": "assistant"
}
] |
def x(self, which, lims, d, scale, reverse=False):
setattr(self.limits, which + 'lims', lims)
setattr(self.limits, 'd' + which, d)
setattr(self.limits, which + 'scale', scale)
if reverse:
setattr(self.limits, 'reverse_' + which + '_axis', True)
return
| set_axis_limits | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, which, lims, d, scale, reverse=False):\n setattr(self.limits, which + 'lims', lims)\n setattr(self.limits, 'd' + which, d)\n setattr(self.limits, which + 'scale', scale)\n if reverse:\n setattr(self.limits, 'reverse_' + which + '_axis', True)\n return\n",
"role": "user"
},
{
"content": "set_axis_limits",
"role": "assistant"
}
] |
def x(self, xlims, dx, xscale, reverse=False):
self._set_axis_limits('x', xlims, dx, xscale, reverse)
return
| set_xlim | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, xlims, dx, xscale, reverse=False):\n self._set_axis_limits('x', xlims, dx, xscale, reverse)\n return\n",
"role": "user"
},
{
"content": "set_xlim",
"role": "assistant"
}
] |
def x(self, xlims, dx, xscale, reverse=False):
self._set_axis_limits('y', xlims, dx, xscale, reverse)
return
| set_ylim | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, xlims, dx, xscale, reverse=False):\n self._set_axis_limits('y', xlims, dx, xscale, reverse)\n return\n",
"role": "user"
},
{
"content": "set_ylim",
"role": "assistant"
}
] |
def x(self, labels=None, **kwargs):
if 'size' in kwargs:
if 'prop' not in kwargs:
kwargs['prop'] = {'size': kwargs['size']}
else:
kwargs['prop']['size'] = kwargs['size']
del kwargs['size']
self.legend.add_legend = True
self.legend.legend_labels = labels
self.legend.legend_kwargs = kwargs
return
| add_legend | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, labels=None, **kwargs):\n if 'size' in kwargs:\n if 'prop' not in kwargs:\n kwargs['prop'] = {'size': kwargs['size']}\n else:\n kwargs['prop']['size'] = kwargs['size']\n del kwargs['size']\n self.legend.add_legend = True\n self.legend.legend_labels = labels\n self.legend.legend_kwargs = kwargs\n return\n",
"role": "user"
},
{
"content": "add_legend",
"role": "assistant"
}
] |
def x(self, name=None, label=None, x_column_label=None, y_column_label=None,
index=None, control=False):
if name is None and label is None and index is None:
raise ValueError('Attempting to add a dataset without' +
'supplying index or file information.')
if index is None:
trans_dict = DataImportContainer()
if name is not None:
trans_dict.file_name = name
if label is not None:
trans_dict.label = label
if x_column_label is not None:
trans_dict.x_column_label = x_column_label
if y_column_label is not None:
trans_dict.y_column_label = y_column_label
if control:
self.control = trans_dict
else:
if 'file' not in self.__dict__:
self.file = []
self.file.append(trans_dict)
elif control:
self.control = DataImportContainer()
self.control.index = index
else:
if 'indices' not in self.__dict__:
self.indices = []
self.indices.append(index)
return
| add_dataset | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, name=None, label=None, x_column_label=None, y_column_label=None,\n index=None, control=False):\n if name is None and label is None and index is None:\n raise ValueError('Attempting to add a dataset without' +\n 'supplying index or file information.')\n if index is None:\n trans_dict = DataImportContainer()\n if name is not None:\n trans_dict.file_name = name\n if label is not None:\n trans_dict.label = label\n if x_column_label is not None:\n trans_dict.x_column_label = x_column_label\n if y_column_label is not None:\n trans_dict.y_column_label = y_column_label\n if control:\n self.control = trans_dict\n else:\n if 'file' not in self.__dict__:\n self.file = []\n self.file.append(trans_dict)\n elif control:\n self.control = DataImportContainer()\n self.control.index = index\n else:\n if 'indices' not in self.__dict__:\n self.indices = []\n self.indices.append(index)\n return\n",
"role": "user"
},
{
"content": "add_dataset",
"role": "assistant"
}
] |
def x(self, output_path, **kwargs):
self.figure.save_figure = True
self.figure.output_path = output_path
self.figure.savefig_kwargs = kwargs
return
| savefig | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, output_path, **kwargs):\n self.figure.save_figure = True\n self.figure.output_path = output_path\n self.figure.savefig_kwargs = kwargs\n return\n",
"role": "user"
},
{
"content": "savefig",
"role": "assistant"
}
] |
def x(self, width, height=None):
self.figure.figure_width = width
self.figure.figure_height = height
return
| set_fig_size | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, width, height=None):\n self.figure.figure_width = width\n self.figure.figure_height = height\n return\n",
"role": "user"
},
{
"content": "set_fig_size",
"role": "assistant"
}
] |
def x(self, space):
self.figure.spacing = space
if 'subplots_adjust_kwargs' not in self.figure.__dict__:
self.figure.subplots_adjust_kwargs = {}
if space == 'wide':
self.figure.subplots_adjust_kwargs['hspace'] = 0.3
self.figure.subplots_adjust_kwargs['wspace'] = 0.3
else:
self.figure.subplots_adjust_kwargs['hspace'] = 0.0
self.figure.subplots_adjust_kwargs['wspace'] = 0.0
return
| set_spacing | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, space):\n self.figure.spacing = space\n if 'subplots_adjust_kwargs' not in self.figure.__dict__:\n self.figure.subplots_adjust_kwargs = {}\n if space == 'wide':\n self.figure.subplots_adjust_kwargs['hspace'] = 0.3\n self.figure.subplots_adjust_kwargs['wspace'] = 0.3\n else:\n self.figure.subplots_adjust_kwargs['hspace'] = 0.0\n self.figure.subplots_adjust_kwargs['wspace'] = 0.0\n return\n",
"role": "user"
},
{
"content": "set_spacing",
"role": "assistant"
}
] |
def x(self, **kwargs):
prop_default = {'bottom': 0.1, 'top': 0.85, 'right': 0.9, 'left': 0.12,
'hspace': 0.3, 'wspace': 0.3}
if 'subplots_adjust_kwargs' in self.figure.__dict__:
for key, value in self.figure.subplots_adjust_kwargs.items():
prop_default[key] = value
for prop, default in prop_default.items():
kwargs[prop] = kwargs.get(prop, default)
self.figure.subplots_adjust_kwargs = kwargs
return
| subplots_adjust | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, **kwargs):\n prop_default = {'bottom': 0.1, 'top': 0.85, 'right': 0.9, 'left': 0.12,\n 'hspace': 0.3, 'wspace': 0.3}\n if 'subplots_adjust_kwargs' in self.figure.__dict__:\n for key, value in self.figure.subplots_adjust_kwargs.items():\n prop_default[key] = value\n for prop, default in prop_default.items():\n kwargs[prop] = kwargs.get(prop, default)\n self.figure.subplots_adjust_kwargs = kwargs\n return\n",
"role": "user"
},
{
"content": "subplots_adjust",
"role": "assistant"
}
] |
def x(self, xlabel, **kwargs):
prop_default = {'x': 0.01, 'y': 0.51, 'fontsize': 20, 'rotation':
'vertical', 'va': 'center'}
for prop, default in prop_default.items():
kwargs[prop] = kwargs.get(prop, default)
self._set_fig_label('x', xlabel, **kwargs)
return
| set_fig_x_label | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, xlabel, **kwargs):\n prop_default = {'x': 0.01, 'y': 0.51, 'fontsize': 20, 'rotation':\n 'vertical', 'va': 'center'}\n for prop, default in prop_default.items():\n kwargs[prop] = kwargs.get(prop, default)\n self._set_fig_label('x', xlabel, **kwargs)\n return\n",
"role": "user"
},
{
"content": "set_fig_x_label",
"role": "assistant"
}
] |
def x(self, ylabel, **kwargs):
prop_default = {'x': 0.45, 'y': 0.02, 'fontsize': 20, 'rotation':
'horizontal', 'ha': 'center'}
for prop, default in prop_default.items():
kwargs[prop] = kwargs.get(prop, default)
self._set_fig_label('y', ylabel, **kwargs)
return
| set_fig_y_label | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, ylabel, **kwargs):\n prop_default = {'x': 0.45, 'y': 0.02, 'fontsize': 20, 'rotation':\n 'horizontal', 'ha': 'center'}\n for prop, default in prop_default.items():\n kwargs[prop] = kwargs.get(prop, default)\n self._set_fig_label('y', ylabel, **kwargs)\n return\n",
"role": "user"
},
{
"content": "set_fig_y_label",
"role": "assistant"
}
] |
def x(self, title, **kwargs):
prop_default = {'fontsize': 20}
for prop, default in prop_default.items():
kwargs[prop] = kwargs.get(prop, default)
self.figure.fig_title = title
self.figure.fig_title_kwargs = kwargs
return
| set_fig_title | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, title, **kwargs):\n prop_default = {'fontsize': 20}\n for prop, default in prop_default.items():\n kwargs[prop] = kwargs.get(prop, default)\n self.figure.fig_title = title\n self.figure.fig_title_kwargs = kwargs\n return\n",
"role": "user"
},
{
"content": "set_fig_title",
"role": "assistant"
}
] |
def x(self, plot_type, **kwargs):
prop_default = {'cbar_label': None, 'cbar_ticks_fontsize': 15,
'cbar_label_fontsize': 20, 'cbar_axes': [], 'cbar_ticks': [],
'cbar_tick_labels': [], 'cbar_pos': 'use_default', 'cbar_label_pad':
None}
for prop, default in prop_default.items():
kwargs[prop] = kwargs.get(prop[5:], default)
if prop[5:] in kwargs:
del kwargs[prop[5:]]
if 'colorbars' not in self.figure.__dict__:
self.figure.colorbars = {}
self.figure.colorbars[plot_type] = kwargs
return
| set_colorbar | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, plot_type, **kwargs):\n prop_default = {'cbar_label': None, 'cbar_ticks_fontsize': 15,\n 'cbar_label_fontsize': 20, 'cbar_axes': [], 'cbar_ticks': [],\n 'cbar_tick_labels': [], 'cbar_pos': 'use_default', 'cbar_label_pad':\n None}\n for prop, default in prop_default.items():\n kwargs[prop] = kwargs.get(prop[5:], default)\n if prop[5:] in kwargs:\n del kwargs[prop[5:]]\n if 'colorbars' not in self.figure.__dict__:\n self.figure.colorbars = {}\n self.figure.colorbars[plot_type] = kwargs\n return\n",
"role": "user"
},
{
"content": "set_colorbar",
"role": "assistant"
}
] |
def x(self, xlabel=None, ylabel=None):
if xlabel is not None:
self.general.x_column_label = xlabel
if ylabel is not None:
self.general.y_column_label = ylabel
if xlabel is None and ylabel is None:
warnings.warn('is not specifying x or y lables even' +
'though column labels function is called.', UserWarning)
return
| set_all_file_column_labels | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, xlabel=None, ylabel=None):\n if xlabel is not None:\n self.general.x_column_label = xlabel\n if ylabel is not None:\n self.general.y_column_label = ylabel\n if xlabel is None and ylabel is None:\n warnings.warn('is not specifying x or y lables even' +\n 'though column labels function is called.', UserWarning)\n return\n",
"role": "user"
},
{
"content": "set_all_file_column_labels",
"role": "assistant"
}
] |
def x(self, which, lim, d, scale, fontsize=None):
setattr(self.general, which + 'lims', lim)
setattr(self.general, 'd' + which, d)
setattr(self.general, which + 'scale', scale)
if fontsize is not None:
setattr(self.general, which + '_tick_label_fontsize', fontsize)
return
| set_all_lims | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, which, lim, d, scale, fontsize=None):\n setattr(self.general, which + 'lims', lim)\n setattr(self.general, 'd' + which, d)\n setattr(self.general, which + 'scale', scale)\n if fontsize is not None:\n setattr(self.general, which + '_tick_label_fontsize', fontsize)\n return\n",
"role": "user"
},
{
"content": "set_all_lims",
"role": "assistant"
}
] |
def x(self, xlim, dx, xscale, fontsize=None):
self._set_all_lims('x', xlim, dx, xscale, fontsize)
return
| set_all_xlims | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, xlim, dx, xscale, fontsize=None):\n self._set_all_lims('x', xlim, dx, xscale, fontsize)\n return\n",
"role": "user"
},
{
"content": "set_all_xlims",
"role": "assistant"
}
] |
def x(self, ylim, dy, yscale, fontsize=None):
self._set_all_lims('y', ylim, dy, yscale, fontsize)
return
| set_all_ylims | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, ylim, dy, yscale, fontsize=None):\n self._set_all_lims('y', ylim, dy, yscale, fontsize)\n return\n",
"role": "user"
},
{
"content": "set_all_ylims",
"role": "assistant"
}
] |
def x(self, axis_to_reverse):
if axis_to_reverse.lower() == 'x':
self.general.reverse_x_axis = True
if axis_to_reverse.lower() == 'y':
self.general.reverse_y_axis = True
if axis_to_reverse.lower() != 'x' or axis_to_reverse.lower() != 'y':
raise ValueError('Axis for reversing needs to be either x or y.')
return
| reverse_axis | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, axis_to_reverse):\n if axis_to_reverse.lower() == 'x':\n self.general.reverse_x_axis = True\n if axis_to_reverse.lower() == 'y':\n self.general.reverse_y_axis = True\n if axis_to_reverse.lower() != 'x' or axis_to_reverse.lower() != 'y':\n raise ValueError('Axis for reversing needs to be either x or y.')\n return\n",
"role": "user"
},
{
"content": "reverse_axis",
"role": "assistant"
}
] |
def x(self):
output_dict = {}
output_dict['general'] = self._iterate_through_class(self.general.__dict__)
output_dict['figure'] = self._iterate_through_class(self.figure.__dict__)
if self.total_plots > 1:
trans_dict = {str(i): self._iterate_through_class(axis.__dict__) for
i, axis in enumerate(self.ax)}
output_dict['plot_info'] = trans_dict
else:
output_dict['plot_info'] = {'0': self._iterate_through_class(self.
ax.__dict__)}
if self.print_input:
print(output_dict)
return output_dict
| return_dict | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n output_dict = {}\n output_dict['general'] = self._iterate_through_class(self.general.__dict__)\n output_dict['figure'] = self._iterate_through_class(self.figure.__dict__)\n if self.total_plots > 1:\n trans_dict = {str(i): self._iterate_through_class(axis.__dict__) for\n i, axis in enumerate(self.ax)}\n output_dict['plot_info'] = trans_dict\n else:\n output_dict['plot_info'] = {'0': self._iterate_through_class(self.\n ax.__dict__)}\n if self.print_input:\n print(output_dict)\n return output_dict\n",
"role": "user"
},
{
"content": "return_dict",
"role": "assistant"
}
] |
def x(self, class_dict):
output_dict = {}
for key in class_dict:
val = class_dict[key]
try:
val = val.__dict__
except AttributeError:
pass
if type(val) is dict:
val = self._iterate_through_class(val)
if type(val) is list:
temp_val = []
for val_i in val:
try:
val_i = val_i.__dict__
except AttributeError:
pass
if type(val_i) is dict:
val_i = self._iterate_through_class(val_i)
temp_val.append(val_i)
val = temp_val
output_dict[key] = val
return output_dict
| iterate_through_class | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, class_dict):\n output_dict = {}\n for key in class_dict:\n val = class_dict[key]\n try:\n val = val.__dict__\n except AttributeError:\n pass\n if type(val) is dict:\n val = self._iterate_through_class(val)\n if type(val) is list:\n temp_val = []\n for val_i in val:\n try:\n val_i = val_i.__dict__\n except AttributeError:\n pass\n if type(val_i) is dict:\n val_i = self._iterate_through_class(val_i)\n temp_val.append(val_i)\n val = temp_val\n output_dict[key] = val\n return output_dict\n",
"role": "user"
},
{
"content": "iterate_through_class",
"role": "assistant"
}
] |
def x(self):
data = ascii.read(self.WORKING_DIRECTORY + '/' + self.file_name)
num_x_pts = len(np.unique(data[self.x_column_label]))
num_y_pts = len(np.unique(data[self.y_column_label]))
self.xvals = np.reshape(np.asarray(data[self.x_column_label]), (
num_y_pts, num_x_pts))
self.yvals = np.reshape(np.asarray(data[self.y_column_label]), (
num_y_pts, num_x_pts))
self.zvals = np.reshape(np.asarray(data[self.z_column_label]), (
num_y_pts, num_x_pts))
return
| txt_read_in | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n data = ascii.read(self.WORKING_DIRECTORY + '/' + self.file_name)\n num_x_pts = len(np.unique(data[self.x_column_label]))\n num_y_pts = len(np.unique(data[self.y_column_label]))\n self.xvals = np.reshape(np.asarray(data[self.x_column_label]), (\n num_y_pts, num_x_pts))\n self.yvals = np.reshape(np.asarray(data[self.y_column_label]), (\n num_y_pts, num_x_pts))\n self.zvals = np.reshape(np.asarray(data[self.z_column_label]), (\n num_y_pts, num_x_pts))\n return\n",
"role": "user"
},
{
"content": "txt_read_in",
"role": "assistant"
}
] |
def x(self):
with h5py.File(self.WORKING_DIRECTORY + '/' + self.file_name) as f:
data = f['data']
num_x_pts = len(np.unique(data[self.x_column_label][:]))
num_y_pts = len(np.unique(data[self.y_column_label][:]))
self.xvals = np.reshape(data[self.x_column_label][:], (num_y_pts,
num_x_pts))
self.yvals = np.reshape(data[self.y_column_label][:], (num_y_pts,
num_x_pts))
self.zvals = np.reshape(data[self.z_column_label][:], (num_y_pts,
num_x_pts))
return
| hdf5_read_in | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n with h5py.File(self.WORKING_DIRECTORY + '/' + self.file_name) as f:\n data = f['data']\n num_x_pts = len(np.unique(data[self.x_column_label][:]))\n num_y_pts = len(np.unique(data[self.y_column_label][:]))\n self.xvals = np.reshape(data[self.x_column_label][:], (num_y_pts,\n num_x_pts))\n self.yvals = np.reshape(data[self.y_column_label][:], (num_y_pts,\n num_x_pts))\n self.zvals = np.reshape(data[self.z_column_label][:], (num_y_pts,\n num_x_pts))\n return\n",
"role": "user"
},
{
"content": "hdf5_read_in",
"role": "assistant"
}
] |
def x(self):
if self.xscale != 'lin':
self.xvals = np.logspace(np.log10(float(self.x_low)), np.log10(
float(self.x_high)), self.num_x)
else:
self.xvals = np.linspace(float(self.x_low), float(self.x_high),
self.num_x)
if self.yscale != 'lin':
self.yvals = np.logspace(np.log10(float(self.y_low)), np.log10(
float(self.y_high)), self.num_y)
else:
self.yvals = np.linspace(float(self.y_low), float(self.y_high),
self.num_y)
self.xvals, self.yvals = np.meshgrid(self.xvals, self.yvals)
self.xvals, self.yvals = self.xvals.ravel(), self.yvals.ravel()
for which in ['x', 'y']:
setattr(self, getattr(self, which + 'val_name'), getattr(self,
which + 'vals'))
self.ecc = 'eccentricity' in self.__dict__
if self.ecc:
if 'observation_time' not in self.__dict__:
if 'start_time' not in self.__dict__:
raise ValueError(
'If no observation time is provided, the time before' +
'merger must be the inital starting condition.')
self.observation_time = self.start_time
elif 'spin' in self.__dict__:
self.spin_1 = self.spin
self.spin_2 = self.spin
for key in ['redshift', 'luminosity_distance', 'comoving_distance']:
if key in self.__dict__:
self.dist_type = key
self.z_or_dist = getattr(self, key)
if self.ecc:
for key in ['start_frequency', 'start_time', 'start_separation']:
if key in self.__dict__:
self.initial_cond_type = key.split('_')[-1]
self.initial_point = getattr(self, key)
self.m1 = self.total_mass / (1.0 + self.mass_ratio)
self.m2 = self.total_mass * self.mass_ratio / (1.0 + self.mass_ratio)
return
| set_parameters | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n if self.xscale != 'lin':\n self.xvals = np.logspace(np.log10(float(self.x_low)), np.log10(\n float(self.x_high)), self.num_x)\n else:\n self.xvals = np.linspace(float(self.x_low), float(self.x_high),\n self.num_x)\n if self.yscale != 'lin':\n self.yvals = np.logspace(np.log10(float(self.y_low)), np.log10(\n float(self.y_high)), self.num_y)\n else:\n self.yvals = np.linspace(float(self.y_low), float(self.y_high),\n self.num_y)\n self.xvals, self.yvals = np.meshgrid(self.xvals, self.yvals)\n self.xvals, self.yvals = self.xvals.ravel(), self.yvals.ravel()\n for which in ['x', 'y']:\n setattr(self, getattr(self, which + 'val_name'), getattr(self, \n which + 'vals'))\n self.ecc = 'eccentricity' in self.__dict__\n if self.ecc:\n if 'observation_time' not in self.__dict__:\n if 'start_time' not in self.__dict__:\n raise ValueError(\n 'If no observation time is provided, the time before' +\n 'merger must be the inital starting condition.')\n self.observation_time = self.start_time\n elif 'spin' in self.__dict__:\n self.spin_1 = self.spin\n self.spin_2 = self.spin\n for key in ['redshift', 'luminosity_distance', 'comoving_distance']:\n if key in self.__dict__:\n self.dist_type = key\n self.z_or_dist = getattr(self, key)\n if self.ecc:\n for key in ['start_frequency', 'start_time', 'start_separation']:\n if key in self.__dict__:\n self.initial_cond_type = key.split('_')[-1]\n self.initial_point = getattr(self, key)\n self.m1 = self.total_mass / (1.0 + self.mass_ratio)\n self.m2 = self.total_mass * self.mass_ratio / (1.0 + self.mass_ratio)\n return\n",
"role": "user"
},
{
"content": "set_parameters",
"role": "assistant"
}
] |
def x(self):
if self.ecc:
required_kwargs = {'dist_type': self.dist_type, 'initial_cond_type':
self.initial_cond_type, 'ecc': True}
input_args = [self.m1, self.m2, self.z_or_dist, self.initial_point,
self.eccentricity, self.observation_time]
else:
required_kwargs = {'dist_type': self.dist_type}
input_args = [self.m1, self.m2, self.spin_1, self.spin_2, self.
z_or_dist, self.start_time, self.end_time]
input_kwargs = {**required_kwargs, **self.general, **self.
sensitivity_input, **self.snr_input, **self.parallel_input}
self.final_dict = snr(*input_args, **input_kwargs)
return
| run_snr | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n if self.ecc:\n required_kwargs = {'dist_type': self.dist_type, 'initial_cond_type':\n self.initial_cond_type, 'ecc': True}\n input_args = [self.m1, self.m2, self.z_or_dist, self.initial_point,\n self.eccentricity, self.observation_time]\n else:\n required_kwargs = {'dist_type': self.dist_type}\n input_args = [self.m1, self.m2, self.spin_1, self.spin_2, self.\n z_or_dist, self.start_time, self.end_time]\n input_kwargs = {**required_kwargs, **self.general, **self.\n sensitivity_input, **self.snr_input, **self.parallel_input}\n self.final_dict = snr(*input_args, **input_kwargs)\n return\n",
"role": "user"
},
{
"content": "run_snr",
"role": "assistant"
}
] |
def x(self, date, time, pid, commerce_id, transaction_id, request_ip, token,
webpay_server):
raise NotImplementedError('Logging Handler must implement event_payment')
| event_payment | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, date, time, pid, commerce_id, transaction_id, request_ip, token,\n webpay_server):\n raise NotImplementedError('Logging Handler must implement event_payment')\n",
"role": "user"
},
{
"content": "event_payment",
"role": "assistant"
}
] |
def x(self, date, time, pid, commerce_id, transaction_id, request_ip, order_id
):
raise NotImplementedError(
'Logging Handler must implement event_confirmation')
| event_confirmation | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, date, time, pid, commerce_id, transaction_id, request_ip, order_id\n ):\n raise NotImplementedError(\n 'Logging Handler must implement event_confirmation')\n",
"role": "user"
},
{
"content": "event_confirmation",
"role": "assistant"
}
] |
def x(self):
res = self.send_request('manager/stop', post=True)
if res.status_code != 200:
raise UnexpectedResponse(
'Attempted to stop manager. {res_code}: {res_text}'.format(
res_code=res.status_code, res_text=res.text))
if settings.VERBOSITY >= verbosity.PROCESS_STOP:
print('Stopped {}'.format(self.get_name()))
time.sleep(0.05)
| stop | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n res = self.send_request('manager/stop', post=True)\n if res.status_code != 200:\n raise UnexpectedResponse(\n 'Attempted to stop manager. {res_code}: {res_text}'.format(\n res_code=res.status_code, res_text=res.text))\n if settings.VERBOSITY >= verbosity.PROCESS_STOP:\n print('Stopped {}'.format(self.get_name()))\n time.sleep(0.05)\n",
"role": "user"
},
{
"content": "stop",
"role": "assistant"
}
] |
def x(self, config_file):
res = self.send_json_request('host/stop', data={'config': config_file})
if res.status_code != 200:
raise UnexpectedResponse(
'Attempted to stop a JSHost. Response: {res_code}: {res_text}'.
format(res_code=res.status_code, res_text=res.text))
return res.json()
| stop_host | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, config_file):\n res = self.send_json_request('host/stop', data={'config': config_file})\n if res.status_code != 200:\n raise UnexpectedResponse(\n 'Attempted to stop a JSHost. Response: {res_code}: {res_text}'.\n format(res_code=res.status_code, res_text=res.text))\n return res.json()\n",
"role": "user"
},
{
"content": "stop_host",
"role": "assistant"
}
] |
def x(self, params={}):
if self and hasattr(self, 'proxies') and self.proxies is not None:
response = requests.request('GET', url=PVWatts.PVWATTS_QUERY_URL,
params=params, headers={'User-Agent': ''.join(['pypvwatts/',
VERSION, ' (Python)'])}, proxies=self.proxies)
else:
response = requests.request('GET', url=PVWatts.PVWATTS_QUERY_URL,
params=params, headers={'User-Agent': ''.join(['pypvwatts/',
VERSION, ' (Python)'])})
if response.status_code == 403:
raise PVWattsError('Forbidden, 403')
return response.json()
| get_data | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, params={}):\n if self and hasattr(self, 'proxies') and self.proxies is not None:\n response = requests.request('GET', url=PVWatts.PVWATTS_QUERY_URL,\n params=params, headers={'User-Agent': ''.join(['pypvwatts/',\n VERSION, ' (Python)'])}, proxies=self.proxies)\n else:\n response = requests.request('GET', url=PVWatts.PVWATTS_QUERY_URL,\n params=params, headers={'User-Agent': ''.join(['pypvwatts/',\n VERSION, ' (Python)'])})\n if response.status_code == 403:\n raise PVWattsError('Forbidden, 403')\n return response.json()\n",
"role": "user"
},
{
"content": "get_data",
"role": "assistant"
}
] |
def x(self, input):
if IRichInput.providedBy(input):
richInput = unicode(input)
symbolInput = unicode(input.symbol())
else:
richInput = None
symbolInput = unicode(input)
action = LOG_FSM_TRANSITION(self.logger, fsm_identifier=self.identifier,
fsm_state=unicode(self.state), fsm_rich_input=richInput, fsm_input=
symbolInput)
with action as theAction:
output = super(FiniteStateLogger, self).receive(input)
theAction.addSuccessFields(fsm_next_state=unicode(self.state),
fsm_output=[unicode(o) for o in output])
if self._action is not None and self._isTerminal(self.state):
self._action.addSuccessFields(fsm_terminal_state=unicode(self.state))
self._action.finish()
self._action = None
return output
| receive | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, input):\n if IRichInput.providedBy(input):\n richInput = unicode(input)\n symbolInput = unicode(input.symbol())\n else:\n richInput = None\n symbolInput = unicode(input)\n action = LOG_FSM_TRANSITION(self.logger, fsm_identifier=self.identifier,\n fsm_state=unicode(self.state), fsm_rich_input=richInput, fsm_input=\n symbolInput)\n with action as theAction:\n output = super(FiniteStateLogger, self).receive(input)\n theAction.addSuccessFields(fsm_next_state=unicode(self.state),\n fsm_output=[unicode(o) for o in output])\n if self._action is not None and self._isTerminal(self.state):\n self._action.addSuccessFields(fsm_terminal_state=unicode(self.state))\n self._action.finish()\n self._action = None\n return output\n",
"role": "user"
},
{
"content": "receive",
"role": "assistant"
}
] |
def x(self, app):
prefix_len = len(self.prefix)
for varname, value in os.environ.items():
if not varname.startswith(self.prefix):
continue
varname = varname[prefix_len:]
value = value or app.config.get(varname)
try:
value = ast.literal_eval(value)
except (SyntaxError, ValueError):
pass
app.config[varname] = value
| init_app | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, app):\n prefix_len = len(self.prefix)\n for varname, value in os.environ.items():\n if not varname.startswith(self.prefix):\n continue\n varname = varname[prefix_len:]\n value = value or app.config.get(varname)\n try:\n value = ast.literal_eval(value)\n except (SyntaxError, ValueError):\n pass\n app.config[varname] = value\n",
"role": "user"
},
{
"content": "init_app",
"role": "assistant"
}
] |
def x(X, metric='l2', sparsity_param=10, kmax=None, keep_ratio=0.95):
n = X.shape[0]
if kmax is None:
kmax = min(n - 1, max(5, n // 10))
nn_dists, nn_inds = nearest_neighbors(X, metric=metric, k=kmax + 1,
return_dists=True)
W = np.zeros((n, n))
for i, pt in enumerate(X):
nbr_inds = nn_inds[i]
mask = nbr_inds != i
nbr_inds = nbr_inds[mask]
nbr_dist = nn_dists[i, mask]
Y = (X[nbr_inds] - pt) / nbr_dist[:, None]
c = _solve_admm(Y, nbr_dist / nbr_dist.sum(), sparsity_param)
c = np.abs(c / nbr_dist)
W[i, nbr_inds] = c / c.sum()
W = ss.csr_matrix(W)
if keep_ratio < 1:
for i in range(n):
row_data = W.data[W.indptr[i]:W.indptr[i + 1]]
order = np.argsort(row_data)[::-1]
stop_idx = np.searchsorted(np.cumsum(row_data[order]), keep_ratio
) + 1
bad_inds = order[stop_idx:]
row_data[bad_inds] = 0
W.eliminate_zeros()
return Graph.from_adj_matrix(W)
| smce_graph | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(X, metric='l2', sparsity_param=10, kmax=None, keep_ratio=0.95):\n n = X.shape[0]\n if kmax is None:\n kmax = min(n - 1, max(5, n // 10))\n nn_dists, nn_inds = nearest_neighbors(X, metric=metric, k=kmax + 1,\n return_dists=True)\n W = np.zeros((n, n))\n for i, pt in enumerate(X):\n nbr_inds = nn_inds[i]\n mask = nbr_inds != i\n nbr_inds = nbr_inds[mask]\n nbr_dist = nn_dists[i, mask]\n Y = (X[nbr_inds] - pt) / nbr_dist[:, None]\n c = _solve_admm(Y, nbr_dist / nbr_dist.sum(), sparsity_param)\n c = np.abs(c / nbr_dist)\n W[i, nbr_inds] = c / c.sum()\n W = ss.csr_matrix(W)\n if keep_ratio < 1:\n for i in range(n):\n row_data = W.data[W.indptr[i]:W.indptr[i + 1]]\n order = np.argsort(row_data)[::-1]\n stop_idx = np.searchsorted(np.cumsum(row_data[order]), keep_ratio\n ) + 1\n bad_inds = order[stop_idx:]\n row_data[bad_inds] = 0\n W.eliminate_zeros()\n return Graph.from_adj_matrix(W)\n",
"role": "user"
},
{
"content": "smce_graph",
"role": "assistant"
}
] |
def x(X, positive=False, sparsity_param=None, kmax=None):
clf, X = _l1_graph_setup(X, positive, sparsity_param)
if kmax is None:
W = _l1_graph_solve_full(clf, X)
else:
W = _l1_graph_solve_k(clf, X, kmax)
return Graph.from_adj_matrix(W)
| sparse_regularized_graph | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(X, positive=False, sparsity_param=None, kmax=None):\n clf, X = _l1_graph_setup(X, positive, sparsity_param)\n if kmax is None:\n W = _l1_graph_solve_full(clf, X)\n else:\n W = _l1_graph_solve_k(clf, X, kmax)\n return Graph.from_adj_matrix(W)\n",
"role": "user"
},
{
"content": "sparse_regularized_graph",
"role": "assistant"
}
] |
def x(self, binary_args, other_args):
if self.length < 100:
raise Exception(
'Run this across 1 processor by setting num_processors kwarg to None.'
)
if self.num_processors == -1:
self.num_processors = mp.cpu_count()
split_val = int(np.ceil(self.length / self.num_splits))
split_inds = [(self.num_splits * i) for i in np.arange(1, split_val)]
inds_split_all = np.split(np.arange(self.length), split_inds)
self.args = []
for i, ind_split in enumerate(inds_split_all):
trans_args = []
for arg in binary_args:
try:
trans_args.append(arg[ind_split])
except TypeError:
trans_args.append(arg)
self.args.append((i, tuple(trans_args)) + other_args)
return
| prep_parallel | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, binary_args, other_args):\n if self.length < 100:\n raise Exception(\n 'Run this across 1 processor by setting num_processors kwarg to None.'\n )\n if self.num_processors == -1:\n self.num_processors = mp.cpu_count()\n split_val = int(np.ceil(self.length / self.num_splits))\n split_inds = [(self.num_splits * i) for i in np.arange(1, split_val)]\n inds_split_all = np.split(np.arange(self.length), split_inds)\n self.args = []\n for i, ind_split in enumerate(inds_split_all):\n trans_args = []\n for arg in binary_args:\n try:\n trans_args.append(arg[ind_split])\n except TypeError:\n trans_args.append(arg)\n self.args.append((i, tuple(trans_args)) + other_args)\n return\n",
"role": "user"
},
{
"content": "prep_parallel",
"role": "assistant"
}
] |
def x(self, para_func):
if self.timer:
start_timer = time.time()
with mp.Pool(self.num_processors) as pool:
print('start pool with {} processors: {} total processes.\n'.format
(self.num_processors, len(self.args)))
results = [pool.apply_async(para_func, arg) for arg in self.args]
out = [r.get() for r in results]
out = {key: np.concatenate([out_i[key] for out_i in out]) for key in
out[0].keys()}
if self.timer:
print('SNR calculation time:', time.time() - start_timer)
return out
| run_parallel | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, para_func):\n if self.timer:\n start_timer = time.time()\n with mp.Pool(self.num_processors) as pool:\n print('start pool with {} processors: {} total processes.\\n'.format\n (self.num_processors, len(self.args)))\n results = [pool.apply_async(para_func, arg) for arg in self.args]\n out = [r.get() for r in results]\n out = {key: np.concatenate([out_i[key] for out_i in out]) for key in\n out[0].keys()}\n if self.timer:\n print('SNR calculation time:', time.time() - start_timer)\n return out\n",
"role": "user"
},
{
"content": "run_parallel",
"role": "assistant"
}
] |
def x(parser_name, parse_callable, *unparse_callables):
def x(f):
@wraps(f)
def x(self, source, *a, **kw):
node = parse_callable(source)
return f(node, *a, **kw)
unparse.__doc__ = f.__doc__.replace("""ast
The AST """,
"""source
The source """)
return unparse
def x(f):
@wraps(f)
def x(self, source):
return f(source)
parse.__name__ = parser_name
parse.__qualname__ = parser_name
return parse
callables = {f.__name__: build_unparse(f) for f in unparse_callables}
callables['__call__'] = build_parse(parse_callable)
callables['__module__'] = PKGNAME
return type(parser_name, (object,), callables)()
| RawParserUnparserFactory | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(parser_name, parse_callable, *unparse_callables):\n\n def x(f):\n\n @wraps(f)\n def x(self, source, *a, **kw):\n node = parse_callable(source)\n return f(node, *a, **kw)\n unparse.__doc__ = f.__doc__.replace(\"\"\"ast\n The AST \"\"\",\n \"\"\"source\n The source \"\"\")\n return unparse\n\n def x(f):\n\n @wraps(f)\n def x(self, source):\n return f(source)\n parse.__name__ = parser_name\n parse.__qualname__ = parser_name\n return parse\n callables = {f.__name__: build_unparse(f) for f in unparse_callables}\n callables['__call__'] = build_parse(parse_callable)\n callables['__module__'] = PKGNAME\n return type(parser_name, (object,), callables)()\n",
"role": "user"
},
{
"content": "RawParserUnparserFactory",
"role": "assistant"
}
] |
def x(module_name, *unparser_names):
parse_callable = import_module(PKGNAME + '.parsers.' + module_name).parse
unparser_module = import_module(PKGNAME + '.unparsers.' + module_name)
return RawParserUnparserFactory(module_name, parse_callable, *[getattr(
unparser_module, name) for name in unparser_names])
| ParserUnparserFactory | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(module_name, *unparser_names):\n parse_callable = import_module(PKGNAME + '.parsers.' + module_name).parse\n unparser_module = import_module(PKGNAME + '.unparsers.' + module_name)\n return RawParserUnparserFactory(module_name, parse_callable, *[getattr(\n unparser_module, name) for name in unparser_names])\n",
"role": "user"
},
{
"content": "ParserUnparserFactory",
"role": "assistant"
}
] |
def x(trajectories, downsampler, *args, **kwargs):
X = np.vstack(trajectories)
traj_lengths = list(map(len, trajectories))
inds = np.sort(downsampler(X, *args, **kwargs))
new_traj = []
for stop in np.cumsum(traj_lengths):
n = np.searchsorted(inds, stop)
new_traj.append(X[inds[:n]])
inds = inds[n:]
return new_traj
| downsample_trajectories | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(trajectories, downsampler, *args, **kwargs):\n X = np.vstack(trajectories)\n traj_lengths = list(map(len, trajectories))\n inds = np.sort(downsampler(X, *args, **kwargs))\n new_traj = []\n for stop in np.cumsum(traj_lengths):\n n = np.searchsorted(inds, stop)\n new_traj.append(X[inds[:n]])\n inds = inds[n:]\n return new_traj\n",
"role": "user"
},
{
"content": "downsample_trajectories",
"role": "assistant"
}
] |
def x(points, close_distance):
num_points = points.shape[0]
indices = set(range(num_points))
selected = []
while indices:
idx = indices.pop()
nn_inds, = nearest_neighbors(points[idx], points, epsilon=
close_distance)
indices.difference_update(nn_inds)
selected.append(idx)
return selected
| epsilon_net | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(points, close_distance):\n num_points = points.shape[0]\n indices = set(range(num_points))\n selected = []\n while indices:\n idx = indices.pop()\n nn_inds, = nearest_neighbors(points[idx], points, epsilon=\n close_distance)\n indices.difference_update(nn_inds)\n selected.append(idx)\n return selected\n",
"role": "user"
},
{
"content": "epsilon_net",
"role": "assistant"
}
] |
def x(points, num_centers, m=2.0, tol=0.0001, max_iter=100, verbose=False):
num_points = points.shape[0]
if num_centers >= num_points:
return np.arange(num_points)
assn = np.random.random((points.shape[0], num_centers))
for i in range(max_iter):
w = assn ** m
w /= w.sum(axis=0)
centers = w.T.dot(points)
d = pairwise_distances(points, centers)
d **= 2.0 / (m - 1)
np.maximum(d, 1e-10, out=d)
new_assn = 1.0 / np.einsum('ik,ij->ik', d, 1.0 / d)
change = np.linalg.norm(new_assn - assn)
if verbose:
print('At iteration %d: change = %g' % (i + 1, change))
if change < tol:
break
assn = new_assn
else:
warnings.warn("fuzzy_c_means didn't converge in %d iterations" %
max_iter)
return d.argmin(axis=0)
| fuzzy_c_means | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(points, num_centers, m=2.0, tol=0.0001, max_iter=100, verbose=False):\n num_points = points.shape[0]\n if num_centers >= num_points:\n return np.arange(num_points)\n assn = np.random.random((points.shape[0], num_centers))\n for i in range(max_iter):\n w = assn ** m\n w /= w.sum(axis=0)\n centers = w.T.dot(points)\n d = pairwise_distances(points, centers)\n d **= 2.0 / (m - 1)\n np.maximum(d, 1e-10, out=d)\n new_assn = 1.0 / np.einsum('ik,ij->ik', d, 1.0 / d)\n change = np.linalg.norm(new_assn - assn)\n if verbose:\n print('At iteration %d: change = %g' % (i + 1, change))\n if change < tol:\n break\n assn = new_assn\n else:\n warnings.warn(\"fuzzy_c_means didn't converge in %d iterations\" %\n max_iter)\n return d.argmin(axis=0)\n",
"role": "user"
},
{
"content": "fuzzy_c_means",
"role": "assistant"
}
] |
def x(url):
if '/' not in url and ':' in url and url.rsplit(':')[-1].isdigit():
url = 'scgi://' + url
url = urlparse.urlsplit(url, scheme='scgi', allow_fragments=False)
try:
transport = TRANSPORTS[url.scheme.lower()]
except KeyError:
if not any((url.netloc, url.query)) and url.path.isdigit():
return transport_from_url('scgi://%s:%s' % (url.scheme, url.path))
else:
raise URLError('Unsupported scheme in URL %r' % url.geturl())
else:
return transport(url)
| transport_from_url | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(url):\n if '/' not in url and ':' in url and url.rsplit(':')[-1].isdigit():\n url = 'scgi://' + url\n url = urlparse.urlsplit(url, scheme='scgi', allow_fragments=False)\n try:\n transport = TRANSPORTS[url.scheme.lower()]\n except KeyError:\n if not any((url.netloc, url.query)) and url.path.isdigit():\n return transport_from_url('scgi://%s:%s' % (url.scheme, url.path))\n else:\n raise URLError('Unsupported scheme in URL %r' % url.geturl())\n else:\n return transport(url)\n",
"role": "user"
},
{
"content": "transport_from_url",
"role": "assistant"
}
] |
def x(data, headers=None):
prolog = 'CONTENT_LENGTH\x00%d\x00SCGI\x001\x00' % len(data)
if headers:
prolog += _encode_headers(headers)
return _encode_netstring(prolog) + data
| encode_payload | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(data, headers=None):\n prolog = 'CONTENT_LENGTH\\x00%d\\x00SCGI\\x001\\x00' % len(data)\n if headers:\n prolog += _encode_headers(headers)\n return _encode_netstring(prolog) + data\n",
"role": "user"
},
{
"content": "encode_payload",
"role": "assistant"
}
] |
def x(headers):
try:
return dict(line.rstrip().split(': ', 1) for line in headers.
splitlines() if line)
except (TypeError, ValueError) as exc:
raise SCGIException('Error in SCGI headers %r (%s)' % (headers, exc))
| parse_headers | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(headers):\n try:\n return dict(line.rstrip().split(': ', 1) for line in headers.\n splitlines() if line)\n except (TypeError, ValueError) as exc:\n raise SCGIException('Error in SCGI headers %r (%s)' % (headers, exc))\n",
"role": "user"
},
{
"content": "parse_headers",
"role": "assistant"
}
] |
def x(resp):
try:
headers, payload = resp.split('\r\n\r\n', 1)
except (TypeError, ValueError) as exc:
raise SCGIException(
'No header delimiter in SCGI response of length %d (%s)' % (len
(resp), exc))
headers = _parse_headers(headers)
clen = headers.get('Content-Length')
if clen is not None:
assert len(payload) == int(clen)
return payload, headers
| parse_response | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(resp):\n try:\n headers, payload = resp.split('\\r\\n\\r\\n', 1)\n except (TypeError, ValueError) as exc:\n raise SCGIException(\n 'No header delimiter in SCGI response of length %d (%s)' % (len\n (resp), exc))\n headers = _parse_headers(headers)\n clen = headers.get('Content-Length')\n if clen is not None:\n assert len(payload) == int(clen)\n return payload, headers\n",
"role": "user"
},
{
"content": "parse_response",
"role": "assistant"
}
] |
def x(url, methodname, *params, **kw):
xmlreq = xmlrpclib.dumps(params, methodname)
xmlresp = SCGIRequest(url).send(xmlreq)
if kw.get('deserialize', True):
xmlresp = xmlresp.replace('<i8>', '<i4>').replace('</i8>', '</i4>')
return xmlrpclib.loads(xmlresp)[0][0]
else:
return xmlresp
| scgi_request | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(url, methodname, *params, **kw):\n xmlreq = xmlrpclib.dumps(params, methodname)\n xmlresp = SCGIRequest(url).send(xmlreq)\n if kw.get('deserialize', True):\n xmlresp = xmlresp.replace('<i8>', '<i4>').replace('</i8>', '</i4>')\n return xmlrpclib.loads(xmlresp)[0][0]\n else:\n return xmlresp\n",
"role": "user"
},
{
"content": "scgi_request",
"role": "assistant"
}
] |
def x(self, data):
sock = socket.socket(*self.sock_args)
try:
sock.connect(self.sock_addr)
except socket.error as exc:
raise socket.error("Can't connect to %r (%s)" % (self.url.geturl(),
exc))
try:
sock.send(data)
while True:
chunk = sock.recv(self.CHUNK_SIZE)
if chunk:
yield chunk
else:
break
finally:
sock.close()
| send | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, data):\n sock = socket.socket(*self.sock_args)\n try:\n sock.connect(self.sock_addr)\n except socket.error as exc:\n raise socket.error(\"Can't connect to %r (%s)\" % (self.url.geturl(),\n exc))\n try:\n sock.send(data)\n while True:\n chunk = sock.recv(self.CHUNK_SIZE)\n if chunk:\n yield chunk\n else:\n break\n finally:\n sock.close()\n",
"role": "user"
},
{
"content": "send",
"role": "assistant"
}
] |
def x(self, data):
try:
proc = subprocess.Popen(self.cmd, stdin=subprocess.PIPE, stdout=
subprocess.PIPE, stderr=subprocess.PIPE)
except OSError as exc:
raise URLError('Calling %r failed (%s)!' % (' '.join(self.cmd), exc))
else:
stdout, stderr = proc.communicate(data)
if proc.returncode:
raise URLError('Calling %r failed with RC=%d!\n%s' % (' '.join(
self.cmd), proc.returncode, stderr))
yield stdout
| send | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, data):\n try:\n proc = subprocess.Popen(self.cmd, stdin=subprocess.PIPE, stdout=\n subprocess.PIPE, stderr=subprocess.PIPE)\n except OSError as exc:\n raise URLError('Calling %r failed (%s)!' % (' '.join(self.cmd), exc))\n else:\n stdout, stderr = proc.communicate(data)\n if proc.returncode:\n raise URLError('Calling %r failed with RC=%d!\\n%s' % (' '.join(\n self.cmd), proc.returncode, stderr))\n yield stdout\n",
"role": "user"
},
{
"content": "send",
"role": "assistant"
}
] |
def x(self, data):
start = time.time()
try:
scgi_resp = ''.join(self.transport.send(_encode_payload(data)))
finally:
self.latency = time.time() - start
resp, self.resp_headers = _parse_response(scgi_resp)
return resp
| send | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, data):\n start = time.time()\n try:\n scgi_resp = ''.join(self.transport.send(_encode_payload(data)))\n finally:\n self.latency = time.time() - start\n resp, self.resp_headers = _parse_response(scgi_resp)\n return resp\n",
"role": "user"
},
{
"content": "send",
"role": "assistant"
}
] |
def x(traceback):
frame = traceback
while frame.tb_next:
frame = frame.tb_next
yield frame.tb_frame
return
| frames | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(traceback):\n frame = traceback\n while frame.tb_next:\n frame = frame.tb_next\n yield frame.tb_frame\n return\n",
"role": "user"
},
{
"content": "frames",
"role": "assistant"
}
] |
def x(self, name, traceback):
foundMethod = False
for frame in self._frames(traceback):
this = frame.f_locals.get('self')
if this is self and frame.f_code.co_name == name:
foundMethod = True
break
return foundMethod
| methodInTraceback | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, name, traceback):\n foundMethod = False\n for frame in self._frames(traceback):\n this = frame.f_locals.get('self')\n if this is self and frame.f_code.co_name == name:\n foundMethod = True\n break\n return foundMethod\n",
"role": "user"
},
{
"content": "methodInTraceback",
"role": "assistant"
}
] |
def x(self, callback, *args, **kwargs):
self.steps.append((callback, args, kwargs))
| addStep | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, callback, *args, **kwargs):\n self.steps.append((callback, args, kwargs))\n",
"role": "user"
},
{
"content": "addStep",
"role": "assistant"
}
] |
def x(self):
while self.steps:
callback, args, kwargs = self.steps.pop()
callback(*args, **kwargs)
| doRollback | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n while self.steps:\n callback, args, kwargs = self.steps.pop()\n callback(*args, **kwargs)\n",
"role": "user"
},
{
"content": "doRollback",
"role": "assistant"
}
] |
def x(X, metric='euclidean', k=None, epsilon=None, weighting='none',
precomputed=False):
if k is None and epsilon is None:
raise ValueError('Must provide `k` or `epsilon`.')
if weighting not in ('binary', 'none'):
raise ValueError('Invalid weighting param: %r' % weighting)
precomputed = precomputed or metric == 'precomputed'
binary = weighting == 'binary'
if not precomputed and epsilon is None:
return _sparse_neighbor_graph(X, k, binary, metric)
if precomputed:
D = X
else:
D = pairwise_distances(X, metric=metric)
return _slow_neighbor_graph(D, k, epsilon, binary)
| neighbor_graph | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(X, metric='euclidean', k=None, epsilon=None, weighting='none',\n precomputed=False):\n if k is None and epsilon is None:\n raise ValueError('Must provide `k` or `epsilon`.')\n if weighting not in ('binary', 'none'):\n raise ValueError('Invalid weighting param: %r' % weighting)\n precomputed = precomputed or metric == 'precomputed'\n binary = weighting == 'binary'\n if not precomputed and epsilon is None:\n return _sparse_neighbor_graph(X, k, binary, metric)\n if precomputed:\n D = X\n else:\n D = pairwise_distances(X, metric=metric)\n return _slow_neighbor_graph(D, k, epsilon, binary)\n",
"role": "user"
},
{
"content": "neighbor_graph",
"role": "assistant"
}
] |
def x(query_pts, target_pts=None, metric='euclidean', k=None, epsilon=None,
return_dists=False, precomputed=False):
if k is None and epsilon is None:
raise ValueError('Must provide `k` or `epsilon`.')
precomputed = precomputed or metric == 'precomputed'
if precomputed and target_pts is not None:
raise ValueError(
'`target_pts` cannot be used with precomputed distances')
query_pts = np.array(query_pts)
if len(query_pts.shape) == 1:
query_pts = query_pts.reshape((1, -1))
if precomputed:
dists = query_pts.copy()
else:
dists = pairwise_distances(query_pts, Y=target_pts, metric=metric)
if epsilon is not None:
if k is not None:
_, not_nn = _min_k_indices(dists, k, inv_ind=True)
dists[np.arange(dists.shape[0]), not_nn.T] = np.inf
is_close = dists <= epsilon
if return_dists:
nnis, nnds = [], []
for i, row in enumerate(is_close):
nns = np.nonzero(row)[0]
nnis.append(nns)
nnds.append(dists[i, nns])
return nnds, nnis
return np.array([np.nonzero(row)[0] for row in is_close])
nns = _min_k_indices(dists, k)
if return_dists:
row_inds = np.arange(len(nns))[:, np.newaxis]
nn_dists = dists[row_inds, nns]
return nn_dists, nns
return nns
| nearest_neighbors | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(query_pts, target_pts=None, metric='euclidean', k=None, epsilon=None,\n return_dists=False, precomputed=False):\n if k is None and epsilon is None:\n raise ValueError('Must provide `k` or `epsilon`.')\n precomputed = precomputed or metric == 'precomputed'\n if precomputed and target_pts is not None:\n raise ValueError(\n '`target_pts` cannot be used with precomputed distances')\n query_pts = np.array(query_pts)\n if len(query_pts.shape) == 1:\n query_pts = query_pts.reshape((1, -1))\n if precomputed:\n dists = query_pts.copy()\n else:\n dists = pairwise_distances(query_pts, Y=target_pts, metric=metric)\n if epsilon is not None:\n if k is not None:\n _, not_nn = _min_k_indices(dists, k, inv_ind=True)\n dists[np.arange(dists.shape[0]), not_nn.T] = np.inf\n is_close = dists <= epsilon\n if return_dists:\n nnis, nnds = [], []\n for i, row in enumerate(is_close):\n nns = np.nonzero(row)[0]\n nnis.append(nns)\n nnds.append(dists[i, nns])\n return nnds, nnis\n return np.array([np.nonzero(row)[0] for row in is_close])\n nns = _min_k_indices(dists, k)\n if return_dists:\n row_inds = np.arange(len(nns))[:, np.newaxis]\n nn_dists = dists[row_inds, nns]\n return nn_dists, nns\n return nns\n",
"role": "user"
},
{
"content": "nearest_neighbors",
"role": "assistant"
}
] |
def x(X, k, binary=False, metric='l2'):
knn = NearestNeighbors(n_neighbors=k, metric=metric).fit(X)
mode = 'connectivity' if binary else 'distance'
try:
adj = knn.kneighbors_graph(None, mode=mode)
except IndexError:
if binary:
adj = knn.kneighbors_graph(X, k + 1, mode=mode)
adj.setdiag(0)
else:
adj = knn.kneighbors_graph(X, k, mode=mode)
return Graph.from_adj_matrix(adj)
| sparse_neighbor_graph | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(X, k, binary=False, metric='l2'):\n knn = NearestNeighbors(n_neighbors=k, metric=metric).fit(X)\n mode = 'connectivity' if binary else 'distance'\n try:\n adj = knn.kneighbors_graph(None, mode=mode)\n except IndexError:\n if binary:\n adj = knn.kneighbors_graph(X, k + 1, mode=mode)\n adj.setdiag(0)\n else:\n adj = knn.kneighbors_graph(X, k, mode=mode)\n return Graph.from_adj_matrix(adj)\n",
"role": "user"
},
{
"content": "sparse_neighbor_graph",
"role": "assistant"
}
] |
def x(X, q=32, k=4, tangent_dim=1, curv_thresh=0.95, decay_rate=0.9,
max_iter=15, verbose=False):
n = len(X)
dist = pairwise_distances(X)
idx = np.argpartition(dist, q)[:, q]
r = np.median(dist[np.arange(n), idx])
W = neighbor_graph(dist, precomputed=True, epsilon=r).matrix('csr')
W.data[:] = 1 / W.data
normalize(W, norm='l1', axis=1, copy=False)
W = W.toarray()
prev_goodness = 1e-12
for it in range(max_iter):
goodness = 0
S = _estimate_tangent_spaces(X, W, tangent_dim)
for i, row in enumerate(W):
nbrs = row.nonzero()[-1]
edges = X[nbrs] - X[i]
edge_norms = (edges ** 2).sum(axis=1)
a1 = (edges.dot(S[i]) ** 2).sum(axis=1) / edge_norms
a2 = (np.einsum('ij,ijk->ik', edges, S[nbrs]) ** 2).sum(axis=1
) / edge_norms
a3 = _principal_angle(S[i], S[nbrs]) ** 2
x = np.minimum(curv_thresh, a1) * np.minimum(curv_thresh, a2
) * np.minimum(curv_thresh, a3)
excess = x.shape[0] - k
if excess > 0:
bad_idx = np.argpartition(x, excess - 1)[:excess]
W[i, nbrs[bad_idx]] *= decay_rate
W[i] /= W[i].sum()
goodness += x.dot(W[i, nbrs])
if verbose:
goodness /= n
print(it, goodness, goodness / prev_goodness)
if goodness / prev_goodness <= 1.0001:
break
prev_goodness = goodness
else:
warnings.warn('Failed to converge after %d iterations.' % max_iter)
indptr, indices, data = [0], [], []
for i, row in enumerate(W):
nbrs = row.nonzero()[-1]
if len(nbrs) > k:
nbrs = nbrs[np.argpartition(row[nbrs], len(nbrs) - k)[-k:]]
indices.extend(nbrs)
indptr.append(len(nbrs))
data.extend(dist[i, nbrs])
indptr = np.cumsum(indptr)
data = np.array(data)
indices = np.array(indices)
W = ss.csr_matrix((data, indices, indptr), shape=W.shape)
return Graph.from_adj_matrix(W)
| saffron | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(X, q=32, k=4, tangent_dim=1, curv_thresh=0.95, decay_rate=0.9,\n max_iter=15, verbose=False):\n n = len(X)\n dist = pairwise_distances(X)\n idx = np.argpartition(dist, q)[:, q]\n r = np.median(dist[np.arange(n), idx])\n W = neighbor_graph(dist, precomputed=True, epsilon=r).matrix('csr')\n W.data[:] = 1 / W.data\n normalize(W, norm='l1', axis=1, copy=False)\n W = W.toarray()\n prev_goodness = 1e-12\n for it in range(max_iter):\n goodness = 0\n S = _estimate_tangent_spaces(X, W, tangent_dim)\n for i, row in enumerate(W):\n nbrs = row.nonzero()[-1]\n edges = X[nbrs] - X[i]\n edge_norms = (edges ** 2).sum(axis=1)\n a1 = (edges.dot(S[i]) ** 2).sum(axis=1) / edge_norms\n a2 = (np.einsum('ij,ijk->ik', edges, S[nbrs]) ** 2).sum(axis=1\n ) / edge_norms\n a3 = _principal_angle(S[i], S[nbrs]) ** 2\n x = np.minimum(curv_thresh, a1) * np.minimum(curv_thresh, a2\n ) * np.minimum(curv_thresh, a3)\n excess = x.shape[0] - k\n if excess > 0:\n bad_idx = np.argpartition(x, excess - 1)[:excess]\n W[i, nbrs[bad_idx]] *= decay_rate\n W[i] /= W[i].sum()\n goodness += x.dot(W[i, nbrs])\n if verbose:\n goodness /= n\n print(it, goodness, goodness / prev_goodness)\n if goodness / prev_goodness <= 1.0001:\n break\n prev_goodness = goodness\n else:\n warnings.warn('Failed to converge after %d iterations.' % max_iter)\n indptr, indices, data = [0], [], []\n for i, row in enumerate(W):\n nbrs = row.nonzero()[-1]\n if len(nbrs) > k:\n nbrs = nbrs[np.argpartition(row[nbrs], len(nbrs) - k)[-k:]]\n indices.extend(nbrs)\n indptr.append(len(nbrs))\n data.extend(dist[i, nbrs])\n indptr = np.cumsum(indptr)\n data = np.array(data)\n indices = np.array(indices)\n W = ss.csr_matrix((data, indices, indptr), shape=W.shape)\n return Graph.from_adj_matrix(W)\n",
"role": "user"
},
{
"content": "saffron",
"role": "assistant"
}
] |
def x(a, B):
if a.shape[1] == 1:
return a.T.dot(B)[0, :, 0]
x1 = np.einsum('abc,adc->abd', B, B).dot(a) - a
x2 = np.einsum('ab,cad->cbd', a.dot(a.T), B) - B
xx = np.vstack((x1, x2))
xx -= xx.mean(axis=1)[:, None]
c = np.einsum('abc,abd->acd', xx, xx)
_, vecs = np.linalg.eigh(c)
fpc = vecs[:, :, -1]
fpc1 = fpc[:len(x1)]
fpc2 = fpc[len(x1):]
lhs = a.dot(fpc1.T).T
rhs = np.einsum('abc,ac->ab', B, fpc2)
return np.einsum('ij,ij->i', lhs, rhs)
| principal_angle | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(a, B):\n if a.shape[1] == 1:\n return a.T.dot(B)[0, :, 0]\n x1 = np.einsum('abc,adc->abd', B, B).dot(a) - a\n x2 = np.einsum('ab,cad->cbd', a.dot(a.T), B) - B\n xx = np.vstack((x1, x2))\n xx -= xx.mean(axis=1)[:, None]\n c = np.einsum('abc,abd->acd', xx, xx)\n _, vecs = np.linalg.eigh(c)\n fpc = vecs[:, :, -1]\n fpc1 = fpc[:len(x1)]\n fpc2 = fpc[len(x1):]\n lhs = a.dot(fpc1.T).T\n rhs = np.einsum('abc,ac->ab', B, fpc2)\n return np.einsum('ij,ij->i', lhs, rhs)\n",
"role": "user"
},
{
"content": "principal_angle",
"role": "assistant"
}
] |
def x(db, encoding, no_limit, zip, case_insensitive):
limit = 15
if no_limit:
limit = float('inf')
with open_db(db, zip, encoding=encoding, case_sensitive=not
case_insensitive) as dbf:
for idx, row in enumerate(dbf, 1):
click.secho('')
for key, val in row._asdict().items():
click.secho(' %s: %s' % (key, val))
if idx == limit:
click.secho(
'Note: Output is limited to %s rows. Use --no-limit option to bypass.'
% limit, fg='red')
break
| show | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(db, encoding, no_limit, zip, case_insensitive):\n limit = 15\n if no_limit:\n limit = float('inf')\n with open_db(db, zip, encoding=encoding, case_sensitive=not\n case_insensitive) as dbf:\n for idx, row in enumerate(dbf, 1):\n click.secho('')\n for key, val in row._asdict().items():\n click.secho(' %s: %s' % (key, val))\n if idx == limit:\n click.secho(\n 'Note: Output is limited to %s rows. Use --no-limit option to bypass.'\n % limit, fg='red')\n break\n",
"role": "user"
},
{
"content": "show",
"role": "assistant"
}
] |
def x(db, zip, case_insensitive):
with open_db(db, zip, case_sensitive=not case_insensitive) as dbf:
click.secho('Rows count: %s' % dbf.prolog.records_count)
click.secho('Fields:')
for field in dbf.fields:
click.secho(' %s: %s' % (field.type, field))
| describe | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(db, zip, case_insensitive):\n with open_db(db, zip, case_sensitive=not case_insensitive) as dbf:\n click.secho('Rows count: %s' % dbf.prolog.records_count)\n click.secho('Fields:')\n for field in dbf.fields:\n click.secho(' %s: %s' % (field.type, field))\n",
"role": "user"
},
{
"content": "describe",
"role": "assistant"
}
] |
def x(pkg, dest, repo, repo_dest, branch):
git(['clone', '--recursive', '-b', branch, repo, repo_dest])
| clone_repo | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(pkg, dest, repo, repo_dest, branch):\n git(['clone', '--recursive', '-b', branch, repo, repo_dest])\n",
"role": "user"
},
{
"content": "clone_repo",
"role": "assistant"
}
] |
def x(pkg, repo_dest):
vars = {'pkg': pkg}
with dir_path(repo_dest):
patch(
""" diff --git a/manage.py b/manage.py
index 40ebb0a..cdfe363 100755
--- a/manage.py
+++ b/manage.py
@@ -3,7 +3,7 @@ import os
import sys
# Edit this if necessary or override the variable in your environment.
-os.environ.setdefault('DJANGO_SETTINGS_MODULE', 'project.settings')
+os.environ.setdefault('DJANGO_SETTINGS_MODULE', '%(pkg)s.settings')
try:
# For local development in a virtualenv:
diff --git a/project/settings/base.py b/project/settings/base.py
index 312f280..c75e673 100644
--- a/project/settings/base.py
+++ b/project/settings/base.py
@@ -7,7 +7,7 @@ from funfactory.settings_base import *
# If you did not install Playdoh with the funfactory installer script
# you may need to edit this value. See the docs about installing from a
# clone.
-PROJECT_MODULE = 'project'
+PROJECT_MODULE = '%(pkg)s'
# Bundles is a dictionary of two dictionaries, css and js, which list css files
# and js files that can be bundled together by the minify app.
diff --git a/setup.py b/setup.py
index 58dbd93..9a38628 100644
--- a/setup.py
+++ b/setup.py
@@ -3,7 +3,7 @@ import os
from setuptools import setup, find_packages
-setup(name='project',
+setup(name='%(pkg)s',
version='1.0',
description='Django application.',
long_description='',
"""
% vars)
git(['mv', 'project', pkg])
git(['commit', '-a', '-m', 'Renamed project module to %s' % pkg])
| init_pkg | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(pkg, repo_dest):\n vars = {'pkg': pkg}\n with dir_path(repo_dest):\n patch(\n \"\"\" diff --git a/manage.py b/manage.py\n index 40ebb0a..cdfe363 100755\n --- a/manage.py\n +++ b/manage.py\n @@ -3,7 +3,7 @@ import os\n import sys\n\n # Edit this if necessary or override the variable in your environment.\n -os.environ.setdefault('DJANGO_SETTINGS_MODULE', 'project.settings')\n +os.environ.setdefault('DJANGO_SETTINGS_MODULE', '%(pkg)s.settings')\n\n try:\n # For local development in a virtualenv:\n diff --git a/project/settings/base.py b/project/settings/base.py\n index 312f280..c75e673 100644\n --- a/project/settings/base.py\n +++ b/project/settings/base.py\n @@ -7,7 +7,7 @@ from funfactory.settings_base import *\n # If you did not install Playdoh with the funfactory installer script\n # you may need to edit this value. See the docs about installing from a\n # clone.\n -PROJECT_MODULE = 'project'\n +PROJECT_MODULE = '%(pkg)s'\n\n # Bundles is a dictionary of two dictionaries, css and js, which list css files\n # and js files that can be bundled together by the minify app.\n diff --git a/setup.py b/setup.py\n index 58dbd93..9a38628 100644\n --- a/setup.py\n +++ b/setup.py\n @@ -3,7 +3,7 @@ import os\n from setuptools import setup, find_packages\n\n\n -setup(name='project',\n +setup(name='%(pkg)s',\n version='1.0',\n description='Django application.',\n long_description='',\n \"\"\"\n % vars)\n git(['mv', 'project', pkg])\n git(['commit', '-a', '-m', 'Renamed project module to %s' % pkg])\n",
"role": "user"
},
{
"content": "init_pkg",
"role": "assistant"
}
] |
def x(pkg, repo_dest, db_user, db_name, db_password, db_host, db_port):
vars = {'pkg': pkg, 'db_user': db_user, 'db_name': db_name,
'db_password': db_password or '', 'db_host': db_host or '',
'db_port': db_port or '', 'hmac_date': datetime.now().strftime(
'%Y-%m-%d'), 'hmac_key': generate_key(32), 'secret_key':
generate_key(32)}
with dir_path(repo_dest):
shutil.copyfile('%s/settings/local.py-dist' % pkg,
'%s/settings/local.py' % pkg)
patch(
""" --- a/%(pkg)s/settings/local.py
+++ b/%(pkg)s/settings/local.py
@@ -9,11 +9,11 @@ from . import base
DATABASES = {
'default': {
'ENGINE': 'django.db.backends.mysql',
- 'NAME': 'playdoh_app',
- 'USER': 'root',
- 'PASSWORD': '',
- 'HOST': '',
- 'PORT': '',
+ 'NAME': '%(db_name)s',
+ 'USER': '%(db_user)s',
+ 'PASSWORD': '%(db_password)s',
+ 'HOST': '%(db_host)s',
+ 'PORT': '%(db_port)s',
'OPTIONS': {
'init_command': 'SET storage_engine=InnoDB',
'charset' : 'utf8',
@@ -51,14 +51,14 @@ DEV = True
# Playdoh ships with Bcrypt+HMAC by default because it's the most secure.
# To use bcrypt, fill in a secret HMAC key. It cannot be blank.
HMAC_KEYS = {
- #'2012-06-06': 'some secret',
+ '%(hmac_date)s': '%(hmac_key)s',
}
from django_sha2 import get_password_hashers
PASSWORD_HASHERS = get_password_hashers(base.BASE_PASSWORD_HASHERS, HMAC_KEYS)
# Make this unique, and don't share it with anybody. It cannot be blank.
-SECRET_KEY = ''
+SECRET_KEY = '%(secret_key)s'
# Uncomment these to activate and customize Celery:
# CELERY_ALWAYS_EAGER = False # required to activate celeryd
"""
% vars)
| create_settings | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(pkg, repo_dest, db_user, db_name, db_password, db_host, db_port):\n vars = {'pkg': pkg, 'db_user': db_user, 'db_name': db_name,\n 'db_password': db_password or '', 'db_host': db_host or '',\n 'db_port': db_port or '', 'hmac_date': datetime.now().strftime(\n '%Y-%m-%d'), 'hmac_key': generate_key(32), 'secret_key':\n generate_key(32)}\n with dir_path(repo_dest):\n shutil.copyfile('%s/settings/local.py-dist' % pkg, \n '%s/settings/local.py' % pkg)\n patch(\n \"\"\" --- a/%(pkg)s/settings/local.py\n +++ b/%(pkg)s/settings/local.py\n @@ -9,11 +9,11 @@ from . import base\n DATABASES = {\n 'default': {\n 'ENGINE': 'django.db.backends.mysql',\n - 'NAME': 'playdoh_app',\n - 'USER': 'root',\n - 'PASSWORD': '',\n - 'HOST': '',\n - 'PORT': '',\n + 'NAME': '%(db_name)s',\n + 'USER': '%(db_user)s',\n + 'PASSWORD': '%(db_password)s',\n + 'HOST': '%(db_host)s',\n + 'PORT': '%(db_port)s',\n 'OPTIONS': {\n 'init_command': 'SET storage_engine=InnoDB',\n 'charset' : 'utf8',\n @@ -51,14 +51,14 @@ DEV = True\n # Playdoh ships with Bcrypt+HMAC by default because it's the most secure.\n # To use bcrypt, fill in a secret HMAC key. It cannot be blank.\n HMAC_KEYS = {\n - #'2012-06-06': 'some secret',\n + '%(hmac_date)s': '%(hmac_key)s',\n }\n\n from django_sha2 import get_password_hashers\n PASSWORD_HASHERS = get_password_hashers(base.BASE_PASSWORD_HASHERS, HMAC_KEYS)\n\n # Make this unique, and don't share it with anybody. It cannot be blank.\n -SECRET_KEY = ''\n +SECRET_KEY = '%(secret_key)s'\n\n # Uncomment these to activate and customize Celery:\n # CELERY_ALWAYS_EAGER = False # required to activate celeryd\n \"\"\"\n % vars)\n",
"role": "user"
},
{
"content": "create_settings",
"role": "assistant"
}
] |
def x(pkg, repo_dest, python):
workon_home = os.environ.get('WORKON_HOME')
venv_cmd = find_executable('virtualenv')
python_bin = find_executable(python)
if not python_bin:
raise EnvironmentError(
'%s is not installed or not available on your $PATH' % python)
if workon_home:
venv = os.path.join(workon_home, pkg)
else:
venv = os.path.join(repo_dest, '.virtualenv')
if venv_cmd:
if not verbose:
log.info('Creating virtual environment in %r' % venv)
args = ['--python', python_bin, venv]
if not verbose:
args.insert(0, '-q')
subprocess.check_call([venv_cmd] + args)
else:
raise EnvironmentError(
'Could not locate the virtualenv. Install with pip install virtualenv.'
)
return venv
| create_virtualenv | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(pkg, repo_dest, python):\n workon_home = os.environ.get('WORKON_HOME')\n venv_cmd = find_executable('virtualenv')\n python_bin = find_executable(python)\n if not python_bin:\n raise EnvironmentError(\n '%s is not installed or not available on your $PATH' % python)\n if workon_home:\n venv = os.path.join(workon_home, pkg)\n else:\n venv = os.path.join(repo_dest, '.virtualenv')\n if venv_cmd:\n if not verbose:\n log.info('Creating virtual environment in %r' % venv)\n args = ['--python', python_bin, venv]\n if not verbose:\n args.insert(0, '-q')\n subprocess.check_call([venv_cmd] + args)\n else:\n raise EnvironmentError(\n 'Could not locate the virtualenv. Install with pip install virtualenv.'\n )\n return venv\n",
"role": "user"
},
{
"content": "create_virtualenv",
"role": "assistant"
}
] |
def x(venv, repo_dest):
with dir_path(repo_dest):
args = ['-r', 'requirements/compiled.txt']
if not verbose:
args.insert(0, '-q')
subprocess.check_call([os.path.join(venv, 'bin', 'pip'), 'install'] +
args)
| install_reqs | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(venv, repo_dest):\n with dir_path(repo_dest):\n args = ['-r', 'requirements/compiled.txt']\n if not verbose:\n args.insert(0, '-q')\n subprocess.check_call([os.path.join(venv, 'bin', 'pip'), 'install'] +\n args)\n",
"role": "user"
},
{
"content": "install_reqs",
"role": "assistant"
}
] |
def x(name):
for pt in os.environ.get('PATH', '').split(':'):
candidate = os.path.join(pt, name)
if os.path.exists(candidate):
return candidate
| find_executable | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(name):\n for pt in os.environ.get('PATH', '').split(':'):\n candidate = os.path.join(pt, name)\n if os.path.exists(candidate):\n return candidate\n",
"role": "user"
},
{
"content": "find_executable",
"role": "assistant"
}
] |
def x(dir):
old_dir = os.getcwd()
os.chdir(dir)
yield
os.chdir(old_dir)
| dir_path | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(dir):\n old_dir = os.getcwd()\n os.chdir(dir)\n yield\n os.chdir(old_dir)\n",
"role": "user"
},
{
"content": "dir_path",
"role": "assistant"
}
] |
def x(maptype='jet', reverse=True):
d = {'A': 89, 'R': 174, 'N': 132, 'D': 133, 'C': 121, 'Q': 146, 'E':
147, 'G': 75, 'H': 155, 'I': 131, 'L': 131, 'K': 146, 'M': 149, 'F':
165, 'P': 115, 'S': 105, 'T': 119, 'W': 204, 'Y': 181, 'V': 117}
aas = sorted(AA_TO_INDEX.keys())
mws = [d[aa] for aa in aas]
if reverse:
mws = [(-1 * x) for x in mws]
mapper = pylab.cm.ScalarMappable(cmap=maptype)
mapper.set_clim(min(mws), max(mws))
mapping_d = {'*': '#000000'}
for aa, h in zip(aas, mws):
tup = mapper.to_rgba(h, bytes=True)
red, green, blue, alpha = tup
mapping_d[aa] = '#%02x%02x%02x' % (red, green, blue)
assert len(mapping_d[aa]) == 7
cmap = mapper.get_cmap()
return cmap, mapping_d, mapper
| MWColorMapping | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(maptype='jet', reverse=True):\n d = {'A': 89, 'R': 174, 'N': 132, 'D': 133, 'C': 121, 'Q': 146, 'E': \n 147, 'G': 75, 'H': 155, 'I': 131, 'L': 131, 'K': 146, 'M': 149, 'F':\n 165, 'P': 115, 'S': 105, 'T': 119, 'W': 204, 'Y': 181, 'V': 117}\n aas = sorted(AA_TO_INDEX.keys())\n mws = [d[aa] for aa in aas]\n if reverse:\n mws = [(-1 * x) for x in mws]\n mapper = pylab.cm.ScalarMappable(cmap=maptype)\n mapper.set_clim(min(mws), max(mws))\n mapping_d = {'*': '#000000'}\n for aa, h in zip(aas, mws):\n tup = mapper.to_rgba(h, bytes=True)\n red, green, blue, alpha = tup\n mapping_d[aa] = '#%02x%02x%02x' % (red, green, blue)\n assert len(mapping_d[aa]) == 7\n cmap = mapper.get_cmap()\n return cmap, mapping_d, mapper\n",
"role": "user"
},
{
"content": "MWColorMapping",
"role": "assistant"
}
] |
def x(maptype='jet', reverse=False):
pos_color = '#FF0000'
neg_color = '#0000FF'
neut_color = '#000000'
mapping_d = {'A': neut_color, 'R': pos_color, 'N': neut_color, 'D':
neg_color, 'C': neut_color, 'Q': neut_color, 'E': neg_color, 'G':
neut_color, 'H': pos_color, 'I': neut_color, 'L': neut_color, 'K':
pos_color, 'M': neut_color, 'F': neut_color, 'P': neut_color, 'S':
neut_color, 'T': neut_color, 'W': neut_color, 'Y': neut_color, 'V':
neut_color}
return None, mapping_d, None
| ChargeColorMapping | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(maptype='jet', reverse=False):\n pos_color = '#FF0000'\n neg_color = '#0000FF'\n neut_color = '#000000'\n mapping_d = {'A': neut_color, 'R': pos_color, 'N': neut_color, 'D':\n neg_color, 'C': neut_color, 'Q': neut_color, 'E': neg_color, 'G':\n neut_color, 'H': pos_color, 'I': neut_color, 'L': neut_color, 'K':\n pos_color, 'M': neut_color, 'F': neut_color, 'P': neut_color, 'S':\n neut_color, 'T': neut_color, 'W': neut_color, 'Y': neut_color, 'V':\n neut_color}\n return None, mapping_d, None\n",
"role": "user"
},
{
"content": "ChargeColorMapping",
"role": "assistant"
}
] |
def x(maptype='jet', reverse=False):
small_color = '#f76ab4'
nucleophilic_color = '#ff7f00'
hydrophobic_color = '#12ab0d'
aromatic_color = '#84380b'
acidic_color = '#e41a1c'
amide_color = '#972aa8'
basic_color = '#3c58e5'
mapping_d = {'G': small_color, 'A': small_color, 'S':
nucleophilic_color, 'T': nucleophilic_color, 'C':
nucleophilic_color, 'V': hydrophobic_color, 'L': hydrophobic_color,
'I': hydrophobic_color, 'M': hydrophobic_color, 'P':
hydrophobic_color, 'F': aromatic_color, 'Y': aromatic_color, 'W':
aromatic_color, 'D': acidic_color, 'E': acidic_color, 'H':
basic_color, 'K': basic_color, 'R': basic_color, 'N': amide_color,
'Q': amide_color, '*': '#000000'}
return None, mapping_d, None
| FunctionalGroupColorMapping | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(maptype='jet', reverse=False):\n small_color = '#f76ab4'\n nucleophilic_color = '#ff7f00'\n hydrophobic_color = '#12ab0d'\n aromatic_color = '#84380b'\n acidic_color = '#e41a1c'\n amide_color = '#972aa8'\n basic_color = '#3c58e5'\n mapping_d = {'G': small_color, 'A': small_color, 'S':\n nucleophilic_color, 'T': nucleophilic_color, 'C':\n nucleophilic_color, 'V': hydrophobic_color, 'L': hydrophobic_color,\n 'I': hydrophobic_color, 'M': hydrophobic_color, 'P':\n hydrophobic_color, 'F': aromatic_color, 'Y': aromatic_color, 'W':\n aromatic_color, 'D': acidic_color, 'E': acidic_color, 'H':\n basic_color, 'K': basic_color, 'R': basic_color, 'N': amide_color,\n 'Q': amide_color, '*': '#000000'}\n return None, mapping_d, None\n",
"role": "user"
},
{
"content": "FunctionalGroupColorMapping",
"role": "assistant"
}
] |
def x(data, format, ordered_alphabets):
eps = _my_eps_formatter(data, format, ordered_alphabets).decode()
gs = weblogolib.GhostscriptAPI()
return gs.convert('pdf', eps, format.logo_width, format.logo_height)
| my_pdf_formatter | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(data, format, ordered_alphabets):\n eps = _my_eps_formatter(data, format, ordered_alphabets).decode()\n gs = weblogolib.GhostscriptAPI()\n return gs.convert('pdf', eps, format.logo_width, format.logo_height)\n",
"role": "user"
},
{
"content": "my_pdf_formatter",
"role": "assistant"
}
] |
def x(logodata, format, ordered_alphabets):
substitutions = {}
from_format = ['creation_date', 'logo_width', 'logo_height',
'lines_per_logo', 'line_width', 'line_height', 'line_margin_right',
'line_margin_left', 'line_margin_bottom', 'line_margin_top',
'title_height', 'xaxis_label_height', 'creator_text', 'logo_title',
'logo_margin', 'stroke_width', 'tic_length', 'stacks_per_line',
'stack_margin', 'yaxis_label', 'yaxis_tic_interval',
'yaxis_minor_tic_interval', 'xaxis_label', 'xaxis_tic_interval',
'number_interval', 'fineprint', 'shrink_fraction',
'errorbar_fraction', 'errorbar_width_fraction', 'errorbar_gray',
'small_fontsize', 'fontsize', 'title_fontsize', 'number_fontsize',
'text_font', 'logo_font', 'title_font', 'logo_label', 'yaxis_scale',
'end_type', 'debug', 'show_title', 'show_xaxis', 'show_xaxis_label',
'show_yaxis', 'show_yaxis_label', 'show_boxes', 'show_errorbars',
'show_fineprint', 'rotate_numbers', 'show_ends', 'stack_height',
'stack_width']
for s in from_format:
substitutions[s] = getattr(format, s)
substitutions['shrink'] = str(format.show_boxes).lower()
def x(color):
return ' '.join(('[', str(color.red), str(color.green), str(color.
blue), ']'))
substitutions['default_color'] = format_color(format.default_color)
colors = []
if hasattr(format.color_scheme, 'rules'):
grouplist = format.color_scheme.rules
else:
grouplist = format.color_scheme.groups
for group in grouplist:
cf = format_color(group.color)
for s in group.symbols:
colors.append(' (' + s + ') ' + cf)
substitutions['color_dict'] = '\n'.join(colors)
data = []
conv_factor = None
data.append('StartLine')
seq_from = format.logo_start - format.first_index
seq_to = format.logo_end - format.first_index + 1
for seq_index in range(seq_from, seq_to):
logo_index = seq_index + format.first_index
stack_index = seq_index - seq_from
if stack_index != 0 and stack_index % format.stacks_per_line == 0:
data.append('')
data.append('EndLine')
data.append('StartLine')
data.append('')
data.append('(%s) StartStack' % format.annotate[seq_index])
if conv_factor:
stack_height = logodata.entropy[seq_index] * std_units[format.
unit_name]
else:
stack_height = 1.0
s_d = dict(zip(logodata.alphabet, logodata.counts[seq_index]))
s = []
for aa in ordered_alphabets[seq_index]:
if aa not in ['B', 'b']:
s.append((s_d[aa], aa))
else:
s.append((s_d[aa], ' '))
C = float(sum(logodata.counts[seq_index]))
if C > 0.0:
fraction_width = 1.0
if format.scale_width:
fraction_width = logodata.weight[seq_index]
for c in s:
data.append(' %f %f (%s) ShowSymbol' % (fraction_width, c[0
] * stack_height / C, c[1]))
if logodata.entropy_interval is not None and conv_factor and C > 0.0:
low, high = logodata.entropy_interval[seq_index]
center = logodata.entropy[seq_index]
low *= conv_factor
high *= conv_factor
center *= conv_factor
if high > format.yaxis_scale:
high = format.yaxis_scale
down = center - low
up = high - center
data.append(' %f %f DrawErrorbar' % (down, up))
data.append('EndStack')
data.append('')
data.append('EndLine')
substitutions['logo_data'] = '\n'.join(data)
template = corebio.utils.resource_string(__name__,
'_weblogo_template.eps', __file__).decode()
logo = string.Template(template).substitute(substitutions)
return logo.encode()
| my_eps_formatter | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(logodata, format, ordered_alphabets):\n substitutions = {}\n from_format = ['creation_date', 'logo_width', 'logo_height',\n 'lines_per_logo', 'line_width', 'line_height', 'line_margin_right',\n 'line_margin_left', 'line_margin_bottom', 'line_margin_top',\n 'title_height', 'xaxis_label_height', 'creator_text', 'logo_title',\n 'logo_margin', 'stroke_width', 'tic_length', 'stacks_per_line',\n 'stack_margin', 'yaxis_label', 'yaxis_tic_interval',\n 'yaxis_minor_tic_interval', 'xaxis_label', 'xaxis_tic_interval',\n 'number_interval', 'fineprint', 'shrink_fraction',\n 'errorbar_fraction', 'errorbar_width_fraction', 'errorbar_gray',\n 'small_fontsize', 'fontsize', 'title_fontsize', 'number_fontsize',\n 'text_font', 'logo_font', 'title_font', 'logo_label', 'yaxis_scale',\n 'end_type', 'debug', 'show_title', 'show_xaxis', 'show_xaxis_label',\n 'show_yaxis', 'show_yaxis_label', 'show_boxes', 'show_errorbars',\n 'show_fineprint', 'rotate_numbers', 'show_ends', 'stack_height',\n 'stack_width']\n for s in from_format:\n substitutions[s] = getattr(format, s)\n substitutions['shrink'] = str(format.show_boxes).lower()\n\n def x(color):\n return ' '.join(('[', str(color.red), str(color.green), str(color.\n blue), ']'))\n substitutions['default_color'] = format_color(format.default_color)\n colors = []\n if hasattr(format.color_scheme, 'rules'):\n grouplist = format.color_scheme.rules\n else:\n grouplist = format.color_scheme.groups\n for group in grouplist:\n cf = format_color(group.color)\n for s in group.symbols:\n colors.append(' (' + s + ') ' + cf)\n substitutions['color_dict'] = '\\n'.join(colors)\n data = []\n conv_factor = None\n data.append('StartLine')\n seq_from = format.logo_start - format.first_index\n seq_to = format.logo_end - format.first_index + 1\n for seq_index in range(seq_from, seq_to):\n logo_index = seq_index + format.first_index\n stack_index = seq_index - seq_from\n if stack_index != 0 and stack_index % format.stacks_per_line == 0:\n data.append('')\n data.append('EndLine')\n data.append('StartLine')\n data.append('')\n data.append('(%s) StartStack' % format.annotate[seq_index])\n if conv_factor:\n stack_height = logodata.entropy[seq_index] * std_units[format.\n unit_name]\n else:\n stack_height = 1.0\n s_d = dict(zip(logodata.alphabet, logodata.counts[seq_index]))\n s = []\n for aa in ordered_alphabets[seq_index]:\n if aa not in ['B', 'b']:\n s.append((s_d[aa], aa))\n else:\n s.append((s_d[aa], ' '))\n C = float(sum(logodata.counts[seq_index]))\n if C > 0.0:\n fraction_width = 1.0\n if format.scale_width:\n fraction_width = logodata.weight[seq_index]\n for c in s:\n data.append(' %f %f (%s) ShowSymbol' % (fraction_width, c[0\n ] * stack_height / C, c[1]))\n if logodata.entropy_interval is not None and conv_factor and C > 0.0:\n low, high = logodata.entropy_interval[seq_index]\n center = logodata.entropy[seq_index]\n low *= conv_factor\n high *= conv_factor\n center *= conv_factor\n if high > format.yaxis_scale:\n high = format.yaxis_scale\n down = center - low\n up = high - center\n data.append(' %f %f DrawErrorbar' % (down, up))\n data.append('EndStack')\n data.append('')\n data.append('EndLine')\n substitutions['logo_data'] = '\\n'.join(data)\n template = corebio.utils.resource_string(__name__,\n '_weblogo_template.eps', __file__).decode()\n logo = string.Template(template).substitute(substitutions)\n return logo.encode()\n",
"role": "user"
},
{
"content": "my_eps_formatter",
"role": "assistant"
}
] |
def x(fin, alphabet=None):
items = []
start = True
for line in fin:
if line.isspace() or line[0] == '#':
continue
stuff = line.split()
if start and stuff[0] != 'PO' and stuff[0] != 'P0':
continue
if stuff[0] == 'XX' or stuff[0] == '//':
break
start = False
items.append(stuff)
if len(items) < 2:
raise ValueError('Vacuous file.')
header = items.pop(0)
hcols = len(header)
rows = len(items)
cols = len(items[0])
if not (header[0] == 'PO' or header[0] == 'P0' or hcols == cols - 1 or
hcols == cols - 2):
raise ValueError('Missing header line!')
cols = len(items[0])
for i in range(1, len(items)):
if cols != len(items[i]):
raise ValueError('Inconsistant length, row %d: ' % i)
if header[0] == 'PO' or header[0] == 'P0':
header.pop(0)
position_header = True
alphabet_header = True
for h in header:
if not corebio.utils.isint(h):
position_header = False
if not position_header and not alphabet_header:
raise ValueError("Can't parse header: %s" % str(header))
if position_header and alphabet_header:
raise ValueError("Can't parse header")
if alphabet_header:
for i, r in enumerate(items):
if not corebio.utils.isint(r[0]) and r[0][0] != 'P':
raise ValueError(
'Expected position as first item on line %d' % i)
r.pop(0)
defacto_alphabet = ''.join(header)
else:
a = []
for i, r in enumerate(items):
if not ischar(r[0]) and r[0][0] != 'P':
raise ValueError(
'Expected position as first item on line %d' % i)
a.append(r.pop(0))
defacto_alphabet = ''.join(a)
defacto_alphabet = corebio.seq.Alphabet(defacto_alphabet)
if alphabet:
if not defacto_alphabet.alphabetic(alphabet):
raise ValueError('Incompatible alphabets: %s , %s (defacto)' %
(alphabet, defacto_alphabet))
else:
alphabets = (unambiguous_rna_alphabet, unambiguous_dna_alphabet,
unambiguous_protein_alphabet)
for a in alphabets:
if defacto_alphabet.alphabetic(a):
alphabet = a
break
if not alphabet:
alphabet = defacto_alphabet
if len(items[0]) == len(header) + 1:
for r in items:
r.pop()
rows = len(items)
cols = len(items[0])
matrix = numpy.zeros((rows, cols), dtype=numpy.float64)
for r in range(rows):
for c in range(cols):
matrix[r, c] = float(items[r][c])
if position_header:
matrix.transpose()
return _my_Motif(defacto_alphabet, matrix).reindex(alphabet)
| read_transfac | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(fin, alphabet=None):\n items = []\n start = True\n for line in fin:\n if line.isspace() or line[0] == '#':\n continue\n stuff = line.split()\n if start and stuff[0] != 'PO' and stuff[0] != 'P0':\n continue\n if stuff[0] == 'XX' or stuff[0] == '//':\n break\n start = False\n items.append(stuff)\n if len(items) < 2:\n raise ValueError('Vacuous file.')\n header = items.pop(0)\n hcols = len(header)\n rows = len(items)\n cols = len(items[0])\n if not (header[0] == 'PO' or header[0] == 'P0' or hcols == cols - 1 or \n hcols == cols - 2):\n raise ValueError('Missing header line!')\n cols = len(items[0])\n for i in range(1, len(items)):\n if cols != len(items[i]):\n raise ValueError('Inconsistant length, row %d: ' % i)\n if header[0] == 'PO' or header[0] == 'P0':\n header.pop(0)\n position_header = True\n alphabet_header = True\n for h in header:\n if not corebio.utils.isint(h):\n position_header = False\n if not position_header and not alphabet_header:\n raise ValueError(\"Can't parse header: %s\" % str(header))\n if position_header and alphabet_header:\n raise ValueError(\"Can't parse header\")\n if alphabet_header:\n for i, r in enumerate(items):\n if not corebio.utils.isint(r[0]) and r[0][0] != 'P':\n raise ValueError(\n 'Expected position as first item on line %d' % i)\n r.pop(0)\n defacto_alphabet = ''.join(header)\n else:\n a = []\n for i, r in enumerate(items):\n if not ischar(r[0]) and r[0][0] != 'P':\n raise ValueError(\n 'Expected position as first item on line %d' % i)\n a.append(r.pop(0))\n defacto_alphabet = ''.join(a)\n defacto_alphabet = corebio.seq.Alphabet(defacto_alphabet)\n if alphabet:\n if not defacto_alphabet.alphabetic(alphabet):\n raise ValueError('Incompatible alphabets: %s , %s (defacto)' %\n (alphabet, defacto_alphabet))\n else:\n alphabets = (unambiguous_rna_alphabet, unambiguous_dna_alphabet,\n unambiguous_protein_alphabet)\n for a in alphabets:\n if defacto_alphabet.alphabetic(a):\n alphabet = a\n break\n if not alphabet:\n alphabet = defacto_alphabet\n if len(items[0]) == len(header) + 1:\n for r in items:\n r.pop()\n rows = len(items)\n cols = len(items[0])\n matrix = numpy.zeros((rows, cols), dtype=numpy.float64)\n for r in range(rows):\n for c in range(cols):\n matrix[r, c] = float(items[r][c])\n if position_header:\n matrix.transpose()\n return _my_Motif(defacto_alphabet, matrix).reindex(alphabet)\n",
"role": "user"
},
{
"content": "read_transfac",
"role": "assistant"
}
] |
def x(self, atoms):
_atoms = self.atoms[:]
for atom in _atoms:
if atom not in atoms:
self.remove_atom(atom)
| pruneToAtoms | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, atoms):\n _atoms = self.atoms[:]\n for atom in _atoms:\n if atom not in atoms:\n self.remove_atom(atom)\n",
"role": "user"
},
{
"content": "pruneToAtoms",
"role": "assistant"
}
] |
def x(connection: connection, timeout: float=1.0) ->Iterable[Event]:
if timeout > 0.0:
log('Polling for events (Blocking, {} seconds)...'.format(timeout),
logger_name=_LOGGER_NAME)
else:
log('Polling for events (Non-Blocking)...', logger_name=_LOGGER_NAME)
if select.select([connection], [], [], timeout) == ([], [], []):
log('...No events found', logger_name=_LOGGER_NAME)
return
else:
log('Events', logger_name=_LOGGER_NAME)
log('------', logger_name=_LOGGER_NAME)
connection.poll()
while connection.notifies:
event = connection.notifies.pop(0)
log(str(event), logger_name=_LOGGER_NAME)
yield Event.fromjson(event.payload)
| poll | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(connection: connection, timeout: float=1.0) ->Iterable[Event]:\n if timeout > 0.0:\n log('Polling for events (Blocking, {} seconds)...'.format(timeout),\n logger_name=_LOGGER_NAME)\n else:\n log('Polling for events (Non-Blocking)...', logger_name=_LOGGER_NAME)\n if select.select([connection], [], [], timeout) == ([], [], []):\n log('...No events found', logger_name=_LOGGER_NAME)\n return\n else:\n log('Events', logger_name=_LOGGER_NAME)\n log('------', logger_name=_LOGGER_NAME)\n connection.poll()\n while connection.notifies:\n event = connection.notifies.pop(0)\n log(str(event), logger_name=_LOGGER_NAME)\n yield Event.fromjson(event.payload)\n",
"role": "user"
},
{
"content": "poll",
"role": "assistant"
}
] |
def x(cls, json_string: str) ->'Event':
obj = json.loads(json_string)
return cls(UUID(obj['event_id']), obj['event_type'], obj['schema_name'],
obj['table_name'], obj['row_id'])
| fromjson | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(cls, json_string: str) ->'Event':\n obj = json.loads(json_string)\n return cls(UUID(obj['event_id']), obj['event_type'], obj['schema_name'],\n obj['table_name'], obj['row_id'])\n",
"role": "user"
},
{
"content": "fromjson",
"role": "assistant"
}
] |
def x(self) ->str:
return json.dumps({'event_id': str(self.id), 'event_type': self.type,
'schema_name': self.schema_name, 'table_name': self.table_name,
'row_id': self.row_id})
| tojson | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self) ->str:\n return json.dumps({'event_id': str(self.id), 'event_type': self.type,\n 'schema_name': self.schema_name, 'table_name': self.table_name,\n 'row_id': self.row_id})\n",
"role": "user"
},
{
"content": "tojson",
"role": "assistant"
}
] |
def x(self):
ordererd = np.sort(np.asarray(list(self.plot_info.keys())).astype(int))
trans_cont_dict = OrderedDict()
for i in ordererd:
trans_cont_dict[str(i)] = self.plot_info[str(i)]
self.plot_info = trans_cont_dict
x = [[] for i in np.arange(len(self.plot_info.keys()))]
y = [[] for i in np.arange(len(self.plot_info.keys()))]
z = [[] for i in np.arange(len(self.plot_info.keys()))]
for k, axis_string in enumerate(self.plot_info.keys()):
if 'file' not in self.plot_info[axis_string].keys():
continue
for j, file_dict in enumerate(self.plot_info[axis_string]['file']):
data_class = ReadInData(**{**self.general, **file_dict, **self.
plot_info[axis_string]['limits']})
x[k].append(data_class.x_append_value)
y[k].append(data_class.y_append_value)
z[k].append(data_class.z_append_value)
for k, axis_string in enumerate(self.plot_info.keys()):
if 'indices' in self.plot_info[axis_string]:
if type(self.plot_info[axis_string]['indices']) == int:
self.plot_info[axis_string]['indices'] = [self.plot_info[
axis_string]['indices']]
for index in self.plot_info[axis_string]['indices']:
index = int(index)
x[k].append(x[index][0])
y[k].append(y[index][0])
z[k].append(z[index][0])
for k, axis_string in enumerate(self.plot_info.keys()):
if 'control' in self.plot_info[axis_string]:
if 'name' in self.plot_info[axis_string]['control'
] or 'label' in self.plot_info[axis_string]['control']:
file_dict = self.plot_info[axis_string]['control']
if 'limits' in self.plot_info[axis_string].keys():
liimits_dict = self.plot_info[axis_string]['limits']
data_class = ReadInData(**{**self.general, **file_dict, **
self.plot_info[axis_string]['limits']})
x[k].append(data_class.x_append_value)
y[k].append(data_class.y_append_value)
z[k].append(data_class.z_append_value)
elif 'index' in self.plot_info[axis_string]['control']:
index = int(self.plot_info[axis_string]['control']['index'])
x[k].append(x[index][0])
y[k].append(y[index][0])
z[k].append(z[index][0])
value_classes = []
for k in range(len(x)):
value_classes.append(PlotVals(x[k], y[k], z[k]))
self.value_classes = value_classes
return
| input_data | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n ordererd = np.sort(np.asarray(list(self.plot_info.keys())).astype(int))\n trans_cont_dict = OrderedDict()\n for i in ordererd:\n trans_cont_dict[str(i)] = self.plot_info[str(i)]\n self.plot_info = trans_cont_dict\n x = [[] for i in np.arange(len(self.plot_info.keys()))]\n y = [[] for i in np.arange(len(self.plot_info.keys()))]\n z = [[] for i in np.arange(len(self.plot_info.keys()))]\n for k, axis_string in enumerate(self.plot_info.keys()):\n if 'file' not in self.plot_info[axis_string].keys():\n continue\n for j, file_dict in enumerate(self.plot_info[axis_string]['file']):\n data_class = ReadInData(**{**self.general, **file_dict, **self.\n plot_info[axis_string]['limits']})\n x[k].append(data_class.x_append_value)\n y[k].append(data_class.y_append_value)\n z[k].append(data_class.z_append_value)\n for k, axis_string in enumerate(self.plot_info.keys()):\n if 'indices' in self.plot_info[axis_string]:\n if type(self.plot_info[axis_string]['indices']) == int:\n self.plot_info[axis_string]['indices'] = [self.plot_info[\n axis_string]['indices']]\n for index in self.plot_info[axis_string]['indices']:\n index = int(index)\n x[k].append(x[index][0])\n y[k].append(y[index][0])\n z[k].append(z[index][0])\n for k, axis_string in enumerate(self.plot_info.keys()):\n if 'control' in self.plot_info[axis_string]:\n if 'name' in self.plot_info[axis_string]['control'\n ] or 'label' in self.plot_info[axis_string]['control']:\n file_dict = self.plot_info[axis_string]['control']\n if 'limits' in self.plot_info[axis_string].keys():\n liimits_dict = self.plot_info[axis_string]['limits']\n data_class = ReadInData(**{**self.general, **file_dict, **\n self.plot_info[axis_string]['limits']})\n x[k].append(data_class.x_append_value)\n y[k].append(data_class.y_append_value)\n z[k].append(data_class.z_append_value)\n elif 'index' in self.plot_info[axis_string]['control']:\n index = int(self.plot_info[axis_string]['control']['index'])\n x[k].append(x[index][0])\n y[k].append(y[index][0])\n z[k].append(z[index][0])\n value_classes = []\n for k in range(len(x)):\n value_classes.append(PlotVals(x[k], y[k], z[k]))\n self.value_classes = value_classes\n return\n",
"role": "user"
},
{
"content": "input_data",
"role": "assistant"
}
] |
def x(self):
fig, ax = plt.subplots(nrows=int(self.num_rows), ncols=int(self.
num_cols), sharex=self.sharex, sharey=self.sharey)
fig.set_size_inches(self.figure_width, self.figure_height)
try:
ax = ax.ravel()
except AttributeError:
ax = [ax]
self.plot_types = [self.plot_info[str(i)]['plot_type'] for i in range(
len(ax))]
if len(self.plot_types) == 1:
if self.plot_types[0] not in self.colorbars:
self.colorbars[self.plot_types[0]] = {'cbar_pos': 5}
elif 'cbar_pos' not in self.colorbars[self.plot_types[0]]:
self.colorbars[self.plot_types[0]]['cbar_pos'] = 5
self.colorbar_classes = {}
for plot_type in self.plot_types:
if plot_type in self.colorbar_classes:
continue
if plot_type == 'Horizon':
self.colorbar_classes[plot_type] = None
elif plot_type in self.colorbars:
self.colorbar_classes[plot_type] = FigColorbar(fig, plot_type,
**self.colorbars[plot_type])
else:
self.colorbar_classes[plot_type] = FigColorbar(fig, plot_type)
if 'Ratio' in self.plot_types or 'Waterfall':
self.subplots_adjust_kwargs['right'] = 0.79
fig.subplots_adjust(**self.subplots_adjust_kwargs)
if 'fig_y_label' in self.__dict__.keys():
fig.text(self.fig_y_label_x, self.fig_y_label_y, '{}'.format(self.
fig_y_label), **self.fig_y_label_kwargs)
if 'fig_x_label' in self.__dict__.keys():
fig.text(self.fig_x_label_x, self.fig_x_label_y, '{}'.format(self.
fig_x_label), **self.fig_x_label_kwargs)
if 'fig_title' in self.__dict__.keys():
fig.text(self.fig_title_kwargs['x'], self.fig_title_kwargs['y'],
'{}'.format(self.fig_title), **self.fig_title_kwargs)
self.fig, self.ax = fig, ax
return
| setup_figure | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n fig, ax = plt.subplots(nrows=int(self.num_rows), ncols=int(self.\n num_cols), sharex=self.sharex, sharey=self.sharey)\n fig.set_size_inches(self.figure_width, self.figure_height)\n try:\n ax = ax.ravel()\n except AttributeError:\n ax = [ax]\n self.plot_types = [self.plot_info[str(i)]['plot_type'] for i in range(\n len(ax))]\n if len(self.plot_types) == 1:\n if self.plot_types[0] not in self.colorbars:\n self.colorbars[self.plot_types[0]] = {'cbar_pos': 5}\n elif 'cbar_pos' not in self.colorbars[self.plot_types[0]]:\n self.colorbars[self.plot_types[0]]['cbar_pos'] = 5\n self.colorbar_classes = {}\n for plot_type in self.plot_types:\n if plot_type in self.colorbar_classes:\n continue\n if plot_type == 'Horizon':\n self.colorbar_classes[plot_type] = None\n elif plot_type in self.colorbars:\n self.colorbar_classes[plot_type] = FigColorbar(fig, plot_type,\n **self.colorbars[plot_type])\n else:\n self.colorbar_classes[plot_type] = FigColorbar(fig, plot_type)\n if 'Ratio' in self.plot_types or 'Waterfall':\n self.subplots_adjust_kwargs['right'] = 0.79\n fig.subplots_adjust(**self.subplots_adjust_kwargs)\n if 'fig_y_label' in self.__dict__.keys():\n fig.text(self.fig_y_label_x, self.fig_y_label_y, '{}'.format(self.\n fig_y_label), **self.fig_y_label_kwargs)\n if 'fig_x_label' in self.__dict__.keys():\n fig.text(self.fig_x_label_x, self.fig_x_label_y, '{}'.format(self.\n fig_x_label), **self.fig_x_label_kwargs)\n if 'fig_title' in self.__dict__.keys():\n fig.text(self.fig_title_kwargs['x'], self.fig_title_kwargs['y'],\n '{}'.format(self.fig_title), **self.fig_title_kwargs)\n self.fig, self.ax = fig, ax\n return\n",
"role": "user"
},
{
"content": "setup_figure",
"role": "assistant"
}
] |
def x(self):
for i, axis in enumerate(self.ax):
trans_plot_class_call = globals()[self.plot_types[i]]
trans_plot_class = trans_plot_class_call(self.fig, axis, self.
value_classes[i].x_arr_list, self.value_classes[i].y_arr_list,
self.value_classes[i].z_arr_list, colorbar=self.
colorbar_classes[self.plot_types[i]], **{**self.general, **self
.figure, **self.plot_info[str(i)], **self.plot_info[str(i)][
'limits'], **self.plot_info[str(i)]['label'], **self.plot_info[
str(i)]['extra'], **self.plot_info[str(i)]['legend']})
trans_plot_class.make_plot()
trans_plot_class.setup_plot()
return
| create_plots | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n for i, axis in enumerate(self.ax):\n trans_plot_class_call = globals()[self.plot_types[i]]\n trans_plot_class = trans_plot_class_call(self.fig, axis, self.\n value_classes[i].x_arr_list, self.value_classes[i].y_arr_list,\n self.value_classes[i].z_arr_list, colorbar=self.\n colorbar_classes[self.plot_types[i]], **{**self.general, **self\n .figure, **self.plot_info[str(i)], **self.plot_info[str(i)][\n 'limits'], **self.plot_info[str(i)]['label'], **self.plot_info[\n str(i)]['extra'], **self.plot_info[str(i)]['legend']})\n trans_plot_class.make_plot()\n trans_plot_class.setup_plot()\n return\n",
"role": "user"
},
{
"content": "create_plots",
"role": "assistant"
}
] |
def x(self, cardinality):
query = 'CREATE TABLE IF NOT EXISTS _{0}_gram ('.format(cardinality)
unique = ''
for i in reversed(range(cardinality)):
if i != 0:
unique += 'word_{0}, '.format(i)
query += 'word_{0} TEXT, '.format(i)
else:
unique += 'word'
query += 'word TEXT, count INTEGER, UNIQUE({0}) );'.format(unique)
self.execute_sql(query)
| create_ngram_table | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, cardinality):\n query = 'CREATE TABLE IF NOT EXISTS _{0}_gram ('.format(cardinality)\n unique = ''\n for i in reversed(range(cardinality)):\n if i != 0:\n unique += 'word_{0}, '.format(i)\n query += 'word_{0} TEXT, '.format(i)\n else:\n unique += 'word'\n query += 'word TEXT, count INTEGER, UNIQUE({0}) );'.format(unique)\n self.execute_sql(query)\n",
"role": "user"
},
{
"content": "create_ngram_table",
"role": "assistant"
}
] |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.