source_code stringlengths 45 7.24k | function_name stringlengths 1 111 | messages list |
|---|---|---|
def x(input_obj):
get_chunk = None
if isinstance(input_obj, Queue):
log_msg = 'queue'
get_chunk = get_queue_chunk_reader(input_obj)
elif callable(input_obj):
log_msg = 'callable'
get_chunk = get_callable_chunk_reader(input_obj)
elif hasattr(input_obj, 'read'):
log_msg = 'file descriptor'
get_chunk = get_file_chunk_reader(input_obj)
elif isinstance(input_obj, basestring):
log_msg = 'string'
get_chunk = get_iter_string_reader(input_obj)
elif isinstance(input_obj, bytes):
log_msg = 'bytes'
get_chunk = get_iter_string_reader(input_obj)
elif isinstance(input_obj, GeneratorType):
log_msg = 'generator'
get_chunk = get_iter_chunk_reader(iter(input_obj))
else:
try:
it = iter(input_obj)
except TypeError:
raise Exception('unknown input object')
else:
log_msg = 'general iterable'
get_chunk = get_iter_chunk_reader(it)
return get_chunk, log_msg
| determine_how_to_read_input | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(input_obj):\n get_chunk = None\n if isinstance(input_obj, Queue):\n log_msg = 'queue'\n get_chunk = get_queue_chunk_reader(input_obj)\n ... |
def x(stdin):
bufsize = 1024
iter_str = (stdin[i:i + bufsize] for i in range(0, len(stdin), bufsize))
return get_iter_chunk_reader(iter_str)
| get_iter_string_reader | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(stdin):\n bufsize = 1024\n iter_str = (stdin[i:i + bufsize] for i in range(0, len(stdin), bufsize))\n return get_iter_chunk_reader(iter_str)\n",
"... |
def x(path):
orig_path = os.getcwd()
os.chdir(path)
try:
yield
finally:
os.chdir(orig_path)
| pushd | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(path):\n orig_path = os.getcwd()\n os.chdir(path)\n try:\n yield\n finally:\n os.chdir(orig_path)\n",
"role": "user"
},
{
... |
def x(**kwargs):
kwargs_str = ','.join([('%s=%r' % (k, v)) for k, v in kwargs.items()])
raise DeprecationWarning(
"""
sh.args() has been deprecated because it was never thread safe. use the
following instead:
sh2 = sh({kwargs})
sh2.your_command()
or
sh2 = sh({kwargs})
from sh2 import your_command
your_command()
"""
.format(kwargs=kwargs_str))
| args | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(**kwargs):\n kwargs_str = ','.join([('%s=%r' % (k, v)) for k, v in kwargs.items()])\n raise DeprecationWarning(\n \"\"\"\n\nsh.args() has been dep... |
def x(orig):
prompt = '[sudo] password for %s: ' % getpass.getuser()
def x():
pw = getpass.getpass(prompt=prompt) + '\n'
yield pw
def x(args, kwargs):
password = kwargs.pop('password', None)
if password is None:
pass_getter = stdin()
else:
pass_getter = password.rstrip('\n') + '\n'
kwargs['_in'] = pass_getter
return args, kwargs
cmd = orig.bake('-S', _arg_preprocess=process)
return cmd
| sudo | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(orig):\n prompt = '[sudo] password for %s: ' % getpass.getuser()\n\n def x():\n pw = getpass.getpass(prompt=prompt) + '\\n'\n yield pw\n\n ... |
def x():
def x(importer):
return (importer.__class__.__name__ == ModuleImporterFromVariables.
__name__)
already_registered = any([(True) for i in sys.meta_path if test(i)])
if not already_registered:
importer = ModuleImporterFromVariables(restrict_to=['SelfWrapper'])
sys.meta_path.insert(0, importer)
return not already_registered
| register_importer | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x():\n\n def x(importer):\n return (importer.__class__.__name__ == ModuleImporterFromVariables.\n __name__)\n already_registered = any([(... |
def x(self):
if not self._process_completed:
self._process_completed = True
exit_code = self.process.wait()
if self.process.timed_out:
raise TimeoutException(-exit_code)
else:
self.handle_command_exit_code(exit_code)
if self.process._stdin_process:
self.process._stdin_process.command.wait()
self.log.info('process completed')
return self
| wait | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n if not self._process_completed:\n self._process_completed = True\n exit_code = self.process.wait()\n if self.process.timed_out... |
def x(self, code):
ca = self.call_args
exc_class = get_exc_exit_code_would_raise(code, ca['ok_code'], ca['piped'])
if exc_class:
exc = exc_class(self.ran, self.process.stdout, self.process.stderr,
ca['truncate_exc'])
raise exc
| handle_command_exit_code | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, code):\n ca = self.call_args\n exc_class = get_exc_exit_code_would_raise(code, ca['ok_code'], ca['piped'])\n if exc_class:\n exc = exc_cl... |
def x(self):
if self._stopped_iteration:
raise StopIteration()
while True:
try:
chunk = self.process._pipe_queue.get(True, 0.001)
except Empty:
if self.call_args['iter_noblock']:
return errno.EWOULDBLOCK
else:
if chunk is None:
self.wait()
self._stopped_iteration = True
raise StopIteration()
try:
return chunk.decode(self.call_args['encoding'], self.
call_args['decode_errors'])
except UnicodeDecodeError:
return chunk
| next | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n if self._stopped_iteration:\n raise StopIteration()\n while True:\n try:\n chunk = self.process._pipe_queue.get(True, 0... |
def x(kwargs):
kwargs = kwargs.copy()
call_args = {}
for parg, default in Command._call_args.items():
key = '_' + parg
if key in kwargs:
call_args[parg] = kwargs[key]
del kwargs[key]
invalid_kwargs = special_kwarg_validator(call_args, Command.
_kwarg_validators)
if invalid_kwargs:
exc_msg = []
for args, error_msg in invalid_kwargs:
exc_msg.append(' %r: %s' % (args, error_msg))
exc_msg = '\n'.join(exc_msg)
raise TypeError('Invalid special arguments:\n\n%s\n' % exc_msg)
return call_args, kwargs
| extract_call_args | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(kwargs):\n kwargs = kwargs.copy()\n call_args = {}\n for parg, default in Command._call_args.items():\n key = '_' + parg\n if key in kwa... |
def x(self):
if self.exit_code is not None:
return False, self.exit_code
acquired = self._wait_lock.acquire(False)
if not acquired:
if self.exit_code is not None:
return False, self.exit_code
return True, self.exit_code
try:
pid, exit_code = no_interrupt(os.waitpid, self.pid, os.WNOHANG)
if pid == self.pid:
self.exit_code = handle_process_exit_code(exit_code)
self._process_just_ended()
return False, self.exit_code
except OSError:
return False, self.exit_code
else:
return True, self.exit_code
finally:
self._wait_lock.release()
| is_alive | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n if self.exit_code is not None:\n return False, self.exit_code\n acquired = self._wait_lock.acquire(False)\n if not acquired:\n ... |
def x(self):
self.log.debug('acquiring wait lock to wait for completion')
with self._wait_lock:
self.log.debug('got wait lock')
witnessed_end = False
if self.exit_code is None:
self.log.debug('exit code not set, waiting on pid')
pid, exit_code = no_interrupt(os.waitpid, self.pid, 0)
self.exit_code = handle_process_exit_code(exit_code)
witnessed_end = True
else:
self.log.debug('exit code already set (%d), no need to wait',
self.exit_code)
self._quit_threads.set()
if self._input_thread:
self._input_thread.join()
timer = threading.Timer(2.0, self._stop_output_event.set)
timer.start()
self._output_thread.join()
timer.cancel()
self._background_thread.join()
if witnessed_end:
self._process_just_ended()
return self.exit_code
| wait | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n self.log.debug('acquiring wait lock to wait for completion')\n with self._wait_lock:\n self.log.debug('got wait lock')\n witnessed... |
def x(self):
try:
chunk = self.get_chunk()
if chunk is None:
raise DoneReadingForever
except DoneReadingForever:
self.log.debug('done reading')
if self.tty_in:
try:
char = termios.tcgetattr(self.stream)[6][termios.VEOF]
except:
char = chr(4).encode()
os.write(self.stream, char)
os.write(self.stream, char)
return True
except NotYetReadyToRead:
self.log.debug('received no data')
return False
if IS_PY3 and hasattr(chunk, 'encode'):
chunk = chunk.encode(self.encoding)
for proc_chunk in self.stream_bufferer.process(chunk):
self.log.debug('got chunk size %d: %r', len(proc_chunk), proc_chunk
[:30])
self.log.debug('writing chunk to process')
try:
os.write(self.stream, proc_chunk)
except OSError:
self.log.debug('OSError writing stdin chunk')
return True
| write | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n try:\n chunk = self.get_chunk()\n if chunk is None:\n raise DoneReadingForever\n except DoneReadingForever:\n se... |
def x(self, mod_fullname, path=None):
parent_frame = inspect.currentframe().f_back
while in_importlib(parent_frame):
parent_frame = parent_frame.f_back
module = fetch_module_from_frame(mod_fullname, parent_frame)
if not module:
return None
if module.__class__.__name__ not in self.restrict_to:
return None
return self
| find_module | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, mod_fullname, path=None):\n parent_frame = inspect.currentframe().f_back\n while in_importlib(parent_frame):\n parent_frame = parent_frame.f... |
def x(ds, prefix='', sep='.', level=0):
singleton = False
if isinstance(ds, dict):
ds = [ds]
singleton = True
new_ds = []
for d in ds:
new_d = copy.deepcopy(d)
for k, v in d.items():
if not isinstance(k, compat.string_types):
k = str(k)
if level == 0:
newkey = k
else:
newkey = prefix + sep + k
if not isinstance(v, dict):
if level != 0:
v = new_d.pop(k)
new_d[newkey] = v
continue
else:
v = new_d.pop(k)
new_d.update(nested_to_record(v, newkey, sep, level + 1))
new_ds.append(new_d)
if singleton:
return new_ds[0]
return new_ds
| nested_to_record | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(ds, prefix='', sep='.', level=0):\n singleton = False\n if isinstance(ds, dict):\n ds = [ds]\n singleton = True\n new_ds = []\n for d... |
def x(data, record_path=None, meta=None, meta_prefix=None, record_prefix=
None, errors='raise', sep='.'):
def x(js, spec):
result = js
if isinstance(spec, list):
for field in spec:
result = result[field]
else:
result = result[spec]
return result
if isinstance(data, list) and len(data) is 0:
return DataFrame()
if isinstance(data, dict):
data = [data]
if record_path is None:
if any([isinstance(x, dict) for x in compat.itervalues(data[0])]):
data = nested_to_record(data, sep=sep)
return DataFrame(data)
elif not isinstance(record_path, list):
record_path = [record_path]
if meta is None:
meta = []
elif not isinstance(meta, list):
meta = [meta]
for i, x in enumerate(meta):
if not isinstance(x, list):
meta[i] = [x]
records = []
lengths = []
meta_vals = defaultdict(list)
if not isinstance(sep, compat.string_types):
sep = str(sep)
meta_keys = [sep.join(val) for val in meta]
def x(data, path, seen_meta, level=0):
if len(path) > 1:
for obj in data:
for val, key in zip(meta, meta_keys):
if level + 1 == len(val):
seen_meta[key] = _pull_field(obj, val[-1])
_recursive_extract(obj[path[0]], path[1:], seen_meta, level
=level + 1)
else:
for obj in data:
recs = _pull_field(obj, path[0])
lengths.append(len(recs))
for val, key in zip(meta, meta_keys):
if level + 1 > len(val):
meta_val = seen_meta[key]
else:
try:
meta_val = _pull_field(obj, val[level:])
except KeyError as e:
if errors == 'ignore':
meta_val = np.nan
else:
raise KeyError(
"Try running with errors='ignore' as key %s is not always present"
, e)
meta_vals[key].append(meta_val)
records.extend(recs)
_recursive_extract(data, record_path, {}, level=0)
result = DataFrame(records)
if record_prefix is not None:
result.rename(columns=lambda x: record_prefix + x, inplace=True)
for k, v in compat.iteritems(meta_vals):
if meta_prefix is not None:
k = meta_prefix + k
if k in result:
raise ValueError(
'Conflicting metadata name %s, need distinguishing prefix ' % k
)
result[k] = np.array(v).repeat(lengths)
return result
| json_normalize | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(data, record_path=None, meta=None, meta_prefix=None, record_prefix=\n None, errors='raise', sep='.'):\n\n def x(js, spec):\n result = js\n ... |
def x(x):
if is_integer_dtype(x):
return 'integer'
elif is_bool_dtype(x):
return 'boolean'
elif is_numeric_dtype(x):
return 'number'
elif is_datetime64_dtype(x) or is_datetime64tz_dtype(x):
return 'datetime'
elif is_timedelta64_dtype(x):
return 'duration'
elif is_categorical_dtype(x):
return 'any'
elif is_string_dtype(x):
return 'string'
else:
return 'any'
| as_json_table_type | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(x):\n if is_integer_dtype(x):\n return 'integer'\n elif is_bool_dtype(x):\n return 'boolean'\n elif is_numeric_dtype(x):\n return... |
def x(data):
if all(name is not None for name in data.index.names):
return data
data = data.copy()
if data.index.nlevels > 1:
names = [(name if name is not None else 'level_{}'.format(i)) for i,
name in enumerate(data.index.names)]
data.index.names = names
else:
data.index.name = data.index.name or 'index'
return data
| set_default_names | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(data):\n if all(name is not None for name in data.index.names):\n return data\n data = data.copy()\n if data.index.nlevels > 1:\n names ... |
def x(path_or_buf=None, orient=None, typ='frame', dtype=True, convert_axes=
True, convert_dates=True, keep_default_dates=True, numpy=False,
precise_float=False, date_unit=None, encoding=None, lines=False):
filepath_or_buffer, _, _ = get_filepath_or_buffer(path_or_buf, encoding
=encoding)
if isinstance(filepath_or_buffer, compat.string_types):
try:
exists = os.path.exists(filepath_or_buffer)
except (TypeError, ValueError):
exists = False
if exists:
fh, handles = _get_handle(filepath_or_buffer, 'r', encoding=
encoding)
json = fh.read()
fh.close()
else:
json = filepath_or_buffer
elif hasattr(filepath_or_buffer, 'read'):
json = filepath_or_buffer.read()
else:
json = filepath_or_buffer
if lines:
lines = list(StringIO(json.strip()))
json = '[' + ','.join(lines) + ']'
obj = None
if typ == 'frame':
obj = FrameParser(json, orient, dtype, convert_axes, convert_dates,
keep_default_dates, numpy, precise_float, date_unit).parse()
if typ == 'series' or obj is None:
if not isinstance(dtype, bool):
dtype = dict(data=dtype)
obj = SeriesParser(json, orient, dtype, convert_axes, convert_dates,
keep_default_dates, numpy, precise_float, date_unit).parse()
return obj
| read_json | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(path_or_buf=None, orient=None, typ='frame', dtype=True, convert_axes=\n True, convert_dates=True, keep_default_dates=True, numpy=False,\n precise_float=F... |
def x(self, name, data, use_dtypes=True, convert_dates=True):
if use_dtypes:
if self.dtype is False:
return data, False
elif self.dtype is True:
pass
else:
dtype = self.dtype.get(name) if isinstance(self.dtype, dict
) else self.dtype
if dtype is not None:
try:
dtype = np.dtype(dtype)
return data.astype(dtype), True
except:
return data, False
if convert_dates:
new_data, result = self._try_convert_to_date(data)
if result:
return new_data, True
result = False
if data.dtype == 'object':
try:
data = data.astype('float64')
result = True
except:
pass
if data.dtype.kind == 'f':
if data.dtype != 'float64':
try:
data = data.astype('float64')
result = True
except:
pass
if len(data) and (data.dtype == 'float' or data.dtype == 'object'):
try:
new_data = data.astype('int64')
if (new_data == data).all():
data = new_data
result = True
except:
pass
if data.dtype == 'int':
try:
data = data.astype('int64')
result = True
except:
pass
return data, result
| try_convert_data | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, name, data, use_dtypes=True, convert_dates=True):\n if use_dtypes:\n if self.dtype is False:\n return data, False\n elif self... |
def x(self, data):
if not len(data):
return data, False
new_data = data
if new_data.dtype == 'object':
try:
new_data = data.astype('int64')
except:
pass
if issubclass(new_data.dtype.type, np.number):
in_range = isnull(new_data.values) | (new_data > self.min_stamp) | (
new_data.values == iNaT)
if not in_range.all():
return data, False
date_units = (self.date_unit,) if self.date_unit else self._STAMP_UNITS
for date_unit in date_units:
try:
new_data = to_datetime(new_data, errors='raise', unit=date_unit)
except ValueError:
continue
except:
break
return new_data, True
return data, False
| try_convert_to_date | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, data):\n if not len(data):\n return data, False\n new_data = data\n if new_data.dtype == 'object':\n try:\n new_data = ... |
def x(show_toolbar=None, precision=None, grid_options=None, column_options=None
):
defaults.set_defaults(show_toolbar=show_toolbar, precision=precision,
grid_options=grid_options, column_options=column_options)
| set_defaults | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(show_toolbar=None, precision=None, grid_options=None, column_options=None\n ):\n defaults.set_defaults(show_toolbar=show_toolbar, precision=precision,\n ... |
def x(dataframe=True, series=True):
try:
from IPython.core.getipython import get_ipython
except ImportError:
raise ImportError('This feature requires IPython 1.0+')
ip = get_ipython()
ip_formatter = ip.display_formatter.ipython_display_formatter
if dataframe:
ip_formatter.for_type(pd.DataFrame, _display_as_qgrid)
else:
ip_formatter.type_printers.pop(pd.DataFrame, None)
if series:
ip_formatter.for_type(pd.Series, _display_as_qgrid)
else:
ip_formatter.type_printers.pop(pd.Series)
| enable | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(dataframe=True, series=True):\n try:\n from IPython.core.getipython import get_ipython\n except ImportError:\n raise ImportError('This feat... |
def x(data_frame, show_toolbar=None, precision=None, grid_options=None,
column_options=None, column_definitions=None, row_edit_callback=None):
if show_toolbar is None:
show_toolbar = defaults.show_toolbar
if precision is None:
precision = defaults.precision
if not isinstance(precision, Integral):
raise TypeError('precision must be int, not %s' % type(precision))
if column_options is None:
column_options = defaults.column_options
else:
options = defaults.column_options.copy()
options.update(column_options)
column_options = options
if grid_options is None:
grid_options = defaults.grid_options
else:
options = defaults.grid_options.copy()
options.update(grid_options)
grid_options = options
if not isinstance(grid_options, dict):
raise TypeError('grid_options must be dict, not %s' % type(
grid_options))
if isinstance(data_frame, pd.Series):
data_frame = pd.DataFrame(data_frame)
elif not isinstance(data_frame, pd.DataFrame):
raise TypeError('data_frame must be DataFrame or Series, not %s' %
type(data_frame))
column_definitions = column_definitions or {}
return QgridWidget(df=data_frame, precision=precision, grid_options=
grid_options, column_options=column_options, column_definitions=
column_definitions, row_edit_callback=row_edit_callback,
show_toolbar=show_toolbar)
| show_grid | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(data_frame, show_toolbar=None, precision=None, grid_options=None,\n column_options=None, column_definitions=None, row_edit_callback=None):\n if show_tool... |
def x(command, strict=False):
class DecoratedCommand(command):
def x(self):
jsdeps = self.distribution.get_command_obj('jsdeps')
if not is_repo and all(exists(t) for t in jsdeps.targets):
command.run(self)
return
try:
self.distribution.run_command('jsdeps')
except Exception as e:
missing = [t for t in jsdeps.targets if not exists(t)]
if strict or missing:
log.warn('rebuilding js and css failed')
if missing:
log.error('missing files: %s' % missing)
raise e
else:
log.warn('rebuilding js and css failed (not a problem)')
log.warn(str(e))
command.run(self)
update_package_data(self.distribution)
return DecoratedCommand
| js_prerelease | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(command, strict=False):\n\n\n class DecoratedCommand(command):\n\n def x(self):\n jsdeps = self.distribution.get_command_obj('jsdeps')\n ... |
def x():
styles_dir = os.path.join(package_dir, 'styles')
themes = [os.path.basename(theme).replace('.less', '') for theme in
glob('{0}/*.less'.format(styles_dir))]
return themes
| get_themes | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x():\n styles_dir = os.path.join(package_dir, 'styles')\n themes = [os.path.basename(theme).replace('.less', '') for theme in\n glob('{0}/*.less'.fo... |
def x(theme=None, monofont=None, monosize=11, nbfont=None, nbfontsize=13,
tcfont=None, tcfontsize=13, dffontsize=93, outfontsize=85, mathfontsize
=100, margins='auto', cellwidth='980', lineheight=170, cursorwidth=2,
cursorcolor='default', altprompt=False, altmd=False, altout=False,
hideprompt=False, vimext=False, toolbar=False, nbname=False, kernellogo
=False, dfonts=False):
wkdir = os.path.abspath('./')
stylefx.reset_default(False)
stylefx.check_directories()
doc = '\nConcatenated font imports, .less styles, & custom variables\n'
s = '*' * 65
style_less = '\n'.join(['/*', s, s, doc, s, s, '*/'])
style_less += '\n\n\n'
style_less += '/* Import Notebook, Markdown, & Code Fonts */\n'
style_less = stylefx.set_font_properties(style_less=style_less,
monofont=monofont, monosize=monosize, nbfont=nbfont, nbfontsize=
nbfontsize, tcfont=tcfont, tcfontsize=tcfontsize, dffontsize=
dffontsize, outfontsize=outfontsize, mathfontsize=mathfontsize,
dfonts=dfonts)
if theme is not None:
cursorcolor = stylefx.get_colors(theme=theme, c=cursorcolor)
style_less = stylefx.style_layout(style_less, theme=theme,
cellwidth=cellwidth, margins=margins, lineheight=lineheight,
altprompt=altprompt, altmd=altmd, altout=altout, hideprompt=
hideprompt, cursorwidth=cursorwidth, cursorcolor=cursorcolor,
vimext=vimext, toolbar=toolbar, nbname=nbname, kernellogo=
kernellogo)
style_css = stylefx.less_to_css(style_less)
style_css = stylefx.set_mathjax_style(style_css, mathfontsize)
stylefx.write_final_css(style_css)
os.chdir(wkdir)
| install_theme | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(theme=None, monofont=None, monosize=11, nbfont=None, nbfontsize=13,\n tcfont=None, tcfontsize=13, dffontsize=93, outfontsize=85, mathfontsize\n =100, mar... |
def x():
themes = [os.path.basename(theme).replace('.less', '') for theme in
glob('{0}/*.less'.format(styles_dir))]
if os.path.exists(theme_name_file):
with open(theme_name_file) as f:
theme = f.readlines()[0]
if theme not in themes:
theme = 'default'
else:
theme = 'default'
return theme
| infer_theme | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x():\n themes = [os.path.basename(theme).replace('.less', '') for theme in\n glob('{0}/*.less'.format(styles_dir))]\n if os.path.exists(theme_name_f... |
def x(theme=None, context='paper', grid=True, gridlines=u'-', ticks=False,
spines=True, fscale=1.2, figsize=(8.0, 7.0)):
rcdict = set_context(context=context, fscale=fscale, figsize=figsize)
if theme is None:
theme = infer_theme()
set_style(rcdict, theme=theme, grid=grid, gridlines=gridlines, ticks=
ticks, spines=spines)
| style | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(theme=None, context='paper', grid=True, gridlines=u'-', ticks=False,\n spines=True, fscale=1.2, figsize=(8.0, 7.0)):\n rcdict = set_context(context=conte... |
def x(rcdict, theme=None, grid=True, gridlines=u'-', ticks=False, spines=True):
styleMap, clist = get_theme_style(theme)
figureFace = styleMap['figureFace']
axisFace = styleMap['axisFace']
textColor = styleMap['textColor']
edgeColor = styleMap['edgeColor']
gridColor = styleMap['gridColor']
if not spines:
edgeColor = 'none'
style_dict = {'figure.edgecolor': figureFace, 'figure.facecolor':
figureFace, 'axes.facecolor': axisFace, 'axes.edgecolor': edgeColor,
'axes.labelcolor': textColor, 'axes.grid': grid, 'grid.linestyle':
gridlines, 'grid.color': gridColor, 'text.color': textColor,
'xtick.color': textColor, 'ytick.color': textColor,
'patch.edgecolor': axisFace, 'patch.facecolor': gridColor,
'savefig.facecolor': figureFace, 'savefig.edgecolor': figureFace}
rcdict.update(style_dict)
if ticks:
rcdict.update({'xtick.major.size': 6, 'ytick.major.size': 6,
'xtick.minor.size': 3, 'ytick.minor.size': 3})
base_style.update(rcdict)
mpl.rcParams.update(rcdict)
try:
import seaborn as sns
sns.set_style(rc=rcdict)
except Exception:
pass
try:
from cycler import cycler
mpl.rcParams['axes.prop_cycle'] = cycler(color=clist)
except Exception:
pass
for code, color in zip('bgrmyck', clist[:7]):
rgb = mpl.colors.colorConverter.to_rgb(color)
mpl.colors.colorConverter.colors[code] = rgb
mpl.colors.colorConverter.cache[code] = rgb
| set_style | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(rcdict, theme=None, grid=True, gridlines=u'-', ticks=False, spines=True):\n styleMap, clist = get_theme_style(theme)\n figureFace = styleMap['figureFace'... |
def x(context='paper', fscale=1.0, figsize=(8.0, 7.0)):
scaling = dict(paper=0.8, notebook=1, talk=1.3, poster=1.6)[context]
context_dict = {k: (v * scaling) for k, v in base_context.items()}
figX, figY = figsize
context_dict['figure.figsize'] = figX * scaling, figY * scaling
font_dict = {k: (v * fscale) for k, v in base_font.items()}
font_dict['font.family'] = ['sans-serif']
font_dict['font.sans-serif'] = ['Helvetica', 'Helvetica Neue', 'Arial',
'DejaVu Sans', 'Liberation Sans', 'sans-serif']
context_dict.update(font_dict)
return context_dict
| set_context | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(context='paper', fscale=1.0, figsize=(8.0, 7.0)):\n scaling = dict(paper=0.8, notebook=1, talk=1.3, poster=1.6)[context]\n context_dict = {k: (v * scalin... |
def x(x=8, y=7.0, aspect=1.0):
mpl.rcParams.update({'figure.figsize': (x * aspect, y)})
| figsize | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(x=8, y=7.0, aspect=1.0):\n mpl.rcParams.update({'figure.figsize': (x * aspect, y)})\n",
"role": "user"
},
{
"content": "figsize",
"role": "a... |
def x(theme):
styleMap, clist = get_default_jtstyle()
if theme == 'default':
return styleMap, clist
syntaxVars = ['@yellow:', '@orange:', '@red:', '@magenta:', '@violet:',
'@blue:', '@cyan:', '@green:']
get_hex_code = lambda line: line.split(':')[-1].split(';')[0][-7:]
themeFile = os.path.join(styles_dir, theme + '.less')
with open(themeFile) as f:
for line in f:
for k, v in styleMap.items():
if k in line.strip():
styleMap[k] = get_hex_code(line)
for c in syntaxVars:
if c in line.strip():
syntaxVars[syntaxVars.index(c)] = get_hex_code(line)
syntaxVars = list(set(syntaxVars))
clist.extend(syntaxVars)
clist = remove_non_colors(clist)
return styleMap, clist
| get_theme_style | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(theme):\n styleMap, clist = get_default_jtstyle()\n if theme == 'default':\n return styleMap, clist\n syntaxVars = ['@yellow:', '@orange:', '@r... |
def x():
colors = [(0.0, 0.0, 1.0), (0.0, 0.5, 0.0), (1.0, 0.0, 0.0), (0.75,
0.75, 0.0), (0.75, 0.75, 0.0), (0.0, 0.75, 0.75), (0.0, 0.0, 0.0)]
for code, color in zip('bgrmyck', colors):
rgb = mpl.colors.colorConverter.to_rgb(color)
mpl.colors.colorConverter.colors[code] = rgb
mpl.colors.colorConverter.cache[code] = rgb
mpl.rcParams.update(mpl.rcParamsDefault)
mpl.rcParams['figure.facecolor'] = 'white'
mpl.rcParams['axes.facecolor'] = 'white'
| reset | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x():\n colors = [(0.0, 0.0, 1.0), (0.0, 0.5, 0.0), (1.0, 0.0, 0.0), (0.75, \n 0.75, 0.0), (0.75, 0.75, 0.0), (0.0, 0.75, 0.75), (0.0, 0.0, 0.0)]\n f... |
def x(style_less):
with fileOpen(tempfile, 'w') as f:
f.write(style_less)
os.chdir(package_dir)
style_css = lesscpy.compile(tempfile)
style_css += '\n\n'
return style_css
| less_to_css | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(style_less):\n with fileOpen(tempfile, 'w') as f:\n f.write(style_less)\n os.chdir(package_dir)\n style_css = lesscpy.compile(tempfile)\n st... |
def x(style_less, nbfont=None, tcfont=None, monofont=None, monosize=11,
tcfontsize=13, nbfontsize=13, prfontsize=95, dffontsize=93, outfontsize
=85, mathfontsize=100, dfonts=False):
fontsizes = [monosize, nbfontsize, tcfontsize, prfontsize, dffontsize,
outfontsize]
(monosize, nbfontsize, tcfontsize, prfontsize, dffontsize, outfontsize
) = convert_fontsizes(fontsizes)
if dfonts == True:
monofont, tcfont, nbfont = ['monospace', 'sans-serif', 'sans-serif']
else:
if monofont is not None:
monofont, monofpath = stored_font_dicts(monofont)
style_less = import_fonts(style_less, monofont, monofpath)
else:
monofont = 'monospace'
if tcfont is not None:
tcfont, tcfontpath = stored_font_dicts(tcfont)
style_less = import_fonts(style_less, tcfont, tcfontpath)
else:
tcfont = 'sans-serif'
if nbfont is not None:
if nbfont == 'proxima':
nbfont, tcfont = ["'Proxima Nova'"] * 2
style_less = proxima_nova_imports(style_less)
else:
nbfont, nbfontpath = stored_font_dicts(nbfont)
style_less = import_fonts(style_less, nbfont, nbfontpath)
else:
nbfont = 'sans-serif'
style_less += '/* Set Font-Type and Font-Size Variables */\n'
style_less += '@monofont: {}; \n'.format(monofont)
style_less += '@notebook-fontfamily: {}; \n'.format(nbfont)
style_less += '@text-cell-fontfamily: {}; \n'.format(tcfont)
style_less += '@monofontsize: {}pt; \n'.format(monosize)
style_less += '@monofontsize-sub: {}pt; \n'.format(float(monosize) - 1)
style_less += '@nb-fontsize: {}pt; \n'.format(nbfontsize)
style_less += '@nb-fontsize-sub: {}pt; \n'.format(float(nbfontsize) - 1)
style_less += '@text-cell-fontsize: {}pt; \n'.format(tcfontsize)
style_less += '@df-header-fontsize: {}pt; \n'.format(float(dffontsize) + 1)
style_less += '@df-fontsize: {}pt; \n'.format(dffontsize)
style_less += '@output-font-size: {}pt; \n'.format(outfontsize)
style_less += '@prompt-fontsize: {}pt; \n'.format(prfontsize)
style_less += '@mathfontsize: {}%; \n'.format(mathfontsize)
style_less += '\n\n'
style_less += '/* Import Theme Colors and Define Layout Variables */\n'
return style_less
| set_font_properties | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(style_less, nbfont=None, tcfont=None, monofont=None, monosize=11,\n tcfontsize=13, nbfontsize=13, prfontsize=95, dffontsize=93, outfontsize\n =85, mathfo... |
def x(style_less, fontname, font_subdir):
ftype_dict = {'woff2': 'woff2', 'woff': 'woff', 'ttf': 'truetype',
'otf': 'opentype', 'svg': 'svg'}
define_font = """@font-face {{font-family: {fontname};
font-weight:{weight};
font-style: {style};
src: local('{fontname}'),
url('fonts{sepp}{fontfile}') format('{ftype}');}}
"""
fontname = fontname.split(',')[0]
fontpath = os.path.join(fonts_dir, font_subdir)
for fontfile in os.listdir(fontpath):
if '.txt' in fontfile or 'DS_' in fontfile:
continue
weight = 'normal'
style = 'normal'
if 'medium' in fontfile:
weight = 'medium'
elif 'ital' in fontfile:
style = 'italic'
ft = ftype_dict[fontfile.split('.')[-1]]
style_less += define_font.format(fontname=fontname, weight=weight,
style=style, sepp='/', fontfile=fontfile, ftype=ft)
send_fonts_to_jupyter(os.path.join(fontpath, fontfile))
return style_less
| import_fonts | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(style_less, fontname, font_subdir):\n ftype_dict = {'woff2': 'woff2', 'woff': 'woff', 'ttf': 'truetype',\n 'otf': 'opentype', 'svg': 'svg'}\n defi... |
def x(style_less, theme='grade3', cursorwidth=2, cursorcolor='default',
cellwidth='980', lineheight=170, margins='auto', vimext=False, toolbar=
False, nbname=False, kernellogo=False, altprompt=False, altmd=False,
altout=False, hideprompt=False):
with fileOpen(theme_name_file, 'w') as f:
f.write(theme)
if os.path.isdir(styles_dir_user) and '{}.less'.format(theme
) in os.listdir(styles_dir_user):
theme_relpath = os.path.relpath(os.path.join(styles_dir_user, theme
), package_dir)
else:
theme_relpath = os.path.relpath(os.path.join(styles_dir, theme),
package_dir)
style_less += '@import "{}";\n'.format(theme_relpath)
textcell_bg = '@cc-input-bg'
promptText = '@input-prompt'
promptBG = '@cc-input-bg'
promptPadding = '.25em'
promptBorder = '2px solid @prompt-line'
tcPromptBorder = '2px solid @tc-prompt-std'
promptMinWidth = 11.5
outpromptMinWidth = promptMinWidth
tcPromptWidth = promptMinWidth + 3
tcPromptFontsize = '@prompt-fontsize'
ccOutputBG = '@cc-output-bg-default'
if theme == 'grade3':
textcell_bg = '@notebook-bg'
if altprompt:
promptPadding = '.1em'
promptMinWidth = 8
outpromptMinWidth = promptMinWidth + 3
tcPromptWidth = promptMinWidth + 3
promptText = 'transparent'
tcPromptBorder = '2px solid transparent'
if altmd:
textcell_bg = '@notebook-bg'
tcPromptBorder = '2px dotted @tc-border-selected'
if altout:
ccOutputBG = '@notebook-bg'
if margins != 'auto':
margins = '{}px'.format(margins)
if '%' not in cellwidth:
cellwidth = str(cellwidth) + 'px'
style_less += '@container-margins: {};\n'.format(margins)
style_less += '@cell-width: {}; \n'.format(cellwidth)
style_less += '@cc-line-height: {}%; \n'.format(lineheight)
style_less += '@text-cell-bg: {}; \n'.format(textcell_bg)
style_less += '@cc-prompt-width: {}ex; \n'.format(promptMinWidth)
style_less += '@cc-prompt-bg: {}; \n'.format(promptBG)
style_less += '@cc-output-bg: {}; \n'.format(ccOutputBG)
style_less += '@prompt-text: {}; \n'.format(promptText)
style_less += '@prompt-padding: {}; \n'.format(promptPadding)
style_less += '@prompt-border: {}; \n'.format(promptBorder)
style_less += '@prompt-min-width: {}ex; \n'.format(promptMinWidth)
style_less += '@out-prompt-min-width: {}ex; \n'.format(outpromptMinWidth)
style_less += '@tc-prompt-width: {}ex; \n'.format(tcPromptWidth)
style_less += '@tc-prompt-border: {}; \n'.format(tcPromptBorder)
style_less += '@cursor-width: {}px; \n'.format(cursorwidth)
style_less += '@cursor-info: @cursor-width solid {}; \n'.format(cursorcolor
)
style_less += '@tc-prompt-fontsize: {}; \n'.format(tcPromptFontsize)
style_less += '\n\n'
with fileOpen(nb_style, 'r') as notebook:
style_less += notebook.read() + '\n'
with fileOpen(cl_style, 'r') as cells:
style_less += cells.read() + '\n'
with fileOpen(ex_style, 'r') as extras:
style_less += extras.read() + '\n'
with fileOpen(cm_style, 'r') as codemirror:
style_less += codemirror.read() + '\n'
with fileOpen(comp_style, 'r') as codemirror:
style_less += codemirror.read() + '\n'
style_less += toggle_settings(toolbar, nbname, hideprompt, kernellogo
) + '\n'
if vimext:
set_vim_style(theme)
return style_less
| style_layout | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(style_less, theme='grade3', cursorwidth=2, cursorcolor='default',\n cellwidth='980', lineheight=170, margins='auto', vimext=False, toolbar=\n False, nbna... |
def x(toolbar=False, nbname=False, hideprompt=False, kernellogo=False):
toggle = ''
if toolbar:
toggle += 'div#maintoolbar {margin-left: 8px !important;}\n'
toggle += '.toolbar.container {width: 100% !important;}\n'
else:
toggle += 'div#maintoolbar {display: none !important;}\n'
if nbname:
toggle += """span.save_widget span.filename {margin-left: 8px; height: initial;font-size: 100%; color: @nb-name-fg; background-color:@cc-input-bg;}
"""
toggle += """span.save_widget span.filename:hover {color:@nb-name-hover; background-color: @cc-input-bg;}
"""
toggle += (
'#menubar {padding-top: 4px; background-color:@notebook-bg;}\n')
else:
toggle += '#header-container {display: none !important;}\n'
if hideprompt:
toggle += 'div.prompt.input_prompt {display: none !important;}\n'
toggle += 'div.prompt.output_prompt {width: 5ex !important;}\n'
toggle += """div.out_prompt_overlay.prompt:hover {width: 5ex !important; min-width: 5ex !important;}
"""
toggle += """.CodeMirror-gutters, .cm-s-ipython .CodeMirror-gutters{ position: absolute; left: 0; top: 0; z-index: 3; width: 2em; display: inline-block !important; }
"""
toggle += """div.cell.code_cell .input { border-left: 5px solid @cm-gutters !important; border-bottom-left-radius: 5px; border-top-left-radius: 5px; }
"""
if kernellogo:
toggle += '@kernel-logo-display: block;'
else:
toggle += '@kernel-logo-display: none;'
return toggle
| toggle_settings | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(toolbar=False, nbname=False, hideprompt=False, kernellogo=False):\n toggle = ''\n if toolbar:\n toggle += 'div#maintoolbar {margin-left: 8px !impo... |
def x(style_css, mathfontsize):
jax_style = (
"""<script>
MathJax.Hub.Config({
"HTML-CSS": {
/*preferredFont: "TeX",*/
/*availableFonts: ["TeX", "STIX"],*/
styles: {
scale: %d,
".MathJax_Display": {
"font-size": %s,
}
}
}
});
</script>
"""
% (int(mathfontsize), '"{}%"'.format(str(mathfontsize))))
style_css += jax_style
return style_css
| set_mathjax_style | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(style_css, mathfontsize):\n jax_style = (\n \"\"\"<script>\n MathJax.Hub.Config({\n \"HTML-CSS\": {\n /*preferredFont: \"TeX\",*... |
def x(theme):
vim_jupyter_nbext = os.path.join(jupyter_nbext, 'vim_binding')
if not os.path.isdir(vim_jupyter_nbext):
os.makedirs(vim_jupyter_nbext)
vim_less = '@import "styles{}";\n'.format(''.join([os.sep, theme]))
with open(vim_style, 'r') as vimstyle:
vim_less += vimstyle.read() + '\n'
with open(vimtemp, 'w') as vtemp:
vtemp.write(vim_less)
os.chdir(package_dir)
vim_css = lesscpy.compile(vimtemp)
vim_css += '\n\n'
vim_custom_css = os.path.join(vim_jupyter_nbext, 'vim_binding.css')
with open(vim_custom_css, 'w') as vim_custom:
vim_custom.write(vim_css)
| set_vim_style | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(theme):\n vim_jupyter_nbext = os.path.join(jupyter_nbext, 'vim_binding')\n if not os.path.isdir(vim_jupyter_nbext):\n os.makedirs(vim_jupyter_nbex... |
def x(verbose=False):
paths = [jupyter_custom, jupyter_nbext]
for fpath in paths:
custom = '{0}{1}{2}.css'.format(fpath, os.sep, 'custom')
try:
os.remove(custom)
except Exception:
pass
try:
delete_font_files()
except Exception:
check_directories()
delete_font_files()
copyfile(defaultCSS, jupyter_customcss)
copyfile(defaultJS, jupyter_customjs)
if os.path.exists(theme_name_file):
os.remove(theme_name_file)
if verbose:
print('Reset css and font defaults in:\n{} &\n{}'.format(*paths))
| reset_default | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(verbose=False):\n paths = [jupyter_custom, jupyter_nbext]\n for fpath in paths:\n custom = '{0}{1}{2}.css'.format(fpath, os.sep, 'custom')\n ... |
def x(name):
from IPython.core.display import HTML
styles_dir = os.path.join(package_dir, 'styles/compiled/')
css_path = glob('{0}/{1}.css'.format(styles_dir, name))[0]
customcss = open(css_path, 'r').read()
return HTML(''.join(['<style> ', customcss, ' </style>']))
| set_nb_theme | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(name):\n from IPython.core.display import HTML\n styles_dir = os.path.join(package_dir, 'styles/compiled/')\n css_path = glob('{0}/{1}.css'.format(sty... |
def x(readme_path, before_key, after_key, options_key, end_key):
with open(readme_path) as readme_file:
readme = readme_file.read()
top, rest = readme.split(before_key)
before, rest = rest.split(after_key)
_, rest = rest.split(options_key)
_, bottom = rest.split(end_key)
return top.rstrip('\n'), before.strip('\n'
), end_key + '\n\n' + bottom.lstrip('\n')
| split_readme | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(readme_path, before_key, after_key, options_key, end_key):\n with open(readme_path) as readme_file:\n readme = readme_file.read()\n top, rest = re... |
def x():
parser = autopep8.create_parser()
string_io = io.StringIO()
parser.print_help(string_io)
return string_io.getvalue().replace(os.path.expanduser('~'), '~')
| help_message | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x():\n parser = autopep8.create_parser()\n string_io = io.StringIO()\n parser.print_help(string_io)\n return string_io.getvalue().replace(os.path.exp... |
def x(source):
compile(source, '<string>', 'exec', dont_inherit=True)
reporter = pyflakes.reporter.Reporter(sys.stderr, sys.stderr)
pyflakes.api.check(source, filename='<string>', reporter=reporter)
| check | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(source):\n compile(source, '<string>', 'exec', dont_inherit=True)\n reporter = pyflakes.reporter.Reporter(sys.stderr, sys.stderr)\n pyflakes.api.check... |
def x(logical_line, tokens, indent_level, hang_closing, indent_char, noqa):
first_row = tokens[0][2][0]
nrows = 1 + tokens[-1][2][0] - first_row
if noqa or nrows == 1:
return
indent_next = logical_line.endswith(':')
row = depth = 0
valid_hangs = (DEFAULT_INDENT_SIZE,) if indent_char != '\t' else (
DEFAULT_INDENT_SIZE, 2 * DEFAULT_INDENT_SIZE)
parens = [0] * nrows
rel_indent = [0] * nrows
open_rows = [[0]]
hangs = [None]
indent_chances = {}
last_indent = tokens[0][2]
indent = [last_indent[1]]
last_token_multiline = None
line = None
last_line = ''
last_line_begins_with_multiline = False
for token_type, text, start, end, line in tokens:
newline = row < start[0] - first_row
if newline:
row = start[0] - first_row
newline = not last_token_multiline and token_type not in (tokenize
.NL, tokenize.NEWLINE)
last_line_begins_with_multiline = last_token_multiline
if newline:
last_indent = start
rel_indent[row] = pycodestyle.expand_indent(line) - indent_level
close_bracket = token_type == tokenize.OP and text in ']})'
for open_row in reversed(open_rows[depth]):
hang = rel_indent[row] - rel_indent[open_row]
hanging_indent = hang in valid_hangs
if hanging_indent:
break
if hangs[depth]:
hanging_indent = hang == hangs[depth]
visual_indent = (not close_bracket and hang > 0 and
indent_chances.get(start[1]))
if close_bracket and indent[depth]:
if start[1] != indent[depth]:
yield start, 'E124 {}'.format(indent[depth])
elif close_bracket and not hang:
if hang_closing:
yield start, 'E133 {}'.format(indent[depth])
elif indent[depth] and start[1] < indent[depth]:
yield start, 'E128 {}'.format(indent[depth])
elif hanging_indent or indent_next and rel_indent[row
] == 2 * DEFAULT_INDENT_SIZE:
if close_bracket and not hang_closing:
yield start, 'E123 {}'.format(indent_level + rel_indent
[open_row])
hangs[depth] = hang
elif visual_indent is True:
indent[depth] = start[1]
elif visual_indent in (text, unicode):
pass
else:
one_indented = indent_level + rel_indent[open_row
] + DEFAULT_INDENT_SIZE
if hang <= 0:
error = 'E122', one_indented
elif indent[depth]:
error = 'E127', indent[depth]
elif not close_bracket and hangs[depth]:
error = 'E131', one_indented
elif hang > DEFAULT_INDENT_SIZE:
error = 'E126', one_indented
else:
hangs[depth] = hang
error = 'E121', one_indented
yield start, '{} {}'.format(*error)
if parens[row] and token_type not in (tokenize.NL, tokenize.COMMENT
) and not indent[depth]:
indent[depth] = start[1]
indent_chances[start[1]] = True
elif token_type in (tokenize.STRING, tokenize.COMMENT) or text in ('u',
'ur', 'b', 'br'):
indent_chances[start[1]] = unicode
elif not indent_chances and not row and not depth and text == 'if':
indent_chances[end[1] + 1] = True
elif text == ':' and line[end[1]:].isspace():
open_rows[depth].append(row)
if token_type == tokenize.OP:
if text in '([{':
depth += 1
indent.append(0)
hangs.append(None)
if len(open_rows) == depth:
open_rows.append([])
open_rows[depth].append(row)
parens[row] += 1
elif text in ')]}' and depth > 0:
prev_indent = indent.pop() or last_indent[1]
hangs.pop()
for d in range(depth):
if indent[d] > prev_indent:
indent[d] = 0
for ind in list(indent_chances):
if ind >= prev_indent:
del indent_chances[ind]
del open_rows[depth + 1:]
depth -= 1
if depth:
indent_chances[indent[depth]] = True
for idx in range(row, -1, -1):
if parens[idx]:
parens[idx] -= 1
break
assert len(indent) == depth + 1
if start[1] not in indent_chances and not last_line.rstrip(
).endswith(','):
indent_chances[start[1]] = text
last_token_multiline = start[0] != end[0]
if last_token_multiline:
rel_indent[end[0] - first_row] = rel_indent[row]
last_line = line
if (indent_next and not last_line_begins_with_multiline and pycodestyle
.expand_indent(line) == indent_level + DEFAULT_INDENT_SIZE):
pos = start[0], indent[0] + 4
desired_indent = indent_level + 2 * DEFAULT_INDENT_SIZE
if visual_indent:
yield pos, 'E129 {}'.format(desired_indent)
else:
yield pos, 'E125 {}'.format(desired_indent)
| continued_indentation | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(logical_line, tokens, indent_level, hang_closing, indent_char, noqa):\n first_row = tokens[0][2][0]\n nrows = 1 + tokens[-1][2][0] - first_row\n if no... |
def x(tokens):
valid = ['\n', '\\', "'", '"', 'a', 'b', 'f', 'n', 'r', 't', 'v', '0',
'1', '2', '3', '4', '5', '6', '7', 'x', 'N', 'u', 'U']
for token_type, text, start_pos, _end_pos, _line in tokens:
if token_type == tokenize.STRING:
quote = text[-3:] if text[-3:] in ('"""', "'''") else text[-1]
quote_pos = text.index(quote)
prefix = text[:quote_pos].lower()
start = quote_pos + len(quote)
string = text[start:-len(quote)]
if 'r' not in prefix:
pos = string.find('\\')
while pos >= 0:
pos += 1
if string[pos] not in valid:
yield start_pos[1
], "W605 invalid escape sequence '\\%s'" % string[
pos]
pos = string.find('\\', pos + 1)
| get_w605_position | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(tokens):\n valid = ['\\n', '\\\\', \"'\", '\"', 'a', 'b', 'f', 'n', 'r', 't', 'v', '0',\n '1', '2', '3', '4', '5', '6', '7', 'x', 'N', 'u', 'U']\n ... |
def x(source, import_line_index):
def x(line):
if line[0] in 'uUbB':
line = line[1:]
if line and line[0] in 'rR':
line = line[1:]
return line and (line[0] == '"' or line[0] == "'")
def x(line):
nodes = ast.parse(line)
for n in nodes.body:
if isinstance(n, ast.ImportFrom) and n.module == '__future__':
return True
return False
allowed_try_keywords = 'try', 'except', 'else', 'finally'
in_docstring = False
docstring_kind = '"""'
for cnt, line in enumerate(source):
if not in_docstring:
m = DOCSTRING_START_REGEX.match(line.lstrip())
if m is not None:
in_docstring = True
docstring_kind = m.group('kind')
remain = line[m.end():m.endpos].rstrip()
if remain[-3:] == docstring_kind:
in_docstring = False
continue
if in_docstring:
if line.rstrip()[-3:] == docstring_kind:
in_docstring = False
continue
if not line.rstrip():
continue
elif line.startswith('#'):
continue
if line.startswith('import ') or line.startswith('from '):
if cnt == import_line_index or is_future_import(line):
continue
return cnt
elif pycodestyle.DUNDER_REGEX.match(line):
continue
elif any(line.startswith(kw) for kw in allowed_try_keywords):
continue
elif is_string_literal(line):
return cnt
else:
return cnt
return 0
| get_module_imports_on_top_of_file | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(source, import_line_index):\n\n def x(line):\n if line[0] in 'uUbB':\n line = line[1:]\n if line and line[0] in 'rR':\n ... |
def x(target, previous_line, original, indent_word=' ', max_line_length=
79, aggressive=False, experimental=False, verbose=False):
indent = _get_indentation(target)
source = target[len(indent):]
assert source.lstrip() == source
assert not target.lstrip().startswith('#')
tokens = list(generate_tokens(source))
candidates = shorten_line(tokens, source, indent, indent_word,
max_line_length, aggressive=aggressive, experimental=experimental,
previous_line=previous_line)
candidates = sorted(sorted(set(candidates).union([target, original])),
key=lambda x: line_shortening_rank(x, indent_word, max_line_length,
experimental=experimental))
if verbose >= 4:
print(('-' * 79 + '\n').join([''] + candidates + ['']), file=
wrap_output(sys.stderr, 'utf-8'))
if candidates:
best_candidate = candidates[0]
if longest_line_length(best_candidate) > longest_line_length(original):
return None
return best_candidate
| get_fixed_long_line | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(target, previous_line, original, indent_word=' ', max_line_length=\n 79, aggressive=False, experimental=False, verbose=False):\n indent = _get_indenta... |
def x(logical_line):
indentation = _get_indentation(logical_line)
return indentation + untokenize_without_newlines(generate_tokens(
logical_line.lstrip())) + '\n'
| join_logical_line | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(logical_line):\n indentation = _get_indentation(logical_line)\n return indentation + untokenize_without_newlines(generate_tokens(\n logical_line.l... |
def x(tokens):
text = ''
last_row = 0
last_column = -1
for t in tokens:
token_string = t[1]
start_row, start_column = t[2]
end_row, end_column = t[3]
if start_row > last_row:
last_column = 0
if (start_column > last_column or token_string == '\n'
) and not text.endswith(' '):
text += ' '
if token_string != '\n':
text += token_string
last_row = end_row
last_column = end_column
return text.rstrip()
| untokenize_without_newlines | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(tokens):\n text = ''\n last_row = 0\n last_column = -1\n for t in tokens:\n token_string = t[1]\n start_row, start_column = t[2]\n ... |
def x(source_lines, result, logical_start, logical_end):
row = result['line'] - 1
col = result['column'] - 1
ls = None
le = None
for i in range(0, len(logical_start), 1):
assert logical_end
x = logical_end[i]
if x[0] > row or x[0] == row and x[1] > col:
le = x
ls = logical_start[i]
break
if ls is None:
return None
original = source_lines[ls[0]:le[0] + 1]
return ls, le, original
| get_logical | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(source_lines, result, logical_start, logical_end):\n row = result['line'] - 1\n col = result['column'] - 1\n ls = None\n le = None\n for i in ra... |
def x(a, b):
split_a = split_and_strip_non_empty_lines(a)
split_b = split_and_strip_non_empty_lines(b)
if len(split_a) != len(split_b):
return False
for index, _ in enumerate(split_a):
if ''.join(split_a[index].split()) != ''.join(split_b[index].split()):
return False
return True
| code_almost_equal | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(a, b):\n split_a = split_and_strip_non_empty_lines(a)\n split_b = split_and_strip_non_empty_lines(b)\n if len(split_a) != len(split_b):\n retur... |
def x(source, aggressive=True, select=None, ignore=None, filename='', where
='global', verbose=False):
if not aggressive:
return source
select = select or []
ignore = ignore or []
return refactor(source, code_to_2to3(select=select, ignore=ignore,
where=where, verbose=verbose), filename=filename)
| fix_2to3 | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(source, aggressive=True, select=None, ignore=None, filename='', where\n ='global', verbose=False):\n if not aggressive:\n return source\n selec... |
def x(source):
assert not isinstance(source, unicode)
counter = collections.defaultdict(int)
for line in source:
if line.endswith(CRLF):
counter[CRLF] += 1
elif line.endswith(CR):
counter[CR] += 1
elif line.endswith(LF):
counter[LF] += 1
return (sorted(counter, key=counter.get, reverse=True) or [LF])[0]
| find_newline | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(source):\n assert not isinstance(source, unicode)\n counter = collections.defaultdict(int)\n for line in source:\n if line.endswith(CRLF):\n ... |
def x(source):
indent_word = ' '
try:
for t in generate_tokens(source):
if t[0] == token.INDENT:
indent_word = t[1]
break
except (SyntaxError, tokenize.TokenError):
pass
return indent_word
| get_indentword | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(source):\n indent_word = ' '\n try:\n for t in generate_tokens(source):\n if t[0] == token.INDENT:\n indent_word = t[... |
def x(old, new, filename):
newline = '\n'
diff = difflib.unified_diff(old, new, 'original/' + filename, 'fixed/' +
filename, lineterm=newline)
text = ''
for line in diff:
text += line
if text and not line.endswith(newline):
text += newline + '\\ No newline at end of file' + newline
return text
| get_diff_text | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(old, new, filename):\n newline = '\\n'\n diff = difflib.unified_diff(old, new, 'original/' + filename, 'fixed/' +\n filename, lineterm=newline)\n ... |
def x(tokens, source, indentation, indent_word, max_line_length, aggressive
=False, experimental=False, previous_line=''):
for candidate in _shorten_line(tokens=tokens, source=source,
indentation=indentation, indent_word=indent_word, aggressive=
aggressive, previous_line=previous_line):
yield candidate
if aggressive:
for key_token_strings in SHORTEN_OPERATOR_GROUPS:
shortened = _shorten_line_at_tokens(tokens=tokens, source=
source, indentation=indentation, indent_word=indent_word,
key_token_strings=key_token_strings, aggressive=aggressive)
if shortened is not None and shortened != source:
yield shortened
if experimental:
for shortened in _shorten_line_at_tokens_new(tokens=tokens, source=
source, indentation=indentation, max_line_length=max_line_length):
yield shortened
| shorten_line | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(tokens, source, indentation, indent_word, max_line_length, aggressive\n =False, experimental=False, previous_line=''):\n for candidate in _shorten_line(t... |
def x(tokens, source, indentation, indent_word, aggressive=False,
previous_line=''):
for token_type, token_string, start_offset, end_offset in token_offsets(
tokens):
if (token_type == tokenize.COMMENT and not
is_probably_part_of_multiline(previous_line) and not
is_probably_part_of_multiline(source) and not source[
start_offset + 1:].strip().lower().startswith(('noqa',
'pragma:', 'pylint:'))):
first = source[:start_offset]
second = source[start_offset:]
yield indentation + second.strip(
) + '\n' + indentation + first.strip() + '\n'
elif token_type == token.OP and token_string != '=':
assert token_type != token.INDENT
first = source[:end_offset]
second_indent = indentation
if first.rstrip().endswith('(') and source[end_offset:].lstrip(
).startswith(')'):
pass
elif first.rstrip().endswith('('):
second_indent += indent_word
elif '(' in first:
second_indent += ' ' * (1 + first.find('('))
else:
second_indent += indent_word
second = second_indent + source[end_offset:].lstrip()
if not second.strip() or second.lstrip().startswith('#'):
continue
if second.lstrip().startswith(','):
continue
if first.rstrip().endswith('.'):
continue
if token_string in '+-*/':
fixed = first + ' \\' + '\n' + second
else:
fixed = first + '\n' + second
if check_syntax(normalize_multiline(fixed) if aggressive else fixed
):
yield indentation + fixed
| shorten_line | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(tokens, source, indentation, indent_word, aggressive=False,\n previous_line=''):\n for token_type, token_string, start_offset, end_offset in token_offset... |
def x(tokens, index, for_or_if=None):
items = [Atom(Token(*tokens[index]))]
index += 1
num_tokens = len(tokens)
while index < num_tokens:
tok = Token(*tokens[index])
if tok.token_string in ',)]}':
if for_or_if == 'for':
return ListComprehension(items), index - 1
elif for_or_if == 'if':
return IfExpression(items), index - 1
items.append(Atom(tok))
if tok.token_string == ')':
return Tuple(items), index
elif tok.token_string == ']':
return List(items), index
elif tok.token_string == '}':
return DictOrSet(items), index
elif tok.token_string in '([{':
container, index = _parse_container(tokens, index)
items.append(container)
elif tok.token_string == 'for':
container, index = _parse_container(tokens, index, 'for')
items.append(container)
elif tok.token_string == 'if':
container, index = _parse_container(tokens, index, 'if')
items.append(container)
else:
items.append(Atom(tok))
index += 1
return None, None
| parse_container | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(tokens, index, for_or_if=None):\n items = [Atom(Token(*tokens[index]))]\n index += 1\n num_tokens = len(tokens)\n while index < num_tokens:\n ... |
def x(tokens):
index = 0
parsed_tokens = []
num_tokens = len(tokens)
while index < num_tokens:
tok = Token(*tokens[index])
assert tok.token_type != token.INDENT
if tok.token_type == tokenize.NEWLINE:
break
if tok.token_string in '([{':
container, index = _parse_container(tokens, index)
if not container:
return None
parsed_tokens.append(container)
else:
parsed_tokens.append(Atom(tok))
index += 1
return parsed_tokens
| parse_tokens | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(tokens):\n index = 0\n parsed_tokens = []\n num_tokens = len(tokens)\n while index < num_tokens:\n tok = Token(*tokens[index])\n asse... |
def x(parsed_tokens, indentation, max_line_length, start_on_prefix_line):
if unicode(parsed_tokens[0]) == 'def':
continued_indent = indentation + ' ' * 2 * DEFAULT_INDENT_SIZE
else:
continued_indent = indentation + ' ' * DEFAULT_INDENT_SIZE
break_after_open_bracket = not start_on_prefix_line
lines = ReformattedLines(max_line_length)
lines.add_indent(len(indentation.lstrip('\r\n')))
if not start_on_prefix_line:
first_token = get_item(parsed_tokens, 0)
second_token = get_item(parsed_tokens, 1)
if first_token and second_token and unicode(second_token)[0
] == '(' and len(indentation) + len(first_token) + 1 == len(
continued_indent):
return None
for item in parsed_tokens:
lines.add_space_if_needed(unicode(item), equal=True)
save_continued_indent = continued_indent
if start_on_prefix_line and isinstance(item, Container):
start_on_prefix_line = False
continued_indent = ' ' * (lines.current_size() + 1)
item.reflow(lines, continued_indent, break_after_open_bracket)
continued_indent = save_continued_indent
return lines.emit()
| reflow_lines | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(parsed_tokens, indentation, max_line_length, start_on_prefix_line):\n if unicode(parsed_tokens[0]) == 'def':\n continued_indent = indentation + ' ' *... |
def x(tokens, source, indentation, max_line_length):
yield indentation + source
parsed_tokens = _parse_tokens(tokens)
if parsed_tokens:
fixed = _reflow_lines(parsed_tokens, indentation, max_line_length,
start_on_prefix_line=True)
if fixed and check_syntax(normalize_multiline(fixed.lstrip())):
yield fixed
fixed = _reflow_lines(parsed_tokens, indentation, max_line_length,
start_on_prefix_line=False)
if fixed and check_syntax(normalize_multiline(fixed.lstrip())):
yield fixed
| shorten_line_at_tokens_new | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(tokens, source, indentation, max_line_length):\n yield indentation + source\n parsed_tokens = _parse_tokens(tokens)\n if parsed_tokens:\n fixed... |
def x(tokens, source, indentation, indent_word, key_token_strings, aggressive):
offsets = []
for index, _t in enumerate(token_offsets(tokens)):
token_type, token_string, start_offset, end_offset = _t
assert token_type != token.INDENT
if token_string in key_token_strings:
unwanted_next_token = {'(': ')', '[': ']', '{': '}'}.get(
token_string)
if unwanted_next_token:
if get_item(tokens, index + 1, default=[None, None])[1
] == unwanted_next_token or get_item(tokens, index + 2,
default=[None, None])[1] == unwanted_next_token:
continue
if index > 2 and token_string == '(' and tokens[index - 1][1
] in ',(%[':
continue
if end_offset < len(source) - 1:
offsets.append(end_offset)
else:
previous_token = get_item(tokens, index - 1)
if (token_type == tokenize.STRING and previous_token and
previous_token[0] == tokenize.STRING):
offsets.append(start_offset)
current_indent = None
fixed = None
for line in split_at_offsets(source, offsets):
if fixed:
fixed += '\n' + current_indent + line
for symbol in '([{':
if line.endswith(symbol):
current_indent += indent_word
else:
fixed = line
assert not current_indent
current_indent = indent_word
assert fixed is not None
if check_syntax(normalize_multiline(fixed) if aggressive > 1 else fixed):
return indentation + fixed
return None
| shorten_line_at_tokens | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(tokens, source, indentation, indent_word, key_token_strings, aggressive):\n offsets = []\n for index, _t in enumerate(token_offsets(tokens)):\n to... |
def x(tokens):
end_offset = 0
previous_end_row = 0
previous_end_column = 0
for t in tokens:
token_type = t[0]
token_string = t[1]
start_row, start_column = t[2]
end_row, end_column = t[3]
end_offset += start_column
if previous_end_row == start_row:
end_offset -= previous_end_column
start_offset = end_offset
end_offset += len(token_string)
yield token_type, token_string, start_offset, end_offset
previous_end_row = end_row
previous_end_column = end_column
| token_offsets | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(tokens):\n end_offset = 0\n previous_end_row = 0\n previous_end_column = 0\n for t in tokens:\n token_type = t[0]\n token_string = t[... |
def x(pep8_options, source):
class QuietReport(pycodestyle.BaseReport):
"""Version of checker that does not print."""
def x(self, options):
super(QuietReport, self).__init__(options)
self.__full_error_results = []
def x(self, line_number, offset, text, check):
code = super(QuietReport, self).error(line_number, offset, text,
check)
if code:
self.__full_error_results.append({'id': code, 'line':
line_number, 'column': offset + 1, 'info': text})
def x(self):
return self.__full_error_results
checker = pycodestyle.Checker('', lines=source, reporter=QuietReport,
**pep8_options)
checker.check_all()
return checker.report.full_error_results()
| execute_pep8 | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(pep8_options, source):\n\n\n class QuietReport(pycodestyle.BaseReport):\n \"\"\"Version of checker that does not print.\"\"\"\n\n def x(self, ... |
def x(tokens):
find_stmt = 1
level = 0
stats = []
for t in tokens:
token_type = t[0]
sline = t[2][0]
line = t[4]
if token_type == tokenize.NEWLINE:
find_stmt = 1
elif token_type == tokenize.INDENT:
find_stmt = 1
level += 1
elif token_type == tokenize.DEDENT:
find_stmt = 1
level -= 1
elif token_type == tokenize.COMMENT:
if find_stmt:
stats.append((sline, -1))
elif token_type == tokenize.NL:
pass
elif find_stmt:
find_stmt = 0
if line:
stats.append((sline, level))
return stats
| reindent_stats | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(tokens):\n find_stmt = 1\n level = 0\n stats = []\n for t in tokens:\n token_type = t[0]\n sline = t[2][0]\n line = t[4]\n ... |
def x(line):
i = 0
while i < len(line) and line[i] == ' ':
i += 1
return i
| leading_space_count | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(line):\n i = 0\n while i < len(line) and line[i] == ' ':\n i += 1\n return i\n",
"role": "user"
},
{
"content": "leading_space_coun... |
def x(source_text, fixer_names, filename=''):
from lib2to3.refactor import RefactoringTool
fixers = [('lib2to3.fixes.fix_' + name) for name in fixer_names]
tool = RefactoringTool(fixer_names=fixers, explicit=fixers)
from lib2to3.pgen2 import tokenize as lib2to3_tokenize
try:
return unicode(tool.refactor_string(source_text, name=filename))
except lib2to3_tokenize.TokenError:
return source_text
| refactor_with_2to3 | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(source_text, fixer_names, filename=''):\n from lib2to3.refactor import RefactoringTool\n fixers = [('lib2to3.fixes.fix_' + name) for name in fixer_names]... |
def x(code):
try:
return compile(code, '<string>', 'exec', dont_inherit=True)
except (SyntaxError, TypeError, ValueError):
return False
| check_syntax | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(code):\n try:\n return compile(code, '<string>', 'exec', dont_inherit=True)\n except (SyntaxError, TypeError, ValueError):\n return False\n... |
def x(source, results, aggressive):
non_docstring_string_line_numbers = multiline_string_lines(source,
include_docstrings=False)
all_string_line_numbers = multiline_string_lines(source,
include_docstrings=True)
commented_out_code_line_numbers = commented_out_code_lines(source)
has_e901 = any(result['id'].lower() == 'e901' for result in results)
for r in results:
issue_id = r['id'].lower()
if r['line'] in non_docstring_string_line_numbers:
if issue_id.startswith(('e1', 'e501', 'w191')):
continue
if r['line'] in all_string_line_numbers:
if issue_id in ['e501']:
continue
if not aggressive and r['line'] + 1 in all_string_line_numbers:
if issue_id.startswith(('w29', 'w39')):
continue
if aggressive <= 0:
if issue_id.startswith(('e711', 'e72', 'w6')):
continue
if aggressive <= 1:
if issue_id.startswith(('e712', 'e713', 'e714')):
continue
if aggressive <= 2:
if issue_id.startswith('e704'):
continue
if r['line'] in commented_out_code_line_numbers:
if issue_id.startswith(('e26', 'e501')):
continue
if has_e901:
if issue_id.startswith(('e1', 'e7')):
continue
yield r
| filter_results | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(source, results, aggressive):\n non_docstring_string_line_numbers = multiline_string_lines(source,\n include_docstrings=False)\n all_string_line_n... |
def x(source, include_docstrings=False):
line_numbers = set()
previous_token_type = ''
try:
for t in generate_tokens(source):
token_type = t[0]
start_row = t[2][0]
end_row = t[3][0]
if token_type == tokenize.STRING and start_row != end_row:
if (include_docstrings or previous_token_type != tokenize.
INDENT):
line_numbers |= set(range(1 + start_row, 1 + end_row))
previous_token_type = token_type
except (SyntaxError, tokenize.TokenError):
pass
return line_numbers
| multiline_string_lines | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(source, include_docstrings=False):\n line_numbers = set()\n previous_token_type = ''\n try:\n for t in generate_tokens(source):\n to... |
def x(source):
line_numbers = []
try:
for t in generate_tokens(source):
token_type = t[0]
token_string = t[1]
start_row = t[2][0]
line = t[4]
if not line.lstrip().startswith('#'):
continue
if token_type == tokenize.COMMENT:
stripped_line = token_string.lstrip('#').strip()
if (' ' in stripped_line and '#' not in stripped_line and
check_syntax(stripped_line)):
line_numbers.append(start_row)
except (SyntaxError, tokenize.TokenError):
pass
return line_numbers
| commented_out_code_lines | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(source):\n line_numbers = []\n try:\n for t in generate_tokens(source):\n token_type = t[0]\n token_string = t[1]\n ... |
def x(line, max_line_length, last_comment=False):
assert len(line) > max_line_length
line = line.rstrip()
indentation = _get_indentation(line) + '# '
max_line_length = min(max_line_length, len(indentation) + 72)
MIN_CHARACTER_REPEAT = 5
if len(line) - len(line.rstrip(line[-1])
) >= MIN_CHARACTER_REPEAT and not line[-1].isalnum():
return line[:max_line_length] + '\n'
elif last_comment and re.match('\\s*#+\\s*\\w+', line):
split_lines = textwrap.wrap(line.lstrip(' \t#'), initial_indent=
indentation, subsequent_indent=indentation, width=
max_line_length, break_long_words=False, break_on_hyphens=False)
return '\n'.join(split_lines) + '\n'
return line + '\n'
| shorten_comment | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(line, max_line_length, last_comment=False):\n assert len(line) > max_line_length\n line = line.rstrip()\n indentation = _get_indentation(line) + '# '\... |
def x(source, options=None, encoding=None, apply_config=False):
options = _get_options(options, apply_config)
if not isinstance(source, unicode):
source = source.decode(encoding or get_encoding())
sio = io.StringIO(source)
return fix_lines(sio.readlines(), options=options)
| fix_code | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(source, options=None, encoding=None, apply_config=False):\n options = _get_options(options, apply_config)\n if not isinstance(source, unicode):\n ... |
def x(raw_options, apply_config):
if not raw_options:
return parse_args([''], apply_config=apply_config)
if isinstance(raw_options, dict):
options = parse_args([''], apply_config=apply_config)
for name, value in raw_options.items():
if not hasattr(options, name):
raise ValueError("No such option '{}'".format(name))
expected_type = type(getattr(options, name))
if not isinstance(expected_type, (str, unicode)):
if isinstance(value, (str, unicode)):
raise ValueError("Option '{}' should not be a string".
format(name))
setattr(options, name, value)
else:
options = raw_options
return options
| get_options | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(raw_options, apply_config):\n if not raw_options:\n return parse_args([''], apply_config=apply_config)\n if isinstance(raw_options, dict):\n ... |
def x(source_lines, options, filename=''):
original_newline = find_newline(source_lines)
tmp_source = ''.join(normalize_line_endings(source_lines, '\n'))
previous_hashes = set()
if options.line_range:
fixed_source = tmp_source
else:
pep8_options = {'ignore': options.ignore, 'select': options.select,
'max_line_length': options.max_line_length, 'hang_closing':
options.hang_closing}
sio = io.StringIO(tmp_source)
contents = sio.readlines()
results = _execute_pep8(pep8_options, contents)
codes = {result['id'] for result in results if result['id'] in
SELECTED_GLOBAL_FIXED_METHOD_CODES}
fixed_source = apply_global_fixes(tmp_source, options, filename=
filename, codes=codes)
passes = 0
long_line_ignore_cache = set()
while hash(fixed_source) not in previous_hashes:
if options.pep8_passes >= 0 and passes > options.pep8_passes:
break
passes += 1
previous_hashes.add(hash(fixed_source))
tmp_source = copy.copy(fixed_source)
fix = FixPEP8(filename, options, contents=tmp_source,
long_line_ignore_cache=long_line_ignore_cache)
fixed_source = fix.fix()
sio = io.StringIO(fixed_source)
return ''.join(normalize_line_endings(sio.readlines(), original_newline))
| fix_lines | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(source_lines, options, filename=''):\n original_newline = find_newline(source_lines)\n tmp_source = ''.join(normalize_line_endings(source_lines, '\\n'))\... |
def x():
for function in list(globals().values()):
if inspect.isfunction(function):
arguments = _get_parameters(function)
if arguments[:1] != ['source']:
continue
code = extract_code_from_function(function)
if code:
yield code, function
| global_fixes | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x():\n for function in list(globals().values()):\n if inspect.isfunction(function):\n arguments = _get_parameters(function)\n if ... |
def x(function):
if not function.__name__.startswith('fix_'):
return None
code = re.sub('^fix_', '', function.__name__)
if not code:
return None
try:
int(code[1:])
except ValueError:
return None
return code
| extract_code_from_function | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(function):\n if not function.__name__.startswith('fix_'):\n return None\n code = re.sub('^fix_', '', function.__name__)\n if not code:\n ... |
def x():
parser = argparse.ArgumentParser(description=docstring_summary(__doc__),
prog='autopep8')
parser.add_argument('--version', action='version', version=
'%(prog)s {} ({})'.format(__version__, _get_package_version()))
parser.add_argument('-v', '--verbose', action='count', default=0, help=
'print verbose messages; multiple -v result in more verbose messages')
parser.add_argument('-d', '--diff', action='store_true', help=
'print the diff for the fixed source')
parser.add_argument('-i', '--in-place', action='store_true', help=
'make changes to files in place')
parser.add_argument('--global-config', metavar='filename', default=
DEFAULT_CONFIG, help=
'path to a global pep8 config file; if this file does not exist then this is ignored (default: {})'
.format(DEFAULT_CONFIG))
parser.add_argument('--ignore-local-config', action='store_true', help=
"don't look for and apply local config files; if not passed, defaults are updated with any config files in the project's root directory"
)
parser.add_argument('-r', '--recursive', action='store_true', help=
'run recursively over directories; must be used with --in-place or --diff'
)
parser.add_argument('-j', '--jobs', type=int, metavar='n', default=1,
help='number of parallel jobs; match CPU count if value is less than 1'
)
parser.add_argument('-p', '--pep8-passes', metavar='n', default=-1,
type=int, help=
'maximum number of additional pep8 passes (default: infinite)')
parser.add_argument('-a', '--aggressive', action='count', default=0,
help=
'enable non-whitespace changes; multiple -a result in more aggressive changes'
)
parser.add_argument('--experimental', action='store_true', help=
'enable experimental fixes')
parser.add_argument('--exclude', metavar='globs', help=
'exclude file/directory names that match these comma-separated globs')
parser.add_argument('--list-fixes', action='store_true', help=
'list codes for fixes; used by --ignore and --select')
parser.add_argument('--ignore', metavar='errors', default='', help=
'do not fix these errors/warnings (default: {})'.format(DEFAULT_IGNORE)
)
parser.add_argument('--select', metavar='errors', default='', help=
'fix only these errors/warnings (e.g. E4,W)')
parser.add_argument('--max-line-length', metavar='n', default=79, type=
int, help='set maximum allowed line length (default: %(default)s)')
parser.add_argument('--line-range', '--range', metavar='line', default=
None, type=int, nargs=2, help=
'only fix errors found within this inclusive range of line numbers (e.g. 1 99); line numbers are indexed at 1'
)
parser.add_argument('--indent-size', default=DEFAULT_INDENT_SIZE, type=
int, help=argparse.SUPPRESS)
parser.add_argument('--hang-closing', action='store_true', help=
'hang-closing option passed to pycodestyle')
parser.add_argument('--exit-code', action='store_true', help=
'change to behavior of exit code. default behavior of return value, 0 is no differences, 1 is error exit. return 2 when add this option. 2 is exists differences.'
)
parser.add_argument('files', nargs='*', help=
"files to format or '-' for standard in")
return parser
| create_parser | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x():\n parser = argparse.ArgumentParser(description=docstring_summary(__doc__),\n prog='autopep8')\n parser.add_argument('--version', action='versio... |
def x(arguments, apply_config=False):
parser = create_parser()
args = parser.parse_args(arguments)
if not args.files and not args.list_fixes:
parser.error('incorrect number of arguments')
args.files = [decode_filename(name) for name in args.files]
if apply_config:
parser = read_config(args, parser)
args = parser.parse_args(arguments)
args.files = [decode_filename(name) for name in args.files]
if '-' in args.files:
if len(args.files) > 1:
parser.error('cannot mix stdin and regular files')
if args.diff:
parser.error('--diff cannot be used with standard input')
if args.in_place:
parser.error('--in-place cannot be used with standard input')
if args.recursive:
parser.error('--recursive cannot be used with standard input')
if len(args.files) > 1 and not (args.in_place or args.diff):
parser.error(
'autopep8 only takes one filename as argument unless the "--in-place" or "--diff" args are used'
)
if args.recursive and not (args.in_place or args.diff):
parser.error('--recursive must be used with --in-place or --diff')
if args.in_place and args.diff:
parser.error('--in-place and --diff are mutually exclusive')
if args.max_line_length <= 0:
parser.error('--max-line-length must be greater than 0')
if args.select:
args.select = _split_comma_separated(args.select)
if args.ignore:
args.ignore = _split_comma_separated(args.ignore)
if not all(any(conflicting_code.startswith(ignore_code) for
ignore_code in args.ignore) for conflicting_code in
CONFLICTING_CODES):
args.ignore.update(CONFLICTING_CODES)
elif not args.select:
if args.aggressive:
args.select = {'E', 'W1', 'W2', 'W3', 'W6'}
else:
args.ignore = _split_comma_separated(DEFAULT_IGNORE)
if args.exclude:
args.exclude = _split_comma_separated(args.exclude)
else:
args.exclude = {}
if args.jobs < 1:
import multiprocessing
args.jobs = multiprocessing.cpu_count()
if args.jobs > 1 and not args.in_place:
parser.error('parallel jobs requires --in-place')
if args.line_range:
if args.line_range[0] <= 0:
parser.error('--range must be positive numbers')
if args.line_range[0] > args.line_range[1]:
parser.error(
'First value of --range should be less than or equal to the second'
)
return args
| parse_args | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(arguments, apply_config=False):\n parser = create_parser()\n args = parser.parse_args(arguments)\n if not args.files and not args.list_fixes:\n ... |
def x(args, parser):
try:
from configparser import ConfigParser as SafeConfigParser
from configparser import Error
except ImportError:
from ConfigParser import SafeConfigParser
from ConfigParser import Error
config = SafeConfigParser()
try:
config.read(args.global_config)
if not args.ignore_local_config:
parent = tail = args.files and os.path.abspath(os.path.
commonprefix(args.files))
while tail:
if config.read([os.path.join(parent, fn) for fn in
PROJECT_CONFIG]):
break
parent, tail = os.path.split(parent)
defaults = {}
option_list = {o.dest: (o.type or type(o.default)) for o in parser.
_actions}
for section in ['pep8', 'pycodestyle', 'flake8']:
if not config.has_section(section):
continue
for k, _ in config.items(section):
norm_opt = k.lstrip('-').replace('-', '_')
if not option_list.get(norm_opt):
continue
opt_type = option_list[norm_opt]
if opt_type is int:
value = config.getint(section, k)
elif opt_type is bool:
value = config.getboolean(section, k)
else:
value = config.get(section, k)
if args.verbose:
print('enable config: section={}, key={}, value={}'.
format(section, k, value))
defaults[norm_opt] = value
parser.set_defaults(**defaults)
except Error:
pass
return parser
| read_config | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(args, parser):\n try:\n from configparser import ConfigParser as SafeConfigParser\n from configparser import Error\n except ImportError:\n ... |
def x(filename):
if isinstance(filename, unicode):
return filename
return filename.decode(sys.getfilesystemencoding())
| decode_filename | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(filename):\n if isinstance(filename, unicode):\n return filename\n return filename.decode(sys.getfilesystemencoding())\n",
"role": "user"
},... |
def x():
yield 'E101', docstring_summary(reindent.__doc__)
instance = FixPEP8(filename=None, options=None, contents='')
for attribute in dir(instance):
code = re.match('fix_([ew][0-9][0-9][0-9])', attribute)
if code:
yield code.group(1).upper(), re.sub('\\s+', ' ',
docstring_summary(getattr(instance, attribute).__doc__))
for code, function in sorted(global_fixes()):
yield code.upper() + (4 - len(code)) * ' ', re.sub('\\s+', ' ',
docstring_summary(function.__doc__))
for code in sorted(CODE_TO_2TO3):
yield code.upper() + (4 - len(code)) * ' ', re.sub('\\s+', ' ',
docstring_summary(fix_2to3.__doc__))
| supported_fixes | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x():\n yield 'E101', docstring_summary(reindent.__doc__)\n instance = FixPEP8(filename=None, options=None, contents='')\n for attribute in dir(instance)... |
def x(candidate, indent_word, max_line_length, experimental=False):
if not candidate.strip():
return 0
rank = 0
lines = candidate.rstrip().split('\n')
offset = 0
if not lines[0].lstrip().startswith('#') and lines[0].rstrip()[-1
] not in '([{':
for opening, closing in ('()', '[]', '{}'):
opening_loc = lines[0].find(opening)
closing_loc = lines[0].find(closing)
if opening_loc >= 0:
if closing_loc < 0 or closing_loc != opening_loc + 1:
offset = max(offset, 1 + opening_loc)
current_longest = max(offset + len(x.strip()) for x in lines)
rank += 4 * max(0, current_longest - max_line_length)
rank += len(lines)
rank += 2 * standard_deviation(len(line) for line in lines)
bad_staring_symbol = {'(': ')', '[': ']', '{': '}'}.get(lines[0][-1])
if len(lines) > 1:
if bad_staring_symbol and lines[1].lstrip().startswith(
bad_staring_symbol):
rank += 20
for lineno, current_line in enumerate(lines):
current_line = current_line.strip()
if current_line.startswith('#'):
continue
for bad_start in ['.', '%', '+', '-', '/']:
if current_line.startswith(bad_start):
rank += 100
if current_line == bad_start:
rank += 1000
if current_line.endswith(('.', '%', '+', '-', '/')
) and "': " in current_line:
rank += 1000
if current_line.endswith(('(', '[', '{', '.')):
if len(current_line) <= len(indent_word):
rank += 100
if current_line.endswith('(') and current_line[:-1].rstrip(
).endswith(','):
rank += 100
if current_line.endswith('[') and len(current_line) > 1 and (
current_line[-2].isalnum() or current_line[-2] in ']'):
rank += 300
if current_line.endswith('.'):
rank += 100
if has_arithmetic_operator(current_line):
rank += 100
if re.match('.*[(\\[{]\\s*[\\-\\+~]$', current_line.rstrip('\\ ')):
rank += 1000
if re.match('.*lambda\\s*\\*$', current_line.rstrip('\\ ')):
rank += 1000
if current_line.endswith(('%', '(', '[', '{')):
rank -= 20
if current_line.startswith('for '):
rank -= 50
if current_line.endswith('\\'):
total_len = len(current_line)
lineno += 1
while lineno < len(lines):
total_len += len(lines[lineno])
if lines[lineno].lstrip().startswith('#'):
total_len = max_line_length
break
if not lines[lineno].endswith('\\'):
break
lineno += 1
if total_len < max_line_length:
rank += 10
else:
rank += 100 if experimental else 1
if ',' in current_line and current_line.endswith(':'):
rank += 10
if current_line.endswith(':'):
rank += 100
rank += 10 * count_unbalanced_brackets(current_line)
return max(0, rank)
| line_shortening_rank | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(candidate, indent_word, max_line_length, experimental=False):\n if not candidate.strip():\n return 0\n rank = 0\n lines = candidate.rstrip().sp... |
def x(numbers):
numbers = list(numbers)
if not numbers:
return 0
mean = sum(numbers) / len(numbers)
return (sum((n - mean) ** 2 for n in numbers) / len(numbers)) ** 0.5
| standard_deviation | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(numbers):\n numbers = list(numbers)\n if not numbers:\n return 0\n mean = sum(numbers) / len(numbers)\n return (sum((n - mean) ** 2 for n in... |
def x(line):
count = 0
for opening, closing in ['()', '[]', '{}']:
count += abs(line.count(opening) - line.count(closing))
return count
| count_unbalanced_brackets | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(line):\n count = 0\n for opening, closing in ['()', '[]', '{}']:\n count += abs(line.count(opening) - line.count(closing))\n return count\n",
... |
def x(line, offsets):
result = []
previous_offset = 0
current_offset = 0
for current_offset in sorted(offsets):
if current_offset < len(line) and previous_offset != current_offset:
result.append(line[previous_offset:current_offset].strip())
previous_offset = current_offset
result.append(line[current_offset:])
return result
| split_at_offsets | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(line, offsets):\n result = []\n previous_offset = 0\n current_offset = 0\n for current_offset in sorted(offsets):\n if current_offset < len(... |
def x(filename, exclude):
base_name = os.path.basename(filename)
if base_name.startswith('.'):
return False
for pattern in exclude:
if fnmatch.fnmatch(base_name, pattern):
return False
if fnmatch.fnmatch(filename, pattern):
return False
if not os.path.isdir(filename) and not is_python_file(filename):
return False
return True
| match_file | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(filename, exclude):\n base_name = os.path.basename(filename)\n if base_name.startswith('.'):\n return False\n for pattern in exclude:\n ... |
def x(filenames, options, output=None):
results = []
filenames = find_files(filenames, options.recursive, options.exclude)
if options.jobs > 1:
import multiprocessing
pool = multiprocessing.Pool(options.jobs)
ret = pool.map(_fix_file, [(name, options) for name in filenames])
results.extend([x for x in ret if x is not None])
else:
for name in filenames:
ret = _fix_file((name, options, output))
if ret is None:
continue
if options.diff:
if ret != '':
results.append(ret)
elif options.in_place:
results.append(ret)
else:
original_source = readlines_from_file(name)
if ''.join(original_source).splitlines() != ret.splitlines():
results.append(ret)
return results
| fix_multiple_files | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(filenames, options, output=None):\n results = []\n filenames = find_files(filenames, options.recursive, options.exclude)\n if options.jobs > 1:\n ... |
def x(output, encoding):
return codecs.getwriter(encoding)(output.buffer if hasattr(output,
'buffer') else output)
| wrap_output | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(output, encoding):\n return codecs.getwriter(encoding)(output.buffer if hasattr(output,\n 'buffer') else output)\n",
"role": "user"
},
{
... |
def x(self):
pep8_options = {'ignore': self.options.ignore, 'select': self.options.
select, 'max_line_length': self.options.max_line_length,
'hang_closing': self.options.hang_closing}
results = _execute_pep8(pep8_options, self.source)
if self.options.verbose:
progress = {}
for r in results:
if r['id'] not in progress:
progress[r['id']] = set()
progress[r['id']].add(r['line'])
print('---> {n} issue(s) to fix {progress}'.format(n=len(results),
progress=progress), file=sys.stderr)
if self.options.line_range:
start, end = self.options.line_range
results = [r for r in results if start <= r['line'] <= end]
self._fix_source(filter_results(source=''.join(self.source), results=
results, aggressive=self.options.aggressive))
if self.options.line_range:
count = sum(sline.count('\n') for sline in self.source[start - 1:end])
self.options.line_range[1] = start + count - 1
return ''.join(self.source)
| fix | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n pep8_options = {'ignore': self.options.ignore, 'select': self.options.\n select, 'max_line_length': self.options.max_line_length,\n '... |
def x(self, result):
num_indent_spaces = int(result['info'].split()[1])
line_index = result['line'] - 1
target = self.source[line_index]
self.source[line_index] = ' ' * num_indent_spaces + target.lstrip()
| fix_reindent | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, result):\n num_indent_spaces = int(result['info'].split()[1])\n line_index = result['line'] - 1\n target = self.source[line_index]\n self.sou... |
def x(self, result):
line_index = result['line'] - 1
target = self.source[line_index]
if not target.lstrip().startswith('#'):
return []
self.source[line_index] = self.indent_word + target
| fix_e112 | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, result):\n line_index = result['line'] - 1\n target = self.source[line_index]\n if not target.lstrip().startswith('#'):\n return []\n ... |
def x(self, result):
line_index = result['line'] - 1
target = self.source[line_index]
indent = _get_indentation(target)
stripped = target.lstrip()
self.source[line_index] = indent[1:] + stripped
| fix_e113 | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, result):\n line_index = result['line'] - 1\n target = self.source[line_index]\n indent = _get_indentation(target)\n stripped = target.lstrip(... |
def x(self, result):
num_indent_spaces = int(result['info'].split()[1])
line_index = result['line'] - 1
target = self.source[line_index]
spaces_to_add = num_indent_spaces - len(_get_indentation(target))
indent = len(_get_indentation(target))
modified_lines = []
while len(_get_indentation(self.source[line_index])) >= indent:
self.source[line_index] = ' ' * spaces_to_add + self.source[line_index]
modified_lines.append(1 + line_index)
line_index -= 1
return modified_lines
| fix_e125 | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, result):\n num_indent_spaces = int(result['info'].split()[1])\n line_index = result['line'] - 1\n target = self.source[line_index]\n spaces_t... |
def x(self, result):
num_indent_spaces = int(result['info'].split()[1])
line_index = result['line'] - 1
target = self.source[line_index]
spaces_to_add = num_indent_spaces - len(_get_indentation(target))
if spaces_to_add >= 0:
self.source[line_index] = ' ' * spaces_to_add + self.source[line_index]
else:
offset = abs(spaces_to_add)
self.source[line_index] = self.source[line_index][offset:]
| fix_e131 | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, result):\n num_indent_spaces = int(result['info'].split()[1])\n line_index = result['line'] - 1\n target = self.source[line_index]\n spaces_t... |
def x(self, result):
target = self.source[result['line'] - 1]
offset = result['column'] - 1
fixed = target[:offset] + target[offset:].replace('\t', ' ')
self.source[result['line'] - 1] = fixed
| fix_e224 | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, result):\n target = self.source[result['line'] - 1]\n offset = result['column'] - 1\n fixed = target[:offset] + target[offset:].replace('\\t', '... |
def x(self, result):
target = self.source[result['line'] - 1]
offset = result['column'] - 1
fixed = target[:offset] + ' ' + target[offset:]
if fixed.replace(' ', '') == target.replace(' ', '') and _get_indentation(
fixed) == _get_indentation(target):
self.source[result['line'] - 1] = fixed
error_code = result.get('id', 0)
try:
ts = generate_tokens(fixed)
except (SyntaxError, tokenize.TokenError):
return
if not check_syntax(fixed.lstrip()):
return
errors = list(pycodestyle.missing_whitespace_around_operator(fixed, ts)
)
for e in reversed(errors):
if error_code != e[1].split()[0]:
continue
offset = e[0][1]
fixed = fixed[:offset] + ' ' + fixed[offset:]
self.source[result['line'] - 1] = fixed
else:
return []
| fix_e225 | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, result):\n target = self.source[result['line'] - 1]\n offset = result['column'] - 1\n fixed = target[:offset] + ' ' + target[offset:]\n if fi... |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.