repository_name stringlengths 5 67 | func_path_in_repository stringlengths 4 234 | func_name stringlengths 0 314 | whole_func_string stringlengths 52 3.87M | language stringclasses 6
values | func_code_string stringlengths 52 3.87M | func_code_tokens listlengths 15 672k | func_documentation_string stringlengths 1 47.2k | func_documentation_tokens listlengths 1 3.92k | split_name stringclasses 1
value | func_code_url stringlengths 85 339 |
|---|---|---|---|---|---|---|---|---|---|---|
gmr/queries | queries/session.py | Session._connect | def _connect(self):
"""Connect to PostgreSQL, either by reusing a connection from the pool
if possible, or by creating the new connection.
:rtype: psycopg2.extensions.connection
:raises: pool.NoIdleConnectionsError
"""
# Attempt to get a cached connection from the conne... | python | def _connect(self):
"""Connect to PostgreSQL, either by reusing a connection from the pool
if possible, or by creating the new connection.
:rtype: psycopg2.extensions.connection
:raises: pool.NoIdleConnectionsError
"""
# Attempt to get a cached connection from the conne... | [
"def",
"_connect",
"(",
"self",
")",
":",
"# Attempt to get a cached connection from the connection pool",
"try",
":",
"connection",
"=",
"self",
".",
"_pool_manager",
".",
"get",
"(",
"self",
".",
"pid",
",",
"self",
")",
"LOGGER",
".",
"debug",
"(",
"\"Re-usin... | Connect to PostgreSQL, either by reusing a connection from the pool
if possible, or by creating the new connection.
:rtype: psycopg2.extensions.connection
:raises: pool.NoIdleConnectionsError | [
"Connect",
"to",
"PostgreSQL",
"either",
"by",
"reusing",
"a",
"connection",
"from",
"the",
"pool",
"if",
"possible",
"or",
"by",
"creating",
"the",
"new",
"connection",
"."
] | train | https://github.com/gmr/queries/blob/a68855013dc6aaf9ed7b6909a4701f8da8796a0a/queries/session.py#L273-L307 |
gmr/queries | queries/session.py | Session._get_cursor | def _get_cursor(self, connection, name=None):
"""Return a cursor for the given cursor_factory. Specify a name to
use server-side cursors.
:param connection: The connection to create a cursor on
:type connection: psycopg2.extensions.connection
:param str name: A cursor name for a... | python | def _get_cursor(self, connection, name=None):
"""Return a cursor for the given cursor_factory. Specify a name to
use server-side cursors.
:param connection: The connection to create a cursor on
:type connection: psycopg2.extensions.connection
:param str name: A cursor name for a... | [
"def",
"_get_cursor",
"(",
"self",
",",
"connection",
",",
"name",
"=",
"None",
")",
":",
"cursor",
"=",
"connection",
".",
"cursor",
"(",
"name",
"=",
"name",
",",
"cursor_factory",
"=",
"self",
".",
"_cursor_factory",
")",
"if",
"name",
"is",
"not",
... | Return a cursor for the given cursor_factory. Specify a name to
use server-side cursors.
:param connection: The connection to create a cursor on
:type connection: psycopg2.extensions.connection
:param str name: A cursor name for a server side cursor
:rtype: psycopg2.extensions.c... | [
"Return",
"a",
"cursor",
"for",
"the",
"given",
"cursor_factory",
".",
"Specify",
"a",
"name",
"to",
"use",
"server",
"-",
"side",
"cursors",
"."
] | train | https://github.com/gmr/queries/blob/a68855013dc6aaf9ed7b6909a4701f8da8796a0a/queries/session.py#L309-L324 |
gmr/queries | queries/session.py | Session._incr_exceptions | def _incr_exceptions(self):
"""Increment the number of exceptions for the current connection."""
self._pool_manager.get_connection(self.pid, self._conn).exceptions += 1 | python | def _incr_exceptions(self):
"""Increment the number of exceptions for the current connection."""
self._pool_manager.get_connection(self.pid, self._conn).exceptions += 1 | [
"def",
"_incr_exceptions",
"(",
"self",
")",
":",
"self",
".",
"_pool_manager",
".",
"get_connection",
"(",
"self",
".",
"pid",
",",
"self",
".",
"_conn",
")",
".",
"exceptions",
"+=",
"1"
] | Increment the number of exceptions for the current connection. | [
"Increment",
"the",
"number",
"of",
"exceptions",
"for",
"the",
"current",
"connection",
"."
] | train | https://github.com/gmr/queries/blob/a68855013dc6aaf9ed7b6909a4701f8da8796a0a/queries/session.py#L326-L328 |
gmr/queries | queries/session.py | Session._incr_executions | def _incr_executions(self):
"""Increment the number of executions for the current connection."""
self._pool_manager.get_connection(self.pid, self._conn).executions += 1 | python | def _incr_executions(self):
"""Increment the number of executions for the current connection."""
self._pool_manager.get_connection(self.pid, self._conn).executions += 1 | [
"def",
"_incr_executions",
"(",
"self",
")",
":",
"self",
".",
"_pool_manager",
".",
"get_connection",
"(",
"self",
".",
"pid",
",",
"self",
".",
"_conn",
")",
".",
"executions",
"+=",
"1"
] | Increment the number of executions for the current connection. | [
"Increment",
"the",
"number",
"of",
"executions",
"for",
"the",
"current",
"connection",
"."
] | train | https://github.com/gmr/queries/blob/a68855013dc6aaf9ed7b6909a4701f8da8796a0a/queries/session.py#L330-L332 |
gmr/queries | queries/session.py | Session._register_unicode | def _register_unicode(connection):
"""Register the cursor to be able to receive Unicode string.
:type connection: psycopg2.extensions.connection
:param connection: Where to register things
"""
psycopg2.extensions.register_type(psycopg2.extensions.UNICODE,
... | python | def _register_unicode(connection):
"""Register the cursor to be able to receive Unicode string.
:type connection: psycopg2.extensions.connection
:param connection: Where to register things
"""
psycopg2.extensions.register_type(psycopg2.extensions.UNICODE,
... | [
"def",
"_register_unicode",
"(",
"connection",
")",
":",
"psycopg2",
".",
"extensions",
".",
"register_type",
"(",
"psycopg2",
".",
"extensions",
".",
"UNICODE",
",",
"connection",
")",
"psycopg2",
".",
"extensions",
".",
"register_type",
"(",
"psycopg2",
".",
... | Register the cursor to be able to receive Unicode string.
:type connection: psycopg2.extensions.connection
:param connection: Where to register things | [
"Register",
"the",
"cursor",
"to",
"be",
"able",
"to",
"receive",
"Unicode",
"string",
"."
] | train | https://github.com/gmr/queries/blob/a68855013dc6aaf9ed7b6909a4701f8da8796a0a/queries/session.py#L345-L355 |
gmr/queries | queries/session.py | Session._status | def _status(self):
"""Return the current connection status as an integer value.
The status should match one of the following constants:
- queries.Session.INTRANS: Connection established, in transaction
- queries.Session.PREPARED: Prepared for second phase of transaction
- queri... | python | def _status(self):
"""Return the current connection status as an integer value.
The status should match one of the following constants:
- queries.Session.INTRANS: Connection established, in transaction
- queries.Session.PREPARED: Prepared for second phase of transaction
- queri... | [
"def",
"_status",
"(",
"self",
")",
":",
"if",
"self",
".",
"_conn",
".",
"status",
"==",
"psycopg2",
".",
"extensions",
".",
"STATUS_BEGIN",
":",
"return",
"self",
".",
"READY",
"return",
"self",
".",
"_conn",
".",
"status"
] | Return the current connection status as an integer value.
The status should match one of the following constants:
- queries.Session.INTRANS: Connection established, in transaction
- queries.Session.PREPARED: Prepared for second phase of transaction
- queries.Session.READY: Connected, n... | [
"Return",
"the",
"current",
"connection",
"status",
"as",
"an",
"integer",
"value",
"."
] | train | https://github.com/gmr/queries/blob/a68855013dc6aaf9ed7b6909a4701f8da8796a0a/queries/session.py#L368-L382 |
gmr/queries | queries/results.py | Results.as_dict | def as_dict(self):
"""Return a single row result as a dictionary. If the results contain
multiple rows, a :py:class:`ValueError` will be raised.
:return: dict
:raises: ValueError
"""
if not self.cursor.rowcount:
return {}
self._rewind()
if s... | python | def as_dict(self):
"""Return a single row result as a dictionary. If the results contain
multiple rows, a :py:class:`ValueError` will be raised.
:return: dict
:raises: ValueError
"""
if not self.cursor.rowcount:
return {}
self._rewind()
if s... | [
"def",
"as_dict",
"(",
"self",
")",
":",
"if",
"not",
"self",
".",
"cursor",
".",
"rowcount",
":",
"return",
"{",
"}",
"self",
".",
"_rewind",
"(",
")",
"if",
"self",
".",
"cursor",
".",
"rowcount",
"==",
"1",
":",
"return",
"dict",
"(",
"self",
... | Return a single row result as a dictionary. If the results contain
multiple rows, a :py:class:`ValueError` will be raised.
:return: dict
:raises: ValueError | [
"Return",
"a",
"single",
"row",
"result",
"as",
"a",
"dictionary",
".",
"If",
"the",
"results",
"contain",
"multiple",
"rows",
"a",
":",
"py",
":",
"class",
":",
"ValueError",
"will",
"be",
"raised",
"."
] | train | https://github.com/gmr/queries/blob/a68855013dc6aaf9ed7b6909a4701f8da8796a0a/queries/results.py#L66-L81 |
gmr/queries | queries/results.py | Results.items | def items(self):
"""Return all of the rows that are in the result set.
:rtype: list
"""
if not self.cursor.rowcount:
return []
self.cursor.scroll(0, 'absolute')
return self.cursor.fetchall() | python | def items(self):
"""Return all of the rows that are in the result set.
:rtype: list
"""
if not self.cursor.rowcount:
return []
self.cursor.scroll(0, 'absolute')
return self.cursor.fetchall() | [
"def",
"items",
"(",
"self",
")",
":",
"if",
"not",
"self",
".",
"cursor",
".",
"rowcount",
":",
"return",
"[",
"]",
"self",
".",
"cursor",
".",
"scroll",
"(",
"0",
",",
"'absolute'",
")",
"return",
"self",
".",
"cursor",
".",
"fetchall",
"(",
")"
... | Return all of the rows that are in the result set.
:rtype: list | [
"Return",
"all",
"of",
"the",
"rows",
"that",
"are",
"in",
"the",
"result",
"set",
"."
] | train | https://github.com/gmr/queries/blob/a68855013dc6aaf9ed7b6909a4701f8da8796a0a/queries/results.py#L98-L108 |
gmr/queries | queries/utils.py | get_current_user | def get_current_user():
"""Return the current username for the logged in user
:rtype: str
"""
if pwd is None:
return getpass.getuser()
else:
try:
return pwd.getpwuid(os.getuid())[0]
except KeyError as error:
LOGGER.error('Could not get logged-in user... | python | def get_current_user():
"""Return the current username for the logged in user
:rtype: str
"""
if pwd is None:
return getpass.getuser()
else:
try:
return pwd.getpwuid(os.getuid())[0]
except KeyError as error:
LOGGER.error('Could not get logged-in user... | [
"def",
"get_current_user",
"(",
")",
":",
"if",
"pwd",
"is",
"None",
":",
"return",
"getpass",
".",
"getuser",
"(",
")",
"else",
":",
"try",
":",
"return",
"pwd",
".",
"getpwuid",
"(",
"os",
".",
"getuid",
"(",
")",
")",
"[",
"0",
"]",
"except",
... | Return the current username for the logged in user
:rtype: str | [
"Return",
"the",
"current",
"username",
"for",
"the",
"logged",
"in",
"user"
] | train | https://github.com/gmr/queries/blob/a68855013dc6aaf9ed7b6909a4701f8da8796a0a/queries/utils.py#L57-L69 |
gmr/queries | queries/utils.py | uri | def uri(host='localhost', port=5432, dbname='postgres', user='postgres',
password=None):
"""Return a PostgreSQL connection URI for the specified values.
:param str host: Host to connect to
:param int port: Port to connect on
:param str dbname: The database name
:param str user: User to conn... | python | def uri(host='localhost', port=5432, dbname='postgres', user='postgres',
password=None):
"""Return a PostgreSQL connection URI for the specified values.
:param str host: Host to connect to
:param int port: Port to connect on
:param str dbname: The database name
:param str user: User to conn... | [
"def",
"uri",
"(",
"host",
"=",
"'localhost'",
",",
"port",
"=",
"5432",
",",
"dbname",
"=",
"'postgres'",
",",
"user",
"=",
"'postgres'",
",",
"password",
"=",
"None",
")",
":",
"if",
"port",
":",
"host",
"=",
"'%s:%s'",
"%",
"(",
"host",
",",
"po... | Return a PostgreSQL connection URI for the specified values.
:param str host: Host to connect to
:param int port: Port to connect on
:param str dbname: The database name
:param str user: User to connect as
:param str password: The password to use, None for no password
:return str: The PostgreSQ... | [
"Return",
"a",
"PostgreSQL",
"connection",
"URI",
"for",
"the",
"specified",
"values",
"."
] | train | https://github.com/gmr/queries/blob/a68855013dc6aaf9ed7b6909a4701f8da8796a0a/queries/utils.py#L82-L98 |
gmr/queries | queries/utils.py | uri_to_kwargs | def uri_to_kwargs(uri):
"""Return a URI as kwargs for connecting to PostgreSQL with psycopg2,
applying default values for non-specified areas of the URI.
:param str uri: The connection URI
:rtype: dict
"""
parsed = urlparse(uri)
default_user = get_current_user()
password = unquote(pars... | python | def uri_to_kwargs(uri):
"""Return a URI as kwargs for connecting to PostgreSQL with psycopg2,
applying default values for non-specified areas of the URI.
:param str uri: The connection URI
:rtype: dict
"""
parsed = urlparse(uri)
default_user = get_current_user()
password = unquote(pars... | [
"def",
"uri_to_kwargs",
"(",
"uri",
")",
":",
"parsed",
"=",
"urlparse",
"(",
"uri",
")",
"default_user",
"=",
"get_current_user",
"(",
")",
"password",
"=",
"unquote",
"(",
"parsed",
".",
"password",
")",
"if",
"parsed",
".",
"password",
"else",
"None",
... | Return a URI as kwargs for connecting to PostgreSQL with psycopg2,
applying default values for non-specified areas of the URI.
:param str uri: The connection URI
:rtype: dict | [
"Return",
"a",
"URI",
"as",
"kwargs",
"for",
"connecting",
"to",
"PostgreSQL",
"with",
"psycopg2",
"applying",
"default",
"values",
"for",
"non",
"-",
"specified",
"areas",
"of",
"the",
"URI",
"."
] | train | https://github.com/gmr/queries/blob/a68855013dc6aaf9ed7b6909a4701f8da8796a0a/queries/utils.py#L101-L127 |
gmr/queries | queries/utils.py | urlparse | def urlparse(url):
"""Parse the URL in a Python2/3 independent fashion.
:param str url: The URL to parse
:rtype: Parsed
"""
value = 'http%s' % url[5:] if url[:5] == 'postgresql' else url
parsed = _urlparse.urlparse(value)
path, query = parsed.path, parsed.query
hostname = parsed.hostna... | python | def urlparse(url):
"""Parse the URL in a Python2/3 independent fashion.
:param str url: The URL to parse
:rtype: Parsed
"""
value = 'http%s' % url[5:] if url[:5] == 'postgresql' else url
parsed = _urlparse.urlparse(value)
path, query = parsed.path, parsed.query
hostname = parsed.hostna... | [
"def",
"urlparse",
"(",
"url",
")",
":",
"value",
"=",
"'http%s'",
"%",
"url",
"[",
"5",
":",
"]",
"if",
"url",
"[",
":",
"5",
"]",
"==",
"'postgresql'",
"else",
"url",
"parsed",
"=",
"_urlparse",
".",
"urlparse",
"(",
"value",
")",
"path",
",",
... | Parse the URL in a Python2/3 independent fashion.
:param str url: The URL to parse
:rtype: Parsed | [
"Parse",
"the",
"URL",
"in",
"a",
"Python2",
"/",
"3",
"independent",
"fashion",
"."
] | train | https://github.com/gmr/queries/blob/a68855013dc6aaf9ed7b6909a4701f8da8796a0a/queries/utils.py#L130-L150 |
gmr/queries | queries/tornado_session.py | Results.free | def free(self):
"""Release the results and connection lock from the TornadoSession
object. This **must** be called after you finish processing the results
from :py:meth:`TornadoSession.query <queries.TornadoSession.query>` or
:py:meth:`TornadoSession.callproc <queries.TornadoSession.call... | python | def free(self):
"""Release the results and connection lock from the TornadoSession
object. This **must** be called after you finish processing the results
from :py:meth:`TornadoSession.query <queries.TornadoSession.query>` or
:py:meth:`TornadoSession.callproc <queries.TornadoSession.call... | [
"def",
"free",
"(",
"self",
")",
":",
"self",
".",
"_freed",
"=",
"True",
"self",
".",
"_cleanup",
"(",
"self",
".",
"cursor",
",",
"self",
".",
"_fd",
")"
] | Release the results and connection lock from the TornadoSession
object. This **must** be called after you finish processing the results
from :py:meth:`TornadoSession.query <queries.TornadoSession.query>` or
:py:meth:`TornadoSession.callproc <queries.TornadoSession.callproc>`
or the conne... | [
"Release",
"the",
"results",
"and",
"connection",
"lock",
"from",
"the",
"TornadoSession",
"object",
".",
"This",
"**",
"must",
"**",
"be",
"called",
"after",
"you",
"finish",
"processing",
"the",
"results",
"from",
":",
"py",
":",
"meth",
":",
"TornadoSessi... | train | https://github.com/gmr/queries/blob/a68855013dc6aaf9ed7b6909a4701f8da8796a0a/queries/tornado_session.py#L103-L113 |
gmr/queries | queries/tornado_session.py | TornadoSession._ensure_pool_exists | def _ensure_pool_exists(self):
"""Create the pool in the pool manager if it does not exist."""
if self.pid not in self._pool_manager:
self._pool_manager.create(self.pid, self._pool_idle_ttl,
self._pool_max_size, self._ioloop.time) | python | def _ensure_pool_exists(self):
"""Create the pool in the pool manager if it does not exist."""
if self.pid not in self._pool_manager:
self._pool_manager.create(self.pid, self._pool_idle_ttl,
self._pool_max_size, self._ioloop.time) | [
"def",
"_ensure_pool_exists",
"(",
"self",
")",
":",
"if",
"self",
".",
"pid",
"not",
"in",
"self",
".",
"_pool_manager",
":",
"self",
".",
"_pool_manager",
".",
"create",
"(",
"self",
".",
"pid",
",",
"self",
".",
"_pool_idle_ttl",
",",
"self",
".",
"... | Create the pool in the pool manager if it does not exist. | [
"Create",
"the",
"pool",
"in",
"the",
"pool",
"manager",
"if",
"it",
"does",
"not",
"exist",
"."
] | train | https://github.com/gmr/queries/blob/a68855013dc6aaf9ed7b6909a4701f8da8796a0a/queries/tornado_session.py#L165-L169 |
gmr/queries | queries/tornado_session.py | TornadoSession._connect | def _connect(self):
"""Connect to PostgreSQL, either by reusing a connection from the pool
if possible, or by creating the new connection.
:rtype: psycopg2.extensions.connection
:raises: pool.NoIdleConnectionsError
"""
future = concurrent.Future()
# Attempt to ... | python | def _connect(self):
"""Connect to PostgreSQL, either by reusing a connection from the pool
if possible, or by creating the new connection.
:rtype: psycopg2.extensions.connection
:raises: pool.NoIdleConnectionsError
"""
future = concurrent.Future()
# Attempt to ... | [
"def",
"_connect",
"(",
"self",
")",
":",
"future",
"=",
"concurrent",
".",
"Future",
"(",
")",
"# Attempt to get a cached connection from the connection pool",
"try",
":",
"connection",
"=",
"self",
".",
"_pool_manager",
".",
"get",
"(",
"self",
".",
"pid",
","... | Connect to PostgreSQL, either by reusing a connection from the pool
if possible, or by creating the new connection.
:rtype: psycopg2.extensions.connection
:raises: pool.NoIdleConnectionsError | [
"Connect",
"to",
"PostgreSQL",
"either",
"by",
"reusing",
"a",
"connection",
"from",
"the",
"pool",
"if",
"possible",
"or",
"by",
"creating",
"the",
"new",
"connection",
"."
] | train | https://github.com/gmr/queries/blob/a68855013dc6aaf9ed7b6909a4701f8da8796a0a/queries/tornado_session.py#L244-L267 |
gmr/queries | queries/tornado_session.py | TornadoSession._create_connection | def _create_connection(self, future):
"""Create a new PostgreSQL connection
:param tornado.concurrent.Future future: future for new conn result
"""
LOGGER.debug('Creating a new connection for %s', self.pid)
# Create a new PostgreSQL connection
kwargs = utils.uri_to_kwa... | python | def _create_connection(self, future):
"""Create a new PostgreSQL connection
:param tornado.concurrent.Future future: future for new conn result
"""
LOGGER.debug('Creating a new connection for %s', self.pid)
# Create a new PostgreSQL connection
kwargs = utils.uri_to_kwa... | [
"def",
"_create_connection",
"(",
"self",
",",
"future",
")",
":",
"LOGGER",
".",
"debug",
"(",
"'Creating a new connection for %s'",
",",
"self",
".",
"pid",
")",
"# Create a new PostgreSQL connection",
"kwargs",
"=",
"utils",
".",
"uri_to_kwargs",
"(",
"self",
"... | Create a new PostgreSQL connection
:param tornado.concurrent.Future future: future for new conn result | [
"Create",
"a",
"new",
"PostgreSQL",
"connection"
] | train | https://github.com/gmr/queries/blob/a68855013dc6aaf9ed7b6909a4701f8da8796a0a/queries/tornado_session.py#L269-L336 |
gmr/queries | queries/tornado_session.py | TornadoSession._execute | def _execute(self, method, query, parameters=None):
"""Issue a query asynchronously on the server, mogrifying the
parameters against the sql statement and yielding the results
as a :py:class:`Results <queries.tornado_session.Results>` object.
This function reduces duplicate code for cal... | python | def _execute(self, method, query, parameters=None):
"""Issue a query asynchronously on the server, mogrifying the
parameters against the sql statement and yielding the results
as a :py:class:`Results <queries.tornado_session.Results>` object.
This function reduces duplicate code for cal... | [
"def",
"_execute",
"(",
"self",
",",
"method",
",",
"query",
",",
"parameters",
"=",
"None",
")",
":",
"future",
"=",
"concurrent",
".",
"Future",
"(",
")",
"def",
"on_connected",
"(",
"cf",
")",
":",
"\"\"\"Invoked by the future returned by self._connect\"\"\""... | Issue a query asynchronously on the server, mogrifying the
parameters against the sql statement and yielding the results
as a :py:class:`Results <queries.tornado_session.Results>` object.
This function reduces duplicate code for callproc and query by getting
the class attribute for the ... | [
"Issue",
"a",
"query",
"asynchronously",
"on",
"the",
"server",
"mogrifying",
"the",
"parameters",
"against",
"the",
"sql",
"statement",
"and",
"yielding",
"the",
"results",
"as",
"a",
":",
"py",
":",
"class",
":",
"Results",
"<queries",
".",
"tornado_session"... | train | https://github.com/gmr/queries/blob/a68855013dc6aaf9ed7b6909a4701f8da8796a0a/queries/tornado_session.py#L338-L406 |
gmr/queries | queries/tornado_session.py | TornadoSession._exec_cleanup | def _exec_cleanup(self, cursor, fd):
"""Close the cursor, remove any references to the fd in internal state
and remove the fd from the ioloop.
:param psycopg2.extensions.cursor cursor: The cursor to close
:param int fd: The connection file descriptor
"""
LOGGER.debug('C... | python | def _exec_cleanup(self, cursor, fd):
"""Close the cursor, remove any references to the fd in internal state
and remove the fd from the ioloop.
:param psycopg2.extensions.cursor cursor: The cursor to close
:param int fd: The connection file descriptor
"""
LOGGER.debug('C... | [
"def",
"_exec_cleanup",
"(",
"self",
",",
"cursor",
",",
"fd",
")",
":",
"LOGGER",
".",
"debug",
"(",
"'Closing cursor and cleaning %s'",
",",
"fd",
")",
"try",
":",
"cursor",
".",
"close",
"(",
")",
"except",
"(",
"psycopg2",
".",
"Error",
",",
"psycopg... | Close the cursor, remove any references to the fd in internal state
and remove the fd from the ioloop.
:param psycopg2.extensions.cursor cursor: The cursor to close
:param int fd: The connection file descriptor | [
"Close",
"the",
"cursor",
"remove",
"any",
"references",
"to",
"the",
"fd",
"in",
"internal",
"state",
"and",
"remove",
"the",
"fd",
"from",
"the",
"ioloop",
"."
] | train | https://github.com/gmr/queries/blob/a68855013dc6aaf9ed7b6909a4701f8da8796a0a/queries/tornado_session.py#L408-L431 |
gmr/queries | queries/tornado_session.py | TornadoSession._cleanup_fd | def _cleanup_fd(self, fd, close=False):
"""Ensure the socket socket is removed from the IOLoop, the
connection stack, and futures stack.
:param int fd: The fd # to cleanup
"""
self._ioloop.remove_handler(fd)
if fd in self._connections:
try:
s... | python | def _cleanup_fd(self, fd, close=False):
"""Ensure the socket socket is removed from the IOLoop, the
connection stack, and futures stack.
:param int fd: The fd # to cleanup
"""
self._ioloop.remove_handler(fd)
if fd in self._connections:
try:
s... | [
"def",
"_cleanup_fd",
"(",
"self",
",",
"fd",
",",
"close",
"=",
"False",
")",
":",
"self",
".",
"_ioloop",
".",
"remove_handler",
"(",
"fd",
")",
"if",
"fd",
"in",
"self",
".",
"_connections",
":",
"try",
":",
"self",
".",
"_pool_manager",
".",
"fre... | Ensure the socket socket is removed from the IOLoop, the
connection stack, and futures stack.
:param int fd: The fd # to cleanup | [
"Ensure",
"the",
"socket",
"socket",
"is",
"removed",
"from",
"the",
"IOLoop",
"the",
"connection",
"stack",
"and",
"futures",
"stack",
"."
] | train | https://github.com/gmr/queries/blob/a68855013dc6aaf9ed7b6909a4701f8da8796a0a/queries/tornado_session.py#L433-L450 |
gmr/queries | queries/tornado_session.py | TornadoSession._incr_exceptions | def _incr_exceptions(self, conn):
"""Increment the number of exceptions for the current connection.
:param psycopg2.extensions.connection conn: the psycopg2 connection
"""
self._pool_manager.get_connection(self.pid, conn).exceptions += 1 | python | def _incr_exceptions(self, conn):
"""Increment the number of exceptions for the current connection.
:param psycopg2.extensions.connection conn: the psycopg2 connection
"""
self._pool_manager.get_connection(self.pid, conn).exceptions += 1 | [
"def",
"_incr_exceptions",
"(",
"self",
",",
"conn",
")",
":",
"self",
".",
"_pool_manager",
".",
"get_connection",
"(",
"self",
".",
"pid",
",",
"conn",
")",
".",
"exceptions",
"+=",
"1"
] | Increment the number of exceptions for the current connection.
:param psycopg2.extensions.connection conn: the psycopg2 connection | [
"Increment",
"the",
"number",
"of",
"exceptions",
"for",
"the",
"current",
"connection",
"."
] | train | https://github.com/gmr/queries/blob/a68855013dc6aaf9ed7b6909a4701f8da8796a0a/queries/tornado_session.py#L452-L458 |
gmr/queries | queries/tornado_session.py | TornadoSession._incr_executions | def _incr_executions(self, conn):
"""Increment the number of executions for the current connection.
:param psycopg2.extensions.connection conn: the psycopg2 connection
"""
self._pool_manager.get_connection(self.pid, conn).executions += 1 | python | def _incr_executions(self, conn):
"""Increment the number of executions for the current connection.
:param psycopg2.extensions.connection conn: the psycopg2 connection
"""
self._pool_manager.get_connection(self.pid, conn).executions += 1 | [
"def",
"_incr_executions",
"(",
"self",
",",
"conn",
")",
":",
"self",
".",
"_pool_manager",
".",
"get_connection",
"(",
"self",
".",
"pid",
",",
"conn",
")",
".",
"executions",
"+=",
"1"
] | Increment the number of executions for the current connection.
:param psycopg2.extensions.connection conn: the psycopg2 connection | [
"Increment",
"the",
"number",
"of",
"executions",
"for",
"the",
"current",
"connection",
"."
] | train | https://github.com/gmr/queries/blob/a68855013dc6aaf9ed7b6909a4701f8da8796a0a/queries/tornado_session.py#L460-L466 |
gmr/queries | queries/tornado_session.py | TornadoSession._on_io_events | def _on_io_events(self, fd=None, _events=None):
"""Invoked by Tornado's IOLoop when there are events for the fd
:param int fd: The file descriptor for the event
:param int _events: The events raised
"""
if fd not in self._connections:
LOGGER.warning('Received IO eve... | python | def _on_io_events(self, fd=None, _events=None):
"""Invoked by Tornado's IOLoop when there are events for the fd
:param int fd: The file descriptor for the event
:param int _events: The events raised
"""
if fd not in self._connections:
LOGGER.warning('Received IO eve... | [
"def",
"_on_io_events",
"(",
"self",
",",
"fd",
"=",
"None",
",",
"_events",
"=",
"None",
")",
":",
"if",
"fd",
"not",
"in",
"self",
".",
"_connections",
":",
"LOGGER",
".",
"warning",
"(",
"'Received IO event for non-existing connection'",
")",
"return",
"s... | Invoked by Tornado's IOLoop when there are events for the fd
:param int fd: The file descriptor for the event
:param int _events: The events raised | [
"Invoked",
"by",
"Tornado",
"s",
"IOLoop",
"when",
"there",
"are",
"events",
"for",
"the",
"fd"
] | train | https://github.com/gmr/queries/blob/a68855013dc6aaf9ed7b6909a4701f8da8796a0a/queries/tornado_session.py#L468-L478 |
gmr/queries | queries/tornado_session.py | TornadoSession._poll_connection | def _poll_connection(self, fd):
"""Check with psycopg2 to see what action to take. If the state is
POLL_OK, we should have a pending callback for that fd.
:param int fd: The socket fd for the postgresql connection
"""
try:
state = self._connections[fd].poll()
... | python | def _poll_connection(self, fd):
"""Check with psycopg2 to see what action to take. If the state is
POLL_OK, we should have a pending callback for that fd.
:param int fd: The socket fd for the postgresql connection
"""
try:
state = self._connections[fd].poll()
... | [
"def",
"_poll_connection",
"(",
"self",
",",
"fd",
")",
":",
"try",
":",
"state",
"=",
"self",
".",
"_connections",
"[",
"fd",
"]",
".",
"poll",
"(",
")",
"except",
"(",
"OSError",
",",
"socket",
".",
"error",
")",
"as",
"error",
":",
"self",
".",
... | Check with psycopg2 to see what action to take. If the state is
POLL_OK, we should have a pending callback for that fd.
:param int fd: The socket fd for the postgresql connection | [
"Check",
"with",
"psycopg2",
"to",
"see",
"what",
"action",
"to",
"take",
".",
"If",
"the",
"state",
"is",
"POLL_OK",
"we",
"should",
"have",
"a",
"pending",
"callback",
"for",
"that",
"fd",
"."
] | train | https://github.com/gmr/queries/blob/a68855013dc6aaf9ed7b6909a4701f8da8796a0a/queries/tornado_session.py#L480-L510 |
jquast/wcwidth | setup.py | main | def main():
"""Setup.py entry point."""
import codecs
setuptools.setup(
name='wcwidth',
version='0.1.7',
description=("Measures number of Terminal column cells "
"of wide-character codes"),
long_description=codecs.open(
os.path.join(HERE, 'REA... | python | def main():
"""Setup.py entry point."""
import codecs
setuptools.setup(
name='wcwidth',
version='0.1.7',
description=("Measures number of Terminal column cells "
"of wide-character codes"),
long_description=codecs.open(
os.path.join(HERE, 'REA... | [
"def",
"main",
"(",
")",
":",
"import",
"codecs",
"setuptools",
".",
"setup",
"(",
"name",
"=",
"'wcwidth'",
",",
"version",
"=",
"'0.1.7'",
",",
"description",
"=",
"(",
"\"Measures number of Terminal column cells \"",
"\"of wide-character codes\"",
")",
",",
"lo... | Setup.py entry point. | [
"Setup",
".",
"py",
"entry",
"point",
"."
] | train | https://github.com/jquast/wcwidth/blob/78800b68911880ef4ef95ae83886154710441871/setup.py#L271-L307 |
jquast/wcwidth | setup.py | SetupUpdate._do_readme_update | def _do_readme_update(self):
"""Patch README.rst to reflect the data files used in release."""
import codecs
import glob
# read in,
data_in = codecs.open(
os.path.join(HERE, 'README.rst'), 'r', 'utf8').read()
# search for beginning and end positions,
... | python | def _do_readme_update(self):
"""Patch README.rst to reflect the data files used in release."""
import codecs
import glob
# read in,
data_in = codecs.open(
os.path.join(HERE, 'README.rst'), 'r', 'utf8').read()
# search for beginning and end positions,
... | [
"def",
"_do_readme_update",
"(",
"self",
")",
":",
"import",
"codecs",
"import",
"glob",
"# read in,",
"data_in",
"=",
"codecs",
".",
"open",
"(",
"os",
".",
"path",
".",
"join",
"(",
"HERE",
",",
"'README.rst'",
")",
",",
"'r'",
",",
"'utf8'",
")",
".... | Patch README.rst to reflect the data files used in release. | [
"Patch",
"README",
".",
"rst",
"to",
"reflect",
"the",
"data",
"files",
"used",
"in",
"release",
"."
] | train | https://github.com/jquast/wcwidth/blob/78800b68911880ef4ef95ae83886154710441871/setup.py#L78-L112 |
jquast/wcwidth | setup.py | SetupUpdate._do_east_asian | def _do_east_asian(self):
"""Fetch and update east-asian tables."""
self._do_retrieve(self.EAW_URL, self.EAW_IN)
(version, date, values) = self._parse_east_asian(
fname=self.EAW_IN,
properties=(u'W', u'F',)
)
table = self._make_table(values)
self._... | python | def _do_east_asian(self):
"""Fetch and update east-asian tables."""
self._do_retrieve(self.EAW_URL, self.EAW_IN)
(version, date, values) = self._parse_east_asian(
fname=self.EAW_IN,
properties=(u'W', u'F',)
)
table = self._make_table(values)
self._... | [
"def",
"_do_east_asian",
"(",
"self",
")",
":",
"self",
".",
"_do_retrieve",
"(",
"self",
".",
"EAW_URL",
",",
"self",
".",
"EAW_IN",
")",
"(",
"version",
",",
"date",
",",
"values",
")",
"=",
"self",
".",
"_parse_east_asian",
"(",
"fname",
"=",
"self"... | Fetch and update east-asian tables. | [
"Fetch",
"and",
"update",
"east",
"-",
"asian",
"tables",
"."
] | train | https://github.com/jquast/wcwidth/blob/78800b68911880ef4ef95ae83886154710441871/setup.py#L114-L122 |
jquast/wcwidth | setup.py | SetupUpdate._do_zero_width | def _do_zero_width(self):
"""Fetch and update zero width tables."""
self._do_retrieve(self.UCD_URL, self.UCD_IN)
(version, date, values) = self._parse_category(
fname=self.UCD_IN,
categories=('Me', 'Mn',)
)
table = self._make_table(values)
self._do... | python | def _do_zero_width(self):
"""Fetch and update zero width tables."""
self._do_retrieve(self.UCD_URL, self.UCD_IN)
(version, date, values) = self._parse_category(
fname=self.UCD_IN,
categories=('Me', 'Mn',)
)
table = self._make_table(values)
self._do... | [
"def",
"_do_zero_width",
"(",
"self",
")",
":",
"self",
".",
"_do_retrieve",
"(",
"self",
".",
"UCD_URL",
",",
"self",
".",
"UCD_IN",
")",
"(",
"version",
",",
"date",
",",
"values",
")",
"=",
"self",
".",
"_parse_category",
"(",
"fname",
"=",
"self",
... | Fetch and update zero width tables. | [
"Fetch",
"and",
"update",
"zero",
"width",
"tables",
"."
] | train | https://github.com/jquast/wcwidth/blob/78800b68911880ef4ef95ae83886154710441871/setup.py#L124-L132 |
jquast/wcwidth | setup.py | SetupUpdate._make_table | def _make_table(values):
"""Return a tuple of lookup tables for given values."""
import collections
table = collections.deque()
start, end = values[0], values[0]
for num, value in enumerate(values):
if num == 0:
table.append((value, value,))
... | python | def _make_table(values):
"""Return a tuple of lookup tables for given values."""
import collections
table = collections.deque()
start, end = values[0], values[0]
for num, value in enumerate(values):
if num == 0:
table.append((value, value,))
... | [
"def",
"_make_table",
"(",
"values",
")",
":",
"import",
"collections",
"table",
"=",
"collections",
".",
"deque",
"(",
")",
"start",
",",
"end",
"=",
"values",
"[",
"0",
"]",
",",
"values",
"[",
"0",
"]",
"for",
"num",
",",
"value",
"in",
"enumerate... | Return a tuple of lookup tables for given values. | [
"Return",
"a",
"tuple",
"of",
"lookup",
"tables",
"for",
"given",
"values",
"."
] | train | https://github.com/jquast/wcwidth/blob/78800b68911880ef4ef95ae83886154710441871/setup.py#L135-L150 |
jquast/wcwidth | setup.py | SetupUpdate._do_retrieve | def _do_retrieve(url, fname):
"""Retrieve given url to target filepath fname."""
folder = os.path.dirname(fname)
if not os.path.exists(folder):
os.makedirs(folder)
print("{}/ created.".format(folder))
if not os.path.exists(fname):
with open(fname, 'wb'... | python | def _do_retrieve(url, fname):
"""Retrieve given url to target filepath fname."""
folder = os.path.dirname(fname)
if not os.path.exists(folder):
os.makedirs(folder)
print("{}/ created.".format(folder))
if not os.path.exists(fname):
with open(fname, 'wb'... | [
"def",
"_do_retrieve",
"(",
"url",
",",
"fname",
")",
":",
"folder",
"=",
"os",
".",
"path",
".",
"dirname",
"(",
"fname",
")",
"if",
"not",
"os",
".",
"path",
".",
"exists",
"(",
"folder",
")",
":",
"os",
".",
"makedirs",
"(",
"folder",
")",
"pr... | Retrieve given url to target filepath fname. | [
"Retrieve",
"given",
"url",
"to",
"target",
"filepath",
"fname",
"."
] | train | https://github.com/jquast/wcwidth/blob/78800b68911880ef4ef95ae83886154710441871/setup.py#L153-L167 |
jquast/wcwidth | setup.py | SetupUpdate._parse_east_asian | def _parse_east_asian(fname, properties=(u'W', u'F',)):
"""Parse unicode east-asian width tables."""
version, date, values = None, None, []
print("parsing {} ..".format(fname))
for line in open(fname, 'rb'):
uline = line.decode('utf-8')
if version is None:
... | python | def _parse_east_asian(fname, properties=(u'W', u'F',)):
"""Parse unicode east-asian width tables."""
version, date, values = None, None, []
print("parsing {} ..".format(fname))
for line in open(fname, 'rb'):
uline = line.decode('utf-8')
if version is None:
... | [
"def",
"_parse_east_asian",
"(",
"fname",
",",
"properties",
"=",
"(",
"u'W'",
",",
"u'F'",
",",
")",
")",
":",
"version",
",",
"date",
",",
"values",
"=",
"None",
",",
"None",
",",
"[",
"]",
"print",
"(",
"\"parsing {} ..\"",
".",
"format",
"(",
"fn... | Parse unicode east-asian width tables. | [
"Parse",
"unicode",
"east",
"-",
"asian",
"width",
"tables",
"."
] | train | https://github.com/jquast/wcwidth/blob/78800b68911880ef4ef95ae83886154710441871/setup.py#L180-L201 |
jquast/wcwidth | setup.py | SetupUpdate._parse_category | def _parse_category(fname, categories):
"""Parse unicode category tables."""
version, date, values = None, None, []
print("parsing {} ..".format(fname))
for line in open(fname, 'rb'):
uline = line.decode('utf-8')
if version is None:
version = uline... | python | def _parse_category(fname, categories):
"""Parse unicode category tables."""
version, date, values = None, None, []
print("parsing {} ..".format(fname))
for line in open(fname, 'rb'):
uline = line.decode('utf-8')
if version is None:
version = uline... | [
"def",
"_parse_category",
"(",
"fname",
",",
"categories",
")",
":",
"version",
",",
"date",
",",
"values",
"=",
"None",
",",
"None",
",",
"[",
"]",
"print",
"(",
"\"parsing {} ..\"",
".",
"format",
"(",
"fname",
")",
")",
"for",
"line",
"in",
"open",
... | Parse unicode category tables. | [
"Parse",
"unicode",
"category",
"tables",
"."
] | train | https://github.com/jquast/wcwidth/blob/78800b68911880ef4ef95ae83886154710441871/setup.py#L204-L226 |
jquast/wcwidth | setup.py | SetupUpdate._do_write | def _do_write(fname, variable, version, date, table):
"""Write combining tables to filesystem as python code."""
# pylint: disable=R0914
# Too many local variables (19/15) (col 4)
print("writing {} ..".format(fname))
import unicodedata
import datetime
impo... | python | def _do_write(fname, variable, version, date, table):
"""Write combining tables to filesystem as python code."""
# pylint: disable=R0914
# Too many local variables (19/15) (col 4)
print("writing {} ..".format(fname))
import unicodedata
import datetime
impo... | [
"def",
"_do_write",
"(",
"fname",
",",
"variable",
",",
"version",
",",
"date",
",",
"table",
")",
":",
"# pylint: disable=R0914",
"# Too many local variables (19/15) (col 4)",
"print",
"(",
"\"writing {} ..\"",
".",
"format",
"(",
"fname",
")",
")",
"import... | Write combining tables to filesystem as python code. | [
"Write",
"combining",
"tables",
"to",
"filesystem",
"as",
"python",
"code",
"."
] | train | https://github.com/jquast/wcwidth/blob/78800b68911880ef4ef95ae83886154710441871/setup.py#L229-L268 |
jquast/wcwidth | bin/wcwidth-libc-comparator.py | report_ucs_msg | def report_ucs_msg(ucs, wcwidth_libc, wcwidth_local):
"""
Return string report of combining character differences.
:param ucs: unicode point.
:type ucs: unicode
:param wcwidth_libc: libc-wcwidth's reported character length.
:type comb_py: int
:param wcwidth_local: wcwidth's reported charact... | python | def report_ucs_msg(ucs, wcwidth_libc, wcwidth_local):
"""
Return string report of combining character differences.
:param ucs: unicode point.
:type ucs: unicode
:param wcwidth_libc: libc-wcwidth's reported character length.
:type comb_py: int
:param wcwidth_local: wcwidth's reported charact... | [
"def",
"report_ucs_msg",
"(",
"ucs",
",",
"wcwidth_libc",
",",
"wcwidth_local",
")",
":",
"ucp",
"=",
"(",
"ucs",
".",
"encode",
"(",
"'unicode_escape'",
")",
"[",
"2",
":",
"]",
".",
"decode",
"(",
"'ascii'",
")",
".",
"upper",
"(",
")",
".",
"lstri... | Return string report of combining character differences.
:param ucs: unicode point.
:type ucs: unicode
:param wcwidth_libc: libc-wcwidth's reported character length.
:type comb_py: int
:param wcwidth_local: wcwidth's reported character length.
:type comb_wc: int
:rtype: unicode | [
"Return",
"string",
"report",
"of",
"combining",
"character",
"differences",
"."
] | train | https://github.com/jquast/wcwidth/blob/78800b68911880ef4ef95ae83886154710441871/bin/wcwidth-libc-comparator.py#L44-L63 |
jquast/wcwidth | bin/wcwidth-libc-comparator.py | main | def main(using_locale=('en_US', 'UTF-8',)):
"""
Program entry point.
Load the entire Unicode table into memory, excluding those that:
- are not named (func unicodedata.name returns empty string),
- are combining characters.
Using ``locale``, for each unicode character string compare l... | python | def main(using_locale=('en_US', 'UTF-8',)):
"""
Program entry point.
Load the entire Unicode table into memory, excluding those that:
- are not named (func unicodedata.name returns empty string),
- are combining characters.
Using ``locale``, for each unicode character string compare l... | [
"def",
"main",
"(",
"using_locale",
"=",
"(",
"'en_US'",
",",
"'UTF-8'",
",",
")",
")",
":",
"all_ucs",
"=",
"(",
"ucs",
"for",
"ucs",
"in",
"[",
"unichr",
"(",
"val",
")",
"for",
"val",
"in",
"range",
"(",
"sys",
".",
"maxunicode",
")",
"]",
"if... | Program entry point.
Load the entire Unicode table into memory, excluding those that:
- are not named (func unicodedata.name returns empty string),
- are combining characters.
Using ``locale``, for each unicode character string compare libc's
wcwidth with local wcwidth.wcwidth() function;... | [
"Program",
"entry",
"point",
"."
] | train | https://github.com/jquast/wcwidth/blob/78800b68911880ef4ef95ae83886154710441871/bin/wcwidth-libc-comparator.py#L89-L123 |
jquast/wcwidth | bin/wcwidth-browser.py | validate_args | def validate_args(opts):
"""Validate and return options provided by docopt parsing."""
if opts['--wide'] is None:
opts['--wide'] = 2
else:
assert opts['--wide'] in ("1", "2"), opts['--wide']
if opts['--alignment'] is None:
opts['--alignment'] = 'left'
else:
assert opt... | python | def validate_args(opts):
"""Validate and return options provided by docopt parsing."""
if opts['--wide'] is None:
opts['--wide'] = 2
else:
assert opts['--wide'] in ("1", "2"), opts['--wide']
if opts['--alignment'] is None:
opts['--alignment'] = 'left'
else:
assert opt... | [
"def",
"validate_args",
"(",
"opts",
")",
":",
"if",
"opts",
"[",
"'--wide'",
"]",
"is",
"None",
":",
"opts",
"[",
"'--wide'",
"]",
"=",
"2",
"else",
":",
"assert",
"opts",
"[",
"'--wide'",
"]",
"in",
"(",
"\"1\"",
",",
"\"2\"",
")",
",",
"opts",
... | Validate and return options provided by docopt parsing. | [
"Validate",
"and",
"return",
"options",
"provided",
"by",
"docopt",
"parsing",
"."
] | train | https://github.com/jquast/wcwidth/blob/78800b68911880ef4ef95ae83886154710441871/bin/wcwidth-browser.py#L661-L675 |
jquast/wcwidth | bin/wcwidth-browser.py | main | def main(opts):
"""Program entry point."""
term = Terminal()
style = Style()
# if the terminal supports colors, use a Style instance with some
# standout colors (magenta, cyan).
if term.number_of_colors:
style = Style(attr_major=term.magenta,
attr_minor=term.bright... | python | def main(opts):
"""Program entry point."""
term = Terminal()
style = Style()
# if the terminal supports colors, use a Style instance with some
# standout colors (magenta, cyan).
if term.number_of_colors:
style = Style(attr_major=term.magenta,
attr_minor=term.bright... | [
"def",
"main",
"(",
"opts",
")",
":",
"term",
"=",
"Terminal",
"(",
")",
"style",
"=",
"Style",
"(",
")",
"# if the terminal supports colors, use a Style instance with some",
"# standout colors (magenta, cyan).",
"if",
"term",
".",
"number_of_colors",
":",
"style",
"=... | Program entry point. | [
"Program",
"entry",
"point",
"."
] | train | https://github.com/jquast/wcwidth/blob/78800b68911880ef4ef95ae83886154710441871/bin/wcwidth-browser.py#L678-L697 |
jquast/wcwidth | bin/wcwidth-browser.py | Screen.hint_width | def hint_width(self):
"""Width of a column segment."""
return sum((len(self.style.delimiter),
self.wide,
len(self.style.delimiter),
len(u' '),
UCS_PRINTLEN + 2,
len(u' '),
self.style.n... | python | def hint_width(self):
"""Width of a column segment."""
return sum((len(self.style.delimiter),
self.wide,
len(self.style.delimiter),
len(u' '),
UCS_PRINTLEN + 2,
len(u' '),
self.style.n... | [
"def",
"hint_width",
"(",
"self",
")",
":",
"return",
"sum",
"(",
"(",
"len",
"(",
"self",
".",
"style",
".",
"delimiter",
")",
",",
"self",
".",
"wide",
",",
"len",
"(",
"self",
".",
"style",
".",
"delimiter",
")",
",",
"len",
"(",
"u' '",
")",
... | Width of a column segment. | [
"Width",
"of",
"a",
"column",
"segment",
"."
] | train | https://github.com/jquast/wcwidth/blob/78800b68911880ef4ef95ae83886154710441871/bin/wcwidth-browser.py#L220-L228 |
jquast/wcwidth | bin/wcwidth-browser.py | Screen.head_item | def head_item(self):
"""Text of a single column heading."""
delimiter = self.style.attr_minor(self.style.delimiter)
hint = self.style.header_hint * self.wide
heading = (u'{delimiter}{hint}{delimiter}'
.format(delimiter=delimiter, hint=hint))
alignment = lambda ... | python | def head_item(self):
"""Text of a single column heading."""
delimiter = self.style.attr_minor(self.style.delimiter)
hint = self.style.header_hint * self.wide
heading = (u'{delimiter}{hint}{delimiter}'
.format(delimiter=delimiter, hint=hint))
alignment = lambda ... | [
"def",
"head_item",
"(",
"self",
")",
":",
"delimiter",
"=",
"self",
".",
"style",
".",
"attr_minor",
"(",
"self",
".",
"style",
".",
"delimiter",
")",
"hint",
"=",
"self",
".",
"style",
".",
"header_hint",
"*",
"self",
".",
"wide",
"heading",
"=",
"... | Text of a single column heading. | [
"Text",
"of",
"a",
"single",
"column",
"heading",
"."
] | train | https://github.com/jquast/wcwidth/blob/78800b68911880ef4ef95ae83886154710441871/bin/wcwidth-browser.py#L231-L241 |
jquast/wcwidth | bin/wcwidth-browser.py | Screen.msg_intro | def msg_intro(self):
"""Introductory message disabled above heading."""
delim = self.style.attr_minor(self.style.delimiter)
txt = self.intro_msg_fmt.format(delim=delim).rstrip()
return self.term.center(txt) | python | def msg_intro(self):
"""Introductory message disabled above heading."""
delim = self.style.attr_minor(self.style.delimiter)
txt = self.intro_msg_fmt.format(delim=delim).rstrip()
return self.term.center(txt) | [
"def",
"msg_intro",
"(",
"self",
")",
":",
"delim",
"=",
"self",
".",
"style",
".",
"attr_minor",
"(",
"self",
".",
"style",
".",
"delimiter",
")",
"txt",
"=",
"self",
".",
"intro_msg_fmt",
".",
"format",
"(",
"delim",
"=",
"delim",
")",
".",
"rstrip... | Introductory message disabled above heading. | [
"Introductory",
"message",
"disabled",
"above",
"heading",
"."
] | train | https://github.com/jquast/wcwidth/blob/78800b68911880ef4ef95ae83886154710441871/bin/wcwidth-browser.py#L244-L248 |
jquast/wcwidth | bin/wcwidth-browser.py | Screen.num_columns | def num_columns(self):
"""Number of columns displayed."""
if self.term.is_a_tty:
return self.term.width // self.hint_width
return 1 | python | def num_columns(self):
"""Number of columns displayed."""
if self.term.is_a_tty:
return self.term.width // self.hint_width
return 1 | [
"def",
"num_columns",
"(",
"self",
")",
":",
"if",
"self",
".",
"term",
".",
"is_a_tty",
":",
"return",
"self",
".",
"term",
".",
"width",
"//",
"self",
".",
"hint_width",
"return",
"1"
] | Number of columns displayed. | [
"Number",
"of",
"columns",
"displayed",
"."
] | train | https://github.com/jquast/wcwidth/blob/78800b68911880ef4ef95ae83886154710441871/bin/wcwidth-browser.py#L256-L260 |
jquast/wcwidth | bin/wcwidth-browser.py | Pager.on_resize | def on_resize(self, *args):
"""Signal handler callback for SIGWINCH."""
# pylint: disable=W0613
# Unused argument 'args'
self.screen.style.name_len = min(self.screen.style.name_len,
self.term.width - 15)
assert self.term.width >= s... | python | def on_resize(self, *args):
"""Signal handler callback for SIGWINCH."""
# pylint: disable=W0613
# Unused argument 'args'
self.screen.style.name_len = min(self.screen.style.name_len,
self.term.width - 15)
assert self.term.width >= s... | [
"def",
"on_resize",
"(",
"self",
",",
"*",
"args",
")",
":",
"# pylint: disable=W0613",
"# Unused argument 'args'",
"self",
".",
"screen",
".",
"style",
".",
"name_len",
"=",
"min",
"(",
"self",
".",
"screen",
".",
"style",
".",
"name_len",
",",
"sel... | Signal handler callback for SIGWINCH. | [
"Signal",
"handler",
"callback",
"for",
"SIGWINCH",
"."
] | train | https://github.com/jquast/wcwidth/blob/78800b68911880ef4ef95ae83886154710441871/bin/wcwidth-browser.py#L305-L315 |
jquast/wcwidth | bin/wcwidth-browser.py | Pager._set_lastpage | def _set_lastpage(self):
"""Calculate value of class attribute ``last_page``."""
self.last_page = (len(self._page_data) - 1) // self.screen.page_size | python | def _set_lastpage(self):
"""Calculate value of class attribute ``last_page``."""
self.last_page = (len(self._page_data) - 1) // self.screen.page_size | [
"def",
"_set_lastpage",
"(",
"self",
")",
":",
"self",
".",
"last_page",
"=",
"(",
"len",
"(",
"self",
".",
"_page_data",
")",
"-",
"1",
")",
"//",
"self",
".",
"screen",
".",
"page_size"
] | Calculate value of class attribute ``last_page``. | [
"Calculate",
"value",
"of",
"class",
"attribute",
"last_page",
"."
] | train | https://github.com/jquast/wcwidth/blob/78800b68911880ef4ef95ae83886154710441871/bin/wcwidth-browser.py#L317-L319 |
jquast/wcwidth | bin/wcwidth-browser.py | Pager.display_initialize | def display_initialize(self):
"""Display 'please wait' message, and narrow build warning."""
echo(self.term.home + self.term.clear)
echo(self.term.move_y(self.term.height // 2))
echo(self.term.center('Initializing page data ...').rstrip())
flushout()
if LIMIT_UCS == 0x10... | python | def display_initialize(self):
"""Display 'please wait' message, and narrow build warning."""
echo(self.term.home + self.term.clear)
echo(self.term.move_y(self.term.height // 2))
echo(self.term.center('Initializing page data ...').rstrip())
flushout()
if LIMIT_UCS == 0x10... | [
"def",
"display_initialize",
"(",
"self",
")",
":",
"echo",
"(",
"self",
".",
"term",
".",
"home",
"+",
"self",
".",
"term",
".",
"clear",
")",
"echo",
"(",
"self",
".",
"term",
".",
"move_y",
"(",
"self",
".",
"term",
".",
"height",
"//",
"2",
"... | Display 'please wait' message, and narrow build warning. | [
"Display",
"please",
"wait",
"message",
"and",
"narrow",
"build",
"warning",
"."
] | train | https://github.com/jquast/wcwidth/blob/78800b68911880ef4ef95ae83886154710441871/bin/wcwidth-browser.py#L321-L334 |
jquast/wcwidth | bin/wcwidth-browser.py | Pager.initialize_page_data | def initialize_page_data(self):
"""Initialize the page data for the given screen."""
if self.term.is_a_tty:
self.display_initialize()
self.character_generator = self.character_factory(self.screen.wide)
page_data = list()
while True:
try:
pa... | python | def initialize_page_data(self):
"""Initialize the page data for the given screen."""
if self.term.is_a_tty:
self.display_initialize()
self.character_generator = self.character_factory(self.screen.wide)
page_data = list()
while True:
try:
pa... | [
"def",
"initialize_page_data",
"(",
"self",
")",
":",
"if",
"self",
".",
"term",
".",
"is_a_tty",
":",
"self",
".",
"display_initialize",
"(",
")",
"self",
".",
"character_generator",
"=",
"self",
".",
"character_factory",
"(",
"self",
".",
"screen",
".",
... | Initialize the page data for the given screen. | [
"Initialize",
"the",
"page",
"data",
"for",
"the",
"given",
"screen",
"."
] | train | https://github.com/jquast/wcwidth/blob/78800b68911880ef4ef95ae83886154710441871/bin/wcwidth-browser.py#L336-L351 |
jquast/wcwidth | bin/wcwidth-browser.py | Pager.page_data | def page_data(self, idx, offset):
"""
Return character data for page of given index and offset.
:param idx: page index.
:type idx: int
:param offset: scrolling region offset of current page.
:type offset: int
:returns: list of tuples in form of ``(ucs, name)``
... | python | def page_data(self, idx, offset):
"""
Return character data for page of given index and offset.
:param idx: page index.
:type idx: int
:param offset: scrolling region offset of current page.
:type offset: int
:returns: list of tuples in form of ``(ucs, name)``
... | [
"def",
"page_data",
"(",
"self",
",",
"idx",
",",
"offset",
")",
":",
"size",
"=",
"self",
".",
"screen",
".",
"page_size",
"while",
"offset",
"<",
"0",
"and",
"idx",
":",
"offset",
"+=",
"size",
"idx",
"-=",
"1",
"offset",
"=",
"max",
"(",
"0",
... | Return character data for page of given index and offset.
:param idx: page index.
:type idx: int
:param offset: scrolling region offset of current page.
:type offset: int
:returns: list of tuples in form of ``(ucs, name)``
:rtype: list[(unicode, unicode)] | [
"Return",
"character",
"data",
"for",
"page",
"of",
"given",
"index",
"and",
"offset",
"."
] | train | https://github.com/jquast/wcwidth/blob/78800b68911880ef4ef95ae83886154710441871/bin/wcwidth-browser.py#L353-L381 |
jquast/wcwidth | bin/wcwidth-browser.py | Pager._run_notty | def _run_notty(self, writer):
"""Pager run method for terminals that are not a tty."""
page_idx = page_offset = 0
while True:
npage_idx, _ = self.draw(writer, page_idx + 1, page_offset)
if npage_idx == self.last_page:
# page displayed was last page, quit.
... | python | def _run_notty(self, writer):
"""Pager run method for terminals that are not a tty."""
page_idx = page_offset = 0
while True:
npage_idx, _ = self.draw(writer, page_idx + 1, page_offset)
if npage_idx == self.last_page:
# page displayed was last page, quit.
... | [
"def",
"_run_notty",
"(",
"self",
",",
"writer",
")",
":",
"page_idx",
"=",
"page_offset",
"=",
"0",
"while",
"True",
":",
"npage_idx",
",",
"_",
"=",
"self",
".",
"draw",
"(",
"writer",
",",
"page_idx",
"+",
"1",
",",
"page_offset",
")",
"if",
"npag... | Pager run method for terminals that are not a tty. | [
"Pager",
"run",
"method",
"for",
"terminals",
"that",
"are",
"not",
"a",
"tty",
"."
] | train | https://github.com/jquast/wcwidth/blob/78800b68911880ef4ef95ae83886154710441871/bin/wcwidth-browser.py#L383-L393 |
jquast/wcwidth | bin/wcwidth-browser.py | Pager._run_tty | def _run_tty(self, writer, reader):
"""Pager run method for terminals that are a tty."""
# allow window-change signal to reflow screen
signal.signal(signal.SIGWINCH, self.on_resize)
page_idx = page_offset = 0
while True:
if self.dirty:
page_idx, page_... | python | def _run_tty(self, writer, reader):
"""Pager run method for terminals that are a tty."""
# allow window-change signal to reflow screen
signal.signal(signal.SIGWINCH, self.on_resize)
page_idx = page_offset = 0
while True:
if self.dirty:
page_idx, page_... | [
"def",
"_run_tty",
"(",
"self",
",",
"writer",
",",
"reader",
")",
":",
"# allow window-change signal to reflow screen",
"signal",
".",
"signal",
"(",
"signal",
".",
"SIGWINCH",
",",
"self",
".",
"on_resize",
")",
"page_idx",
"=",
"page_offset",
"=",
"0",
"whi... | Pager run method for terminals that are a tty. | [
"Pager",
"run",
"method",
"for",
"terminals",
"that",
"are",
"a",
"tty",
"."
] | train | https://github.com/jquast/wcwidth/blob/78800b68911880ef4ef95ae83886154710441871/bin/wcwidth-browser.py#L395-L416 |
jquast/wcwidth | bin/wcwidth-browser.py | Pager.run | def run(self, writer, reader):
"""
Pager entry point.
In interactive mode (terminal is a tty), run until
``process_keystroke()`` detects quit keystroke ('q'). In
non-interactive mode, exit after displaying all unicode points.
:param writer: callable writes to output st... | python | def run(self, writer, reader):
"""
Pager entry point.
In interactive mode (terminal is a tty), run until
``process_keystroke()`` detects quit keystroke ('q'). In
non-interactive mode, exit after displaying all unicode points.
:param writer: callable writes to output st... | [
"def",
"run",
"(",
"self",
",",
"writer",
",",
"reader",
")",
":",
"self",
".",
"_page_data",
"=",
"self",
".",
"initialize_page_data",
"(",
")",
"self",
".",
"_set_lastpage",
"(",
")",
"if",
"not",
"self",
".",
"term",
".",
"is_a_tty",
":",
"self",
... | Pager entry point.
In interactive mode (terminal is a tty), run until
``process_keystroke()`` detects quit keystroke ('q'). In
non-interactive mode, exit after displaying all unicode points.
:param writer: callable writes to output stream, receiving unicode.
:type writer: call... | [
"Pager",
"entry",
"point",
"."
] | train | https://github.com/jquast/wcwidth/blob/78800b68911880ef4ef95ae83886154710441871/bin/wcwidth-browser.py#L418-L437 |
jquast/wcwidth | bin/wcwidth-browser.py | Pager.process_keystroke | def process_keystroke(self, inp, idx, offset):
"""
Process keystroke ``inp``, adjusting screen parameters.
:param inp: return value of Terminal.inkey().
:type inp: blessed.keyboard.Keystroke
:param idx: page index.
:type idx: int
:param offset: scrolling region o... | python | def process_keystroke(self, inp, idx, offset):
"""
Process keystroke ``inp``, adjusting screen parameters.
:param inp: return value of Terminal.inkey().
:type inp: blessed.keyboard.Keystroke
:param idx: page index.
:type idx: int
:param offset: scrolling region o... | [
"def",
"process_keystroke",
"(",
"self",
",",
"inp",
",",
"idx",
",",
"offset",
")",
":",
"if",
"inp",
".",
"lower",
"(",
")",
"in",
"(",
"u'q'",
",",
"u'Q'",
")",
":",
"# exit",
"return",
"(",
"-",
"1",
",",
"-",
"1",
")",
"self",
".",
"_proce... | Process keystroke ``inp``, adjusting screen parameters.
:param inp: return value of Terminal.inkey().
:type inp: blessed.keyboard.Keystroke
:param idx: page index.
:type idx: int
:param offset: scrolling region offset of current page.
:type offset: int
:returns: ... | [
"Process",
"keystroke",
"inp",
"adjusting",
"screen",
"parameters",
"."
] | train | https://github.com/jquast/wcwidth/blob/78800b68911880ef4ef95ae83886154710441871/bin/wcwidth-browser.py#L439-L457 |
jquast/wcwidth | bin/wcwidth-browser.py | Pager._process_keystroke_commands | def _process_keystroke_commands(self, inp):
"""Process keystrokes that issue commands (side effects)."""
if inp in (u'1', u'2'):
# chose 1 or 2-character wide
if int(inp) != self.screen.wide:
self.screen.wide = int(inp)
self.on_resize(None, None)
... | python | def _process_keystroke_commands(self, inp):
"""Process keystrokes that issue commands (side effects)."""
if inp in (u'1', u'2'):
# chose 1 or 2-character wide
if int(inp) != self.screen.wide:
self.screen.wide = int(inp)
self.on_resize(None, None)
... | [
"def",
"_process_keystroke_commands",
"(",
"self",
",",
"inp",
")",
":",
"if",
"inp",
"in",
"(",
"u'1'",
",",
"u'2'",
")",
":",
"# chose 1 or 2-character wide",
"if",
"int",
"(",
"inp",
")",
"!=",
"self",
".",
"screen",
".",
"wide",
":",
"self",
".",
"... | Process keystrokes that issue commands (side effects). | [
"Process",
"keystrokes",
"that",
"issue",
"commands",
"(",
"side",
"effects",
")",
"."
] | train | https://github.com/jquast/wcwidth/blob/78800b68911880ef4ef95ae83886154710441871/bin/wcwidth-browser.py#L459-L481 |
jquast/wcwidth | bin/wcwidth-browser.py | Pager._process_keystroke_movement | def _process_keystroke_movement(self, inp, idx, offset):
"""Process keystrokes that adjust index and offset."""
term = self.term
if inp in (u'y', u'k') or inp.code in (term.KEY_UP,):
# scroll backward 1 line
idx, offset = (idx, offset - self.screen.num_columns)
el... | python | def _process_keystroke_movement(self, inp, idx, offset):
"""Process keystrokes that adjust index and offset."""
term = self.term
if inp in (u'y', u'k') or inp.code in (term.KEY_UP,):
# scroll backward 1 line
idx, offset = (idx, offset - self.screen.num_columns)
el... | [
"def",
"_process_keystroke_movement",
"(",
"self",
",",
"inp",
",",
"idx",
",",
"offset",
")",
":",
"term",
"=",
"self",
".",
"term",
"if",
"inp",
"in",
"(",
"u'y'",
",",
"u'k'",
")",
"or",
"inp",
".",
"code",
"in",
"(",
"term",
".",
"KEY_UP",
",",... | Process keystrokes that adjust index and offset. | [
"Process",
"keystrokes",
"that",
"adjust",
"index",
"and",
"offset",
"."
] | train | https://github.com/jquast/wcwidth/blob/78800b68911880ef4ef95ae83886154710441871/bin/wcwidth-browser.py#L483-L511 |
jquast/wcwidth | bin/wcwidth-browser.py | Pager.draw | def draw(self, writer, idx, offset):
"""
Draw the current page view to ``writer``.
:param writer: callable writes to output stream, receiving unicode.
:type writer: callable
:param idx: current page index.
:type idx: int
:param offset: scrolling region offset of ... | python | def draw(self, writer, idx, offset):
"""
Draw the current page view to ``writer``.
:param writer: callable writes to output stream, receiving unicode.
:type writer: callable
:param idx: current page index.
:type idx: int
:param offset: scrolling region offset of ... | [
"def",
"draw",
"(",
"self",
",",
"writer",
",",
"idx",
",",
"offset",
")",
":",
"# as our screen can be resized while we're mid-calculation,",
"# our self.dirty flag can become re-toggled; because we are",
"# not re-flowing our pagination, we must begin over again.",
"while",
"self",... | Draw the current page view to ``writer``.
:param writer: callable writes to output stream, receiving unicode.
:type writer: callable
:param idx: current page index.
:type idx: int
:param offset: scrolling region offset of current page.
:type offset: int
:returns:... | [
"Draw",
"the",
"current",
"page",
"view",
"to",
"writer",
"."
] | train | https://github.com/jquast/wcwidth/blob/78800b68911880ef4ef95ae83886154710441871/bin/wcwidth-browser.py#L513-L537 |
jquast/wcwidth | bin/wcwidth-browser.py | Pager.draw_heading | def draw_heading(self, writer):
"""
Conditionally redraw screen when ``dirty`` attribute is valued REFRESH.
When Pager attribute ``dirty`` is ``STATE_REFRESH``, cursor is moved
to (0,0), screen is cleared, and heading is displayed.
:param writer: callable writes to output strea... | python | def draw_heading(self, writer):
"""
Conditionally redraw screen when ``dirty`` attribute is valued REFRESH.
When Pager attribute ``dirty`` is ``STATE_REFRESH``, cursor is moved
to (0,0), screen is cleared, and heading is displayed.
:param writer: callable writes to output strea... | [
"def",
"draw_heading",
"(",
"self",
",",
"writer",
")",
":",
"if",
"self",
".",
"dirty",
"==",
"self",
".",
"STATE_REFRESH",
":",
"writer",
"(",
"u''",
".",
"join",
"(",
"(",
"self",
".",
"term",
".",
"home",
",",
"self",
".",
"term",
".",
"clear",... | Conditionally redraw screen when ``dirty`` attribute is valued REFRESH.
When Pager attribute ``dirty`` is ``STATE_REFRESH``, cursor is moved
to (0,0), screen is cleared, and heading is displayed.
:param writer: callable writes to output stream, receiving unicode.
:returns: True if clas... | [
"Conditionally",
"redraw",
"screen",
"when",
"dirty",
"attribute",
"is",
"valued",
"REFRESH",
"."
] | train | https://github.com/jquast/wcwidth/blob/78800b68911880ef4ef95ae83886154710441871/bin/wcwidth-browser.py#L539-L554 |
jquast/wcwidth | bin/wcwidth-browser.py | Pager.draw_status | def draw_status(self, writer, idx):
"""
Conditionally draw status bar when output terminal is a tty.
:param writer: callable writes to output stream, receiving unicode.
:param idx: current page position index.
:type idx: int
"""
if self.term.is_a_tty:
... | python | def draw_status(self, writer, idx):
"""
Conditionally draw status bar when output terminal is a tty.
:param writer: callable writes to output stream, receiving unicode.
:param idx: current page position index.
:type idx: int
"""
if self.term.is_a_tty:
... | [
"def",
"draw_status",
"(",
"self",
",",
"writer",
",",
"idx",
")",
":",
"if",
"self",
".",
"term",
".",
"is_a_tty",
":",
"writer",
"(",
"self",
".",
"term",
".",
"hide_cursor",
"(",
")",
")",
"style",
"=",
"self",
".",
"screen",
".",
"style",
"writ... | Conditionally draw status bar when output terminal is a tty.
:param writer: callable writes to output stream, receiving unicode.
:param idx: current page position index.
:type idx: int | [
"Conditionally",
"draw",
"status",
"bar",
"when",
"output",
"terminal",
"is",
"a",
"tty",
"."
] | train | https://github.com/jquast/wcwidth/blob/78800b68911880ef4ef95ae83886154710441871/bin/wcwidth-browser.py#L556-L578 |
jquast/wcwidth | bin/wcwidth-browser.py | Pager.page_view | def page_view(self, data):
"""
Generator yields text to be displayed for the current unicode pageview.
:param data: The current page's data as tuple of ``(ucs, name)``.
:rtype: generator
"""
if self.term.is_a_tty:
yield self.term.move(self.screen.row_begins, ... | python | def page_view(self, data):
"""
Generator yields text to be displayed for the current unicode pageview.
:param data: The current page's data as tuple of ``(ucs, name)``.
:rtype: generator
"""
if self.term.is_a_tty:
yield self.term.move(self.screen.row_begins, ... | [
"def",
"page_view",
"(",
"self",
",",
"data",
")",
":",
"if",
"self",
".",
"term",
".",
"is_a_tty",
":",
"yield",
"self",
".",
"term",
".",
"move",
"(",
"self",
".",
"screen",
".",
"row_begins",
",",
"0",
")",
"# sequence clears to end-of-line",
"clear_e... | Generator yields text to be displayed for the current unicode pageview.
:param data: The current page's data as tuple of ``(ucs, name)``.
:rtype: generator | [
"Generator",
"yields",
"text",
"to",
"be",
"displayed",
"for",
"the",
"current",
"unicode",
"pageview",
"."
] | train | https://github.com/jquast/wcwidth/blob/78800b68911880ef4ef95ae83886154710441871/bin/wcwidth-browser.py#L580-L615 |
jquast/wcwidth | bin/wcwidth-browser.py | Pager.text_entry | def text_entry(self, ucs, name):
"""
Display a single column segment row describing ``(ucs, name)``.
:param ucs: target unicode point character string.
:param name: name of unicode point.
:rtype: unicode
"""
style = self.screen.style
if len(name) > style.... | python | def text_entry(self, ucs, name):
"""
Display a single column segment row describing ``(ucs, name)``.
:param ucs: target unicode point character string.
:param name: name of unicode point.
:rtype: unicode
"""
style = self.screen.style
if len(name) > style.... | [
"def",
"text_entry",
"(",
"self",
",",
"ucs",
",",
"name",
")",
":",
"style",
"=",
"self",
".",
"screen",
".",
"style",
"if",
"len",
"(",
"name",
")",
">",
"style",
".",
"name_len",
":",
"idx",
"=",
"max",
"(",
"0",
",",
"style",
".",
"name_len",... | Display a single column segment row describing ``(ucs, name)``.
:param ucs: target unicode point character string.
:param name: name of unicode point.
:rtype: unicode | [
"Display",
"a",
"single",
"column",
"segment",
"row",
"describing",
"(",
"ucs",
"name",
")",
"."
] | train | https://github.com/jquast/wcwidth/blob/78800b68911880ef4ef95ae83886154710441871/bin/wcwidth-browser.py#L617-L658 |
matthew-brett/delocate | delocate/tools.py | back_tick | def back_tick(cmd, ret_err=False, as_str=True, raise_err=None):
""" Run command `cmd`, return stdout, or stdout, stderr if `ret_err`
Roughly equivalent to ``check_output`` in Python 2.7
Parameters
----------
cmd : sequence
command to execute
ret_err : bool, optional
If True, re... | python | def back_tick(cmd, ret_err=False, as_str=True, raise_err=None):
""" Run command `cmd`, return stdout, or stdout, stderr if `ret_err`
Roughly equivalent to ``check_output`` in Python 2.7
Parameters
----------
cmd : sequence
command to execute
ret_err : bool, optional
If True, re... | [
"def",
"back_tick",
"(",
"cmd",
",",
"ret_err",
"=",
"False",
",",
"as_str",
"=",
"True",
",",
"raise_err",
"=",
"None",
")",
":",
"if",
"raise_err",
"is",
"None",
":",
"raise_err",
"=",
"False",
"if",
"ret_err",
"else",
"True",
"cmd_is_seq",
"=",
"isi... | Run command `cmd`, return stdout, or stdout, stderr if `ret_err`
Roughly equivalent to ``check_output`` in Python 2.7
Parameters
----------
cmd : sequence
command to execute
ret_err : bool, optional
If True, return stderr in addition to stdout. If False, just return
stdout... | [
"Run",
"command",
"cmd",
"return",
"stdout",
"or",
"stdout",
"stderr",
"if",
"ret_err"
] | train | https://github.com/matthew-brett/delocate/blob/ed48de15fce31c3f52f1a9f32cae1b02fc55aa60/delocate/tools.py#L17-L69 |
matthew-brett/delocate | delocate/tools.py | unique_by_index | def unique_by_index(sequence):
""" unique elements in `sequence` in the order in which they occur
Parameters
----------
sequence : iterable
Returns
-------
uniques : list
unique elements of sequence, ordered by the order in which the element
occurs in `sequence`
"""
... | python | def unique_by_index(sequence):
""" unique elements in `sequence` in the order in which they occur
Parameters
----------
sequence : iterable
Returns
-------
uniques : list
unique elements of sequence, ordered by the order in which the element
occurs in `sequence`
"""
... | [
"def",
"unique_by_index",
"(",
"sequence",
")",
":",
"uniques",
"=",
"[",
"]",
"for",
"element",
"in",
"sequence",
":",
"if",
"element",
"not",
"in",
"uniques",
":",
"uniques",
".",
"append",
"(",
"element",
")",
"return",
"uniques"
] | unique elements in `sequence` in the order in which they occur
Parameters
----------
sequence : iterable
Returns
-------
uniques : list
unique elements of sequence, ordered by the order in which the element
occurs in `sequence` | [
"unique",
"elements",
"in",
"sequence",
"in",
"the",
"order",
"in",
"which",
"they",
"occur"
] | train | https://github.com/matthew-brett/delocate/blob/ed48de15fce31c3f52f1a9f32cae1b02fc55aa60/delocate/tools.py#L72-L89 |
matthew-brett/delocate | delocate/tools.py | ensure_permissions | def ensure_permissions(mode_flags=stat.S_IWUSR):
"""decorator to ensure a filename has given permissions.
If changed, original permissions are restored after the decorated
modification.
"""
def decorator(f):
def modify(filename, *args, **kwargs):
m = chmod_perms(filename) if ex... | python | def ensure_permissions(mode_flags=stat.S_IWUSR):
"""decorator to ensure a filename has given permissions.
If changed, original permissions are restored after the decorated
modification.
"""
def decorator(f):
def modify(filename, *args, **kwargs):
m = chmod_perms(filename) if ex... | [
"def",
"ensure_permissions",
"(",
"mode_flags",
"=",
"stat",
".",
"S_IWUSR",
")",
":",
"def",
"decorator",
"(",
"f",
")",
":",
"def",
"modify",
"(",
"filename",
",",
"*",
"args",
",",
"*",
"*",
"kwargs",
")",
":",
"m",
"=",
"chmod_perms",
"(",
"filen... | decorator to ensure a filename has given permissions.
If changed, original permissions are restored after the decorated
modification. | [
"decorator",
"to",
"ensure",
"a",
"filename",
"has",
"given",
"permissions",
"."
] | train | https://github.com/matthew-brett/delocate/blob/ed48de15fce31c3f52f1a9f32cae1b02fc55aa60/delocate/tools.py#L97-L117 |
matthew-brett/delocate | delocate/tools.py | get_install_names | def get_install_names(filename):
""" Return install names from library named in `filename`
Returns tuple of install names
tuple will be empty if no install names, or if this is not an object file.
Parameters
----------
filename : str
filename of library
Returns
-------
in... | python | def get_install_names(filename):
""" Return install names from library named in `filename`
Returns tuple of install names
tuple will be empty if no install names, or if this is not an object file.
Parameters
----------
filename : str
filename of library
Returns
-------
in... | [
"def",
"get_install_names",
"(",
"filename",
")",
":",
"lines",
"=",
"_cmd_out_err",
"(",
"[",
"'otool'",
",",
"'-L'",
",",
"filename",
"]",
")",
"if",
"not",
"_line0_says_object",
"(",
"lines",
"[",
"0",
"]",
",",
"filename",
")",
":",
"return",
"(",
... | Return install names from library named in `filename`
Returns tuple of install names
tuple will be empty if no install names, or if this is not an object file.
Parameters
----------
filename : str
filename of library
Returns
-------
install_names : tuple
tuple of inst... | [
"Return",
"install",
"names",
"from",
"library",
"named",
"in",
"filename"
] | train | https://github.com/matthew-brett/delocate/blob/ed48de15fce31c3f52f1a9f32cae1b02fc55aa60/delocate/tools.py#L197-L222 |
matthew-brett/delocate | delocate/tools.py | get_install_id | def get_install_id(filename):
""" Return install id from library named in `filename`
Returns None if no install id, or if this is not an object file.
Parameters
----------
filename : str
filename of library
Returns
-------
install_id : str
install id of library `filena... | python | def get_install_id(filename):
""" Return install id from library named in `filename`
Returns None if no install id, or if this is not an object file.
Parameters
----------
filename : str
filename of library
Returns
-------
install_id : str
install id of library `filena... | [
"def",
"get_install_id",
"(",
"filename",
")",
":",
"lines",
"=",
"_cmd_out_err",
"(",
"[",
"'otool'",
",",
"'-D'",
",",
"filename",
"]",
")",
"if",
"not",
"_line0_says_object",
"(",
"lines",
"[",
"0",
"]",
",",
"filename",
")",
":",
"return",
"None",
... | Return install id from library named in `filename`
Returns None if no install id, or if this is not an object file.
Parameters
----------
filename : str
filename of library
Returns
-------
install_id : str
install id of library `filename`, or None if no install id | [
"Return",
"install",
"id",
"from",
"library",
"named",
"in",
"filename"
] | train | https://github.com/matthew-brett/delocate/blob/ed48de15fce31c3f52f1a9f32cae1b02fc55aa60/delocate/tools.py#L225-L247 |
matthew-brett/delocate | delocate/tools.py | set_install_name | def set_install_name(filename, oldname, newname):
""" Set install name `oldname` to `newname` in library filename
Parameters
----------
filename : str
filename of library
oldname : str
current install name in library
newname : str
replacement name for `oldname`
"""
... | python | def set_install_name(filename, oldname, newname):
""" Set install name `oldname` to `newname` in library filename
Parameters
----------
filename : str
filename of library
oldname : str
current install name in library
newname : str
replacement name for `oldname`
"""
... | [
"def",
"set_install_name",
"(",
"filename",
",",
"oldname",
",",
"newname",
")",
":",
"names",
"=",
"get_install_names",
"(",
"filename",
")",
"if",
"oldname",
"not",
"in",
"names",
":",
"raise",
"InstallNameError",
"(",
"'{0} not in install names for {1}'",
".",
... | Set install name `oldname` to `newname` in library filename
Parameters
----------
filename : str
filename of library
oldname : str
current install name in library
newname : str
replacement name for `oldname` | [
"Set",
"install",
"name",
"oldname",
"to",
"newname",
"in",
"library",
"filename"
] | train | https://github.com/matthew-brett/delocate/blob/ed48de15fce31c3f52f1a9f32cae1b02fc55aa60/delocate/tools.py#L251-L267 |
matthew-brett/delocate | delocate/tools.py | set_install_id | def set_install_id(filename, install_id):
""" Set install id for library named in `filename`
Parameters
----------
filename : str
filename of library
install_id : str
install id for library `filename`
Raises
------
RuntimeError if `filename` has not install id
"""
... | python | def set_install_id(filename, install_id):
""" Set install id for library named in `filename`
Parameters
----------
filename : str
filename of library
install_id : str
install id for library `filename`
Raises
------
RuntimeError if `filename` has not install id
"""
... | [
"def",
"set_install_id",
"(",
"filename",
",",
"install_id",
")",
":",
"if",
"get_install_id",
"(",
"filename",
")",
"is",
"None",
":",
"raise",
"InstallNameError",
"(",
"'{0} has no install id'",
".",
"format",
"(",
"filename",
")",
")",
"back_tick",
"(",
"["... | Set install id for library named in `filename`
Parameters
----------
filename : str
filename of library
install_id : str
install id for library `filename`
Raises
------
RuntimeError if `filename` has not install id | [
"Set",
"install",
"id",
"for",
"library",
"named",
"in",
"filename"
] | train | https://github.com/matthew-brett/delocate/blob/ed48de15fce31c3f52f1a9f32cae1b02fc55aa60/delocate/tools.py#L271-L287 |
matthew-brett/delocate | delocate/tools.py | get_rpaths | def get_rpaths(filename):
""" Return a tuple of rpaths from the library `filename`
If `filename` is not a library then the returned tuple will be empty.
Parameters
----------
filaname : str
filename of library
Returns
-------
rpath : tuple
rpath paths in `filename`
... | python | def get_rpaths(filename):
""" Return a tuple of rpaths from the library `filename`
If `filename` is not a library then the returned tuple will be empty.
Parameters
----------
filaname : str
filename of library
Returns
-------
rpath : tuple
rpath paths in `filename`
... | [
"def",
"get_rpaths",
"(",
"filename",
")",
":",
"try",
":",
"lines",
"=",
"_cmd_out_err",
"(",
"[",
"'otool'",
",",
"'-l'",
",",
"filename",
"]",
")",
"except",
"RuntimeError",
":",
"return",
"(",
")",
"if",
"not",
"_line0_says_object",
"(",
"lines",
"["... | Return a tuple of rpaths from the library `filename`
If `filename` is not a library then the returned tuple will be empty.
Parameters
----------
filaname : str
filename of library
Returns
-------
rpath : tuple
rpath paths in `filename` | [
"Return",
"a",
"tuple",
"of",
"rpaths",
"from",
"the",
"library",
"filename"
] | train | https://github.com/matthew-brett/delocate/blob/ed48de15fce31c3f52f1a9f32cae1b02fc55aa60/delocate/tools.py#L292-L325 |
matthew-brett/delocate | delocate/tools.py | dir2zip | def dir2zip(in_dir, zip_fname):
""" Make a zip file `zip_fname` with contents of directory `in_dir`
The recorded filenames are relative to `in_dir`, so doing a standard zip
unpack of the resulting `zip_fname` in an empty directory will result in
the original directory contents.
Parameters
----... | python | def dir2zip(in_dir, zip_fname):
""" Make a zip file `zip_fname` with contents of directory `in_dir`
The recorded filenames are relative to `in_dir`, so doing a standard zip
unpack of the resulting `zip_fname` in an empty directory will result in
the original directory contents.
Parameters
----... | [
"def",
"dir2zip",
"(",
"in_dir",
",",
"zip_fname",
")",
":",
"z",
"=",
"zipfile",
".",
"ZipFile",
"(",
"zip_fname",
",",
"'w'",
",",
"compression",
"=",
"zipfile",
".",
"ZIP_DEFLATED",
")",
"for",
"root",
",",
"dirs",
",",
"files",
"in",
"os",
".",
"... | Make a zip file `zip_fname` with contents of directory `in_dir`
The recorded filenames are relative to `in_dir`, so doing a standard zip
unpack of the resulting `zip_fname` in an empty directory will result in
the original directory contents.
Parameters
----------
in_dir : str
Director... | [
"Make",
"a",
"zip",
"file",
"zip_fname",
"with",
"contents",
"of",
"directory",
"in_dir"
] | train | https://github.com/matthew-brett/delocate/blob/ed48de15fce31c3f52f1a9f32cae1b02fc55aa60/delocate/tools.py#L357-L393 |
matthew-brett/delocate | delocate/tools.py | find_package_dirs | def find_package_dirs(root_path):
""" Find python package directories in directory `root_path`
Parameters
----------
root_path : str
Directory to search for package subdirectories
Returns
-------
package_sdirs : set
Set of strings where each is a subdirectory of `root_path`... | python | def find_package_dirs(root_path):
""" Find python package directories in directory `root_path`
Parameters
----------
root_path : str
Directory to search for package subdirectories
Returns
-------
package_sdirs : set
Set of strings where each is a subdirectory of `root_path`... | [
"def",
"find_package_dirs",
"(",
"root_path",
")",
":",
"package_sdirs",
"=",
"set",
"(",
")",
"for",
"entry",
"in",
"os",
".",
"listdir",
"(",
"root_path",
")",
":",
"fname",
"=",
"entry",
"if",
"root_path",
"==",
"'.'",
"else",
"pjoin",
"(",
"root_path... | Find python package directories in directory `root_path`
Parameters
----------
root_path : str
Directory to search for package subdirectories
Returns
-------
package_sdirs : set
Set of strings where each is a subdirectory of `root_path`, containing
an ``__init__.py`` fi... | [
"Find",
"python",
"package",
"directories",
"in",
"directory",
"root_path"
] | train | https://github.com/matthew-brett/delocate/blob/ed48de15fce31c3f52f1a9f32cae1b02fc55aa60/delocate/tools.py#L396-L415 |
matthew-brett/delocate | delocate/tools.py | cmp_contents | def cmp_contents(filename1, filename2):
""" Returns True if contents of the files are the same
Parameters
----------
filename1 : str
filename of first file to compare
filename2 : str
filename of second file to compare
Returns
-------
tf : bool
True if binary con... | python | def cmp_contents(filename1, filename2):
""" Returns True if contents of the files are the same
Parameters
----------
filename1 : str
filename of first file to compare
filename2 : str
filename of second file to compare
Returns
-------
tf : bool
True if binary con... | [
"def",
"cmp_contents",
"(",
"filename1",
",",
"filename2",
")",
":",
"with",
"open_readable",
"(",
"filename1",
",",
"'rb'",
")",
"as",
"fobj",
":",
"contents1",
"=",
"fobj",
".",
"read",
"(",
")",
"with",
"open_readable",
"(",
"filename2",
",",
"'rb'",
... | Returns True if contents of the files are the same
Parameters
----------
filename1 : str
filename of first file to compare
filename2 : str
filename of second file to compare
Returns
-------
tf : bool
True if binary contents of `filename1` is same as binary contents ... | [
"Returns",
"True",
"if",
"contents",
"of",
"the",
"files",
"are",
"the",
"same"
] | train | https://github.com/matthew-brett/delocate/blob/ed48de15fce31c3f52f1a9f32cae1b02fc55aa60/delocate/tools.py#L418-L438 |
matthew-brett/delocate | delocate/tools.py | get_archs | def get_archs(libname):
""" Return architecture types from library `libname`
Parameters
----------
libname : str
filename of binary for which to return arch codes
Returns
-------
arch_names : frozenset
Empty (frozen)set if no arch codes. If not empty, contains one or more
... | python | def get_archs(libname):
""" Return architecture types from library `libname`
Parameters
----------
libname : str
filename of binary for which to return arch codes
Returns
-------
arch_names : frozenset
Empty (frozen)set if no arch codes. If not empty, contains one or more
... | [
"def",
"get_archs",
"(",
"libname",
")",
":",
"if",
"not",
"exists",
"(",
"libname",
")",
":",
"raise",
"RuntimeError",
"(",
"libname",
"+",
"\" is not a file\"",
")",
"try",
":",
"stdout",
"=",
"back_tick",
"(",
"[",
"'lipo'",
",",
"'-info'",
",",
"libn... | Return architecture types from library `libname`
Parameters
----------
libname : str
filename of binary for which to return arch codes
Returns
-------
arch_names : frozenset
Empty (frozen)set if no arch codes. If not empty, contains one or more
of 'ppc', 'ppc64', 'i386... | [
"Return",
"architecture",
"types",
"from",
"library",
"libname"
] | train | https://github.com/matthew-brett/delocate/blob/ed48de15fce31c3f52f1a9f32cae1b02fc55aa60/delocate/tools.py#L441-L476 |
matthew-brett/delocate | delocate/tools.py | validate_signature | def validate_signature(filename):
""" Remove invalid signatures from a binary file
If the file signature is missing or valid then it will be ignored
Invalid signatures are replaced with an ad-hoc signature. This is the
closest you can get to removing a signature on MacOS
Parameters
---------... | python | def validate_signature(filename):
""" Remove invalid signatures from a binary file
If the file signature is missing or valid then it will be ignored
Invalid signatures are replaced with an ad-hoc signature. This is the
closest you can get to removing a signature on MacOS
Parameters
---------... | [
"def",
"validate_signature",
"(",
"filename",
")",
":",
"out",
",",
"err",
"=",
"back_tick",
"(",
"[",
"'codesign'",
",",
"'--verify'",
",",
"filename",
"]",
",",
"ret_err",
"=",
"True",
",",
"as_str",
"=",
"True",
",",
"raise_err",
"=",
"False",
")",
... | Remove invalid signatures from a binary file
If the file signature is missing or valid then it will be ignored
Invalid signatures are replaced with an ad-hoc signature. This is the
closest you can get to removing a signature on MacOS
Parameters
----------
filename : str
Filepath to a... | [
"Remove",
"invalid",
"signatures",
"from",
"a",
"binary",
"file"
] | train | https://github.com/matthew-brett/delocate/blob/ed48de15fce31c3f52f1a9f32cae1b02fc55aa60/delocate/tools.py#L513-L534 |
matthew-brett/delocate | versioneer.py | os_path_relpath | def os_path_relpath(path, start=os.path.curdir):
"""Return a relative version of a path"""
if not path:
raise ValueError("no path specified")
start_list = [x for x in os.path.abspath(start).split(os.path.sep) if x]
path_list = [x for x in os.path.abspath(path).split(os.path.sep) if x]
# W... | python | def os_path_relpath(path, start=os.path.curdir):
"""Return a relative version of a path"""
if not path:
raise ValueError("no path specified")
start_list = [x for x in os.path.abspath(start).split(os.path.sep) if x]
path_list = [x for x in os.path.abspath(path).split(os.path.sep) if x]
# W... | [
"def",
"os_path_relpath",
"(",
"path",
",",
"start",
"=",
"os",
".",
"path",
".",
"curdir",
")",
":",
"if",
"not",
"path",
":",
"raise",
"ValueError",
"(",
"\"no path specified\"",
")",
"start_list",
"=",
"[",
"x",
"for",
"x",
"in",
"os",
".",
"path",
... | Return a relative version of a path | [
"Return",
"a",
"relative",
"version",
"of",
"a",
"path"
] | train | https://github.com/matthew-brett/delocate/blob/ed48de15fce31c3f52f1a9f32cae1b02fc55aa60/versioneer.py#L596-L611 |
matthew-brett/delocate | delocate/fuse.py | fuse_trees | def fuse_trees(to_tree, from_tree, lib_exts=('.so', '.dylib', '.a')):
""" Fuse path `from_tree` into path `to_tree`
For each file in `from_tree` - check for library file extension (in
`lib_exts` - if present, check if there is a file with matching relative
path in `to_tree`, if so, use :func:`delocate.... | python | def fuse_trees(to_tree, from_tree, lib_exts=('.so', '.dylib', '.a')):
""" Fuse path `from_tree` into path `to_tree`
For each file in `from_tree` - check for library file extension (in
`lib_exts` - if present, check if there is a file with matching relative
path in `to_tree`, if so, use :func:`delocate.... | [
"def",
"fuse_trees",
"(",
"to_tree",
",",
"from_tree",
",",
"lib_exts",
"=",
"(",
"'.so'",
",",
"'.dylib'",
",",
"'.a'",
")",
")",
":",
"for",
"from_dirpath",
",",
"dirnames",
",",
"filenames",
"in",
"os",
".",
"walk",
"(",
"from_tree",
")",
":",
"to_d... | Fuse path `from_tree` into path `to_tree`
For each file in `from_tree` - check for library file extension (in
`lib_exts` - if present, check if there is a file with matching relative
path in `to_tree`, if so, use :func:`delocate.tools.lipo_fuse` to fuse the
two libraries together and write into `to_tre... | [
"Fuse",
"path",
"from_tree",
"into",
"path",
"to_tree"
] | train | https://github.com/matthew-brett/delocate/blob/ed48de15fce31c3f52f1a9f32cae1b02fc55aa60/delocate/fuse.py#L36-L77 |
matthew-brett/delocate | delocate/fuse.py | fuse_wheels | def fuse_wheels(to_wheel, from_wheel, out_wheel):
""" Fuse `from_wheel` into `to_wheel`, write to `out_wheel`
Parameters
---------
to_wheel : str
filename of wheel to fuse into
from_wheel : str
filename of wheel to fuse from
out_wheel : str
filename of new wheel from fus... | python | def fuse_wheels(to_wheel, from_wheel, out_wheel):
""" Fuse `from_wheel` into `to_wheel`, write to `out_wheel`
Parameters
---------
to_wheel : str
filename of wheel to fuse into
from_wheel : str
filename of wheel to fuse from
out_wheel : str
filename of new wheel from fus... | [
"def",
"fuse_wheels",
"(",
"to_wheel",
",",
"from_wheel",
",",
"out_wheel",
")",
":",
"to_wheel",
",",
"from_wheel",
",",
"out_wheel",
"=",
"[",
"abspath",
"(",
"w",
")",
"for",
"w",
"in",
"(",
"to_wheel",
",",
"from_wheel",
",",
"out_wheel",
")",
"]",
... | Fuse `from_wheel` into `to_wheel`, write to `out_wheel`
Parameters
---------
to_wheel : str
filename of wheel to fuse into
from_wheel : str
filename of wheel to fuse from
out_wheel : str
filename of new wheel from fusion of `to_wheel` and `from_wheel` | [
"Fuse",
"from_wheel",
"into",
"to_wheel",
"write",
"to",
"out_wheel"
] | train | https://github.com/matthew-brett/delocate/blob/ed48de15fce31c3f52f1a9f32cae1b02fc55aa60/delocate/fuse.py#L80-L99 |
matthew-brett/delocate | delocate/delocating.py | delocate_tree_libs | def delocate_tree_libs(lib_dict, lib_path, root_path):
""" Move needed libraries in `lib_dict` into `lib_path`
`lib_dict` has keys naming libraries required by the files in the
corresponding value. Call the keys, "required libs". Call the values
"requiring objects".
Copy all the required libs to... | python | def delocate_tree_libs(lib_dict, lib_path, root_path):
""" Move needed libraries in `lib_dict` into `lib_path`
`lib_dict` has keys naming libraries required by the files in the
corresponding value. Call the keys, "required libs". Call the values
"requiring objects".
Copy all the required libs to... | [
"def",
"delocate_tree_libs",
"(",
"lib_dict",
",",
"lib_path",
",",
"root_path",
")",
":",
"copied_libs",
"=",
"{",
"}",
"delocated_libs",
"=",
"set",
"(",
")",
"copied_basenames",
"=",
"set",
"(",
")",
"rp_root_path",
"=",
"realpath",
"(",
"root_path",
")",... | Move needed libraries in `lib_dict` into `lib_path`
`lib_dict` has keys naming libraries required by the files in the
corresponding value. Call the keys, "required libs". Call the values
"requiring objects".
Copy all the required libs to `lib_path`. Fix up the rpaths and install
names in the re... | [
"Move",
"needed",
"libraries",
"in",
"lib_dict",
"into",
"lib_path"
] | train | https://github.com/matthew-brett/delocate/blob/ed48de15fce31c3f52f1a9f32cae1b02fc55aa60/delocate/delocating.py#L27-L101 |
matthew-brett/delocate | delocate/delocating.py | copy_recurse | def copy_recurse(lib_path, copy_filt_func = None, copied_libs = None):
""" Analyze `lib_path` for library dependencies and copy libraries
`lib_path` is a directory containing libraries. The libraries might
themselves have dependencies. This function analyzes the dependencies and
copies library depend... | python | def copy_recurse(lib_path, copy_filt_func = None, copied_libs = None):
""" Analyze `lib_path` for library dependencies and copy libraries
`lib_path` is a directory containing libraries. The libraries might
themselves have dependencies. This function analyzes the dependencies and
copies library depend... | [
"def",
"copy_recurse",
"(",
"lib_path",
",",
"copy_filt_func",
"=",
"None",
",",
"copied_libs",
"=",
"None",
")",
":",
"if",
"copied_libs",
"is",
"None",
":",
"copied_libs",
"=",
"{",
"}",
"else",
":",
"copied_libs",
"=",
"dict",
"(",
"copied_libs",
")",
... | Analyze `lib_path` for library dependencies and copy libraries
`lib_path` is a directory containing libraries. The libraries might
themselves have dependencies. This function analyzes the dependencies and
copies library dependencies that match the filter `copy_filt_func`. It also
adjusts the dependin... | [
"Analyze",
"lib_path",
"for",
"library",
"dependencies",
"and",
"copy",
"libraries"
] | train | https://github.com/matthew-brett/delocate/blob/ed48de15fce31c3f52f1a9f32cae1b02fc55aa60/delocate/delocating.py#L104-L147 |
matthew-brett/delocate | delocate/delocating.py | _copy_required | def _copy_required(lib_path, copy_filt_func, copied_libs):
""" Copy libraries required for files in `lib_path` to `lib_path`
Augment `copied_libs` dictionary with any newly copied libraries, modifying
`copied_libs` in-place - see Notes.
This is one pass of ``copy_recurse``
Parameters
--------... | python | def _copy_required(lib_path, copy_filt_func, copied_libs):
""" Copy libraries required for files in `lib_path` to `lib_path`
Augment `copied_libs` dictionary with any newly copied libraries, modifying
`copied_libs` in-place - see Notes.
This is one pass of ``copy_recurse``
Parameters
--------... | [
"def",
"_copy_required",
"(",
"lib_path",
",",
"copy_filt_func",
",",
"copied_libs",
")",
":",
"# Paths will be prepended with `lib_path`",
"lib_dict",
"=",
"tree_libs",
"(",
"lib_path",
")",
"# Map library paths after copy ('copied') to path before copy ('orig')",
"rp_lp",
"="... | Copy libraries required for files in `lib_path` to `lib_path`
Augment `copied_libs` dictionary with any newly copied libraries, modifying
`copied_libs` in-place - see Notes.
This is one pass of ``copy_recurse``
Parameters
----------
lib_path : str
Directory containing libraries
co... | [
"Copy",
"libraries",
"required",
"for",
"files",
"in",
"lib_path",
"to",
"lib_path"
] | train | https://github.com/matthew-brett/delocate/blob/ed48de15fce31c3f52f1a9f32cae1b02fc55aa60/delocate/delocating.py#L150-L233 |
matthew-brett/delocate | delocate/delocating.py | delocate_path | def delocate_path(tree_path, lib_path,
lib_filt_func = None,
copy_filt_func = filter_system_libs):
""" Copy required libraries for files in `tree_path` into `lib_path`
Parameters
----------
tree_path : str
Root path of tree to search for required libraries
... | python | def delocate_path(tree_path, lib_path,
lib_filt_func = None,
copy_filt_func = filter_system_libs):
""" Copy required libraries for files in `tree_path` into `lib_path`
Parameters
----------
tree_path : str
Root path of tree to search for required libraries
... | [
"def",
"delocate_path",
"(",
"tree_path",
",",
"lib_path",
",",
"lib_filt_func",
"=",
"None",
",",
"copy_filt_func",
"=",
"filter_system_libs",
")",
":",
"if",
"lib_filt_func",
"==",
"\"dylibs-only\"",
":",
"lib_filt_func",
"=",
"_dylibs_only",
"if",
"not",
"exist... | Copy required libraries for files in `tree_path` into `lib_path`
Parameters
----------
tree_path : str
Root path of tree to search for required libraries
lib_path : str
Directory into which we copy required libraries
lib_filt_func : None or str or callable, optional
If None,... | [
"Copy",
"required",
"libraries",
"for",
"files",
"in",
"tree_path",
"into",
"lib_path"
] | train | https://github.com/matthew-brett/delocate/blob/ed48de15fce31c3f52f1a9f32cae1b02fc55aa60/delocate/delocating.py#L246-L289 |
matthew-brett/delocate | delocate/delocating.py | _merge_lib_dict | def _merge_lib_dict(d1, d2):
""" Merges lib_dict `d2` into lib_dict `d1`
"""
for required, requirings in d2.items():
if required in d1:
d1[required].update(requirings)
else:
d1[required] = requirings
return None | python | def _merge_lib_dict(d1, d2):
""" Merges lib_dict `d2` into lib_dict `d1`
"""
for required, requirings in d2.items():
if required in d1:
d1[required].update(requirings)
else:
d1[required] = requirings
return None | [
"def",
"_merge_lib_dict",
"(",
"d1",
",",
"d2",
")",
":",
"for",
"required",
",",
"requirings",
"in",
"d2",
".",
"items",
"(",
")",
":",
"if",
"required",
"in",
"d1",
":",
"d1",
"[",
"required",
"]",
".",
"update",
"(",
"requirings",
")",
"else",
"... | Merges lib_dict `d2` into lib_dict `d1` | [
"Merges",
"lib_dict",
"d2",
"into",
"lib_dict",
"d1"
] | train | https://github.com/matthew-brett/delocate/blob/ed48de15fce31c3f52f1a9f32cae1b02fc55aa60/delocate/delocating.py#L292-L300 |
matthew-brett/delocate | delocate/delocating.py | delocate_wheel | def delocate_wheel(in_wheel,
out_wheel = None,
lib_sdir = '.dylibs',
lib_filt_func = None,
copy_filt_func = filter_system_libs,
require_archs = None,
check_verbose = False,
):
""" Upda... | python | def delocate_wheel(in_wheel,
out_wheel = None,
lib_sdir = '.dylibs',
lib_filt_func = None,
copy_filt_func = filter_system_libs,
require_archs = None,
check_verbose = False,
):
""" Upda... | [
"def",
"delocate_wheel",
"(",
"in_wheel",
",",
"out_wheel",
"=",
"None",
",",
"lib_sdir",
"=",
"'.dylibs'",
",",
"lib_filt_func",
"=",
"None",
",",
"copy_filt_func",
"=",
"filter_system_libs",
",",
"require_archs",
"=",
"None",
",",
"check_verbose",
"=",
"False"... | Update wheel by copying required libraries to `lib_sdir` in wheel
Create `lib_sdir` in wheel tree only if we are copying one or more
libraries.
If `out_wheel` is None (the default), overwrite the wheel `in_wheel`
in-place.
Parameters
----------
in_wheel : str
Filename of wheel to ... | [
"Update",
"wheel",
"by",
"copying",
"required",
"libraries",
"to",
"lib_sdir",
"in",
"wheel"
] | train | https://github.com/matthew-brett/delocate/blob/ed48de15fce31c3f52f1a9f32cae1b02fc55aa60/delocate/delocating.py#L303-L407 |
matthew-brett/delocate | delocate/delocating.py | patch_wheel | def patch_wheel(in_wheel, patch_fname, out_wheel=None):
""" Apply ``-p1`` style patch in `patch_fname` to contents of `in_wheel`
If `out_wheel` is None (the default), overwrite the wheel `in_wheel`
in-place.
Parameters
----------
in_wheel : str
Filename of wheel to process
patch_fn... | python | def patch_wheel(in_wheel, patch_fname, out_wheel=None):
""" Apply ``-p1`` style patch in `patch_fname` to contents of `in_wheel`
If `out_wheel` is None (the default), overwrite the wheel `in_wheel`
in-place.
Parameters
----------
in_wheel : str
Filename of wheel to process
patch_fn... | [
"def",
"patch_wheel",
"(",
"in_wheel",
",",
"patch_fname",
",",
"out_wheel",
"=",
"None",
")",
":",
"in_wheel",
"=",
"abspath",
"(",
"in_wheel",
")",
"patch_fname",
"=",
"abspath",
"(",
"patch_fname",
")",
"if",
"out_wheel",
"is",
"None",
":",
"out_wheel",
... | Apply ``-p1`` style patch in `patch_fname` to contents of `in_wheel`
If `out_wheel` is None (the default), overwrite the wheel `in_wheel`
in-place.
Parameters
----------
in_wheel : str
Filename of wheel to process
patch_fname : str
Filename of patch file. Will be applied with ... | [
"Apply",
"-",
"p1",
"style",
"patch",
"in",
"patch_fname",
"to",
"contents",
"of",
"in_wheel"
] | train | https://github.com/matthew-brett/delocate/blob/ed48de15fce31c3f52f1a9f32cae1b02fc55aa60/delocate/delocating.py#L410-L443 |
matthew-brett/delocate | delocate/delocating.py | check_archs | def check_archs(copied_libs, require_archs=(), stop_fast=False):
""" Check compatibility of archs in `copied_libs` dict
Parameters
----------
copied_libs : dict
dict containing the (key, value) pairs of (``copied_lib_path``,
``dependings_dict``), where ``copied_lib_path`` is a library r... | python | def check_archs(copied_libs, require_archs=(), stop_fast=False):
""" Check compatibility of archs in `copied_libs` dict
Parameters
----------
copied_libs : dict
dict containing the (key, value) pairs of (``copied_lib_path``,
``dependings_dict``), where ``copied_lib_path`` is a library r... | [
"def",
"check_archs",
"(",
"copied_libs",
",",
"require_archs",
"=",
"(",
")",
",",
"stop_fast",
"=",
"False",
")",
":",
"if",
"isinstance",
"(",
"require_archs",
",",
"string_types",
")",
":",
"require_archs",
"=",
"(",
"[",
"'i386'",
",",
"'x86_64'",
"]"... | Check compatibility of archs in `copied_libs` dict
Parameters
----------
copied_libs : dict
dict containing the (key, value) pairs of (``copied_lib_path``,
``dependings_dict``), where ``copied_lib_path`` is a library real path
that has been copied during delocation, and ``dependings... | [
"Check",
"compatibility",
"of",
"archs",
"in",
"copied_libs",
"dict"
] | train | https://github.com/matthew-brett/delocate/blob/ed48de15fce31c3f52f1a9f32cae1b02fc55aa60/delocate/delocating.py#L446-L505 |
matthew-brett/delocate | delocate/delocating.py | bads_report | def bads_report(bads, path_prefix=None):
""" Return a nice report of bad architectures in `bads`
Parameters
----------
bads : set
set of length 2 or 3 tuples. A length 2 tuple is of form
``(depending_lib, missing_archs)`` meaning that an arch in
`require_archs` was missing from ... | python | def bads_report(bads, path_prefix=None):
""" Return a nice report of bad architectures in `bads`
Parameters
----------
bads : set
set of length 2 or 3 tuples. A length 2 tuple is of form
``(depending_lib, missing_archs)`` meaning that an arch in
`require_archs` was missing from ... | [
"def",
"bads_report",
"(",
"bads",
",",
"path_prefix",
"=",
"None",
")",
":",
"path_processor",
"=",
"(",
"(",
"lambda",
"x",
":",
"x",
")",
"if",
"path_prefix",
"is",
"None",
"else",
"get_rp_stripper",
"(",
"path_prefix",
")",
")",
"reports",
"=",
"[",
... | Return a nice report of bad architectures in `bads`
Parameters
----------
bads : set
set of length 2 or 3 tuples. A length 2 tuple is of form
``(depending_lib, missing_archs)`` meaning that an arch in
`require_archs` was missing from ``depending_lib``. A length 3 tuple
is o... | [
"Return",
"a",
"nice",
"report",
"of",
"bad",
"architectures",
"in",
"bads"
] | train | https://github.com/matthew-brett/delocate/blob/ed48de15fce31c3f52f1a9f32cae1b02fc55aa60/delocate/delocating.py#L508-L552 |
matthew-brett/delocate | delocate/libsana.py | tree_libs | def tree_libs(start_path, filt_func=None):
""" Return analysis of library dependencies within `start_path`
Parameters
----------
start_path : str
root path of tree to search for libraries depending on other libraries.
filt_func : None or callable, optional
If None, inspect all files... | python | def tree_libs(start_path, filt_func=None):
""" Return analysis of library dependencies within `start_path`
Parameters
----------
start_path : str
root path of tree to search for libraries depending on other libraries.
filt_func : None or callable, optional
If None, inspect all files... | [
"def",
"tree_libs",
"(",
"start_path",
",",
"filt_func",
"=",
"None",
")",
":",
"lib_dict",
"=",
"{",
"}",
"for",
"dirpath",
",",
"dirnames",
",",
"basenames",
"in",
"os",
".",
"walk",
"(",
"start_path",
")",
":",
"for",
"base",
"in",
"basenames",
":",... | Return analysis of library dependencies within `start_path`
Parameters
----------
start_path : str
root path of tree to search for libraries depending on other libraries.
filt_func : None or callable, optional
If None, inspect all files for library dependencies. If callable,
acc... | [
"Return",
"analysis",
"of",
"library",
"dependencies",
"within",
"start_path"
] | train | https://github.com/matthew-brett/delocate/blob/ed48de15fce31c3f52f1a9f32cae1b02fc55aa60/delocate/libsana.py#L14-L65 |
matthew-brett/delocate | delocate/libsana.py | resolve_rpath | def resolve_rpath(lib_path, rpaths):
""" Return `lib_path` with its `@rpath` resolved
If the `lib_path` doesn't have `@rpath` then it's returned as is.
If `lib_path` has `@rpath` then returns the first `rpaths`/`lib_path`
combination found. If the library can't be found in `rpaths` then a
detaile... | python | def resolve_rpath(lib_path, rpaths):
""" Return `lib_path` with its `@rpath` resolved
If the `lib_path` doesn't have `@rpath` then it's returned as is.
If `lib_path` has `@rpath` then returns the first `rpaths`/`lib_path`
combination found. If the library can't be found in `rpaths` then a
detaile... | [
"def",
"resolve_rpath",
"(",
"lib_path",
",",
"rpaths",
")",
":",
"if",
"not",
"lib_path",
".",
"startswith",
"(",
"'@rpath/'",
")",
":",
"return",
"lib_path",
"lib_rpath",
"=",
"lib_path",
".",
"split",
"(",
"'/'",
",",
"1",
")",
"[",
"1",
"]",
"for",... | Return `lib_path` with its `@rpath` resolved
If the `lib_path` doesn't have `@rpath` then it's returned as is.
If `lib_path` has `@rpath` then returns the first `rpaths`/`lib_path`
combination found. If the library can't be found in `rpaths` then a
detailed warning is printed and `lib_path` is return... | [
"Return",
"lib_path",
"with",
"its",
"@rpath",
"resolved"
] | train | https://github.com/matthew-brett/delocate/blob/ed48de15fce31c3f52f1a9f32cae1b02fc55aa60/delocate/libsana.py#L68-L104 |
matthew-brett/delocate | delocate/libsana.py | get_prefix_stripper | def get_prefix_stripper(strip_prefix):
""" Return function to strip `strip_prefix` prefix from string if present
Parameters
----------
prefix : str
Prefix to strip from the beginning of string if present
Returns
-------
stripper : func
function such that ``stripper(a_string... | python | def get_prefix_stripper(strip_prefix):
""" Return function to strip `strip_prefix` prefix from string if present
Parameters
----------
prefix : str
Prefix to strip from the beginning of string if present
Returns
-------
stripper : func
function such that ``stripper(a_string... | [
"def",
"get_prefix_stripper",
"(",
"strip_prefix",
")",
":",
"n",
"=",
"len",
"(",
"strip_prefix",
")",
"def",
"stripper",
"(",
"path",
")",
":",
"return",
"path",
"if",
"not",
"path",
".",
"startswith",
"(",
"strip_prefix",
")",
"else",
"path",
"[",
"n"... | Return function to strip `strip_prefix` prefix from string if present
Parameters
----------
prefix : str
Prefix to strip from the beginning of string if present
Returns
-------
stripper : func
function such that ``stripper(a_string)`` will strip `prefix` from
``a_string... | [
"Return",
"function",
"to",
"strip",
"strip_prefix",
"prefix",
"from",
"string",
"if",
"present"
] | train | https://github.com/matthew-brett/delocate/blob/ed48de15fce31c3f52f1a9f32cae1b02fc55aa60/delocate/libsana.py#L107-L124 |
matthew-brett/delocate | delocate/libsana.py | stripped_lib_dict | def stripped_lib_dict(lib_dict, strip_prefix):
""" Return `lib_dict` with `strip_prefix` removed from start of paths
Use to give form of `lib_dict` that appears relative to some base path
given by `strip_prefix`. Particularly useful for analyzing wheels where we
unpack to a temporary path before analy... | python | def stripped_lib_dict(lib_dict, strip_prefix):
""" Return `lib_dict` with `strip_prefix` removed from start of paths
Use to give form of `lib_dict` that appears relative to some base path
given by `strip_prefix`. Particularly useful for analyzing wheels where we
unpack to a temporary path before analy... | [
"def",
"stripped_lib_dict",
"(",
"lib_dict",
",",
"strip_prefix",
")",
":",
"relative_dict",
"=",
"{",
"}",
"stripper",
"=",
"get_prefix_stripper",
"(",
"strip_prefix",
")",
"for",
"lib_path",
",",
"dependings_dict",
"in",
"lib_dict",
".",
"items",
"(",
")",
"... | Return `lib_dict` with `strip_prefix` removed from start of paths
Use to give form of `lib_dict` that appears relative to some base path
given by `strip_prefix`. Particularly useful for analyzing wheels where we
unpack to a temporary path before analyzing.
Parameters
----------
lib_dict : dic... | [
"Return",
"lib_dict",
"with",
"strip_prefix",
"removed",
"from",
"start",
"of",
"paths"
] | train | https://github.com/matthew-brett/delocate/blob/ed48de15fce31c3f52f1a9f32cae1b02fc55aa60/delocate/libsana.py#L145-L175 |
matthew-brett/delocate | delocate/libsana.py | wheel_libs | def wheel_libs(wheel_fname, filt_func = None):
""" Return analysis of library dependencies with a Python wheel
Use this routine for a dump of the dependency tree.
Parameters
----------
wheel_fname : str
Filename of wheel
filt_func : None or callable, optional
If None, inspect a... | python | def wheel_libs(wheel_fname, filt_func = None):
""" Return analysis of library dependencies with a Python wheel
Use this routine for a dump of the dependency tree.
Parameters
----------
wheel_fname : str
Filename of wheel
filt_func : None or callable, optional
If None, inspect a... | [
"def",
"wheel_libs",
"(",
"wheel_fname",
",",
"filt_func",
"=",
"None",
")",
":",
"with",
"TemporaryDirectory",
"(",
")",
"as",
"tmpdir",
":",
"zip2dir",
"(",
"wheel_fname",
",",
"tmpdir",
")",
"lib_dict",
"=",
"tree_libs",
"(",
"tmpdir",
",",
"filt_func",
... | Return analysis of library dependencies with a Python wheel
Use this routine for a dump of the dependency tree.
Parameters
----------
wheel_fname : str
Filename of wheel
filt_func : None or callable, optional
If None, inspect all files for library dependencies. If callable,
... | [
"Return",
"analysis",
"of",
"library",
"dependencies",
"with",
"a",
"Python",
"wheel"
] | train | https://github.com/matthew-brett/delocate/blob/ed48de15fce31c3f52f1a9f32cae1b02fc55aa60/delocate/libsana.py#L178-L205 |
matthew-brett/delocate | delocate/wheeltools.py | _open_for_csv | def _open_for_csv(name, mode):
""" Deal with Python 2/3 open API differences """
if sys.version_info[0] < 3:
return open_rw(name, mode + 'b')
return open_rw(name, mode, newline='', encoding='utf-8') | python | def _open_for_csv(name, mode):
""" Deal with Python 2/3 open API differences """
if sys.version_info[0] < 3:
return open_rw(name, mode + 'b')
return open_rw(name, mode, newline='', encoding='utf-8') | [
"def",
"_open_for_csv",
"(",
"name",
",",
"mode",
")",
":",
"if",
"sys",
".",
"version_info",
"[",
"0",
"]",
"<",
"3",
":",
"return",
"open_rw",
"(",
"name",
",",
"mode",
"+",
"'b'",
")",
"return",
"open_rw",
"(",
"name",
",",
"mode",
",",
"newline... | Deal with Python 2/3 open API differences | [
"Deal",
"with",
"Python",
"2",
"/",
"3",
"open",
"API",
"differences"
] | train | https://github.com/matthew-brett/delocate/blob/ed48de15fce31c3f52f1a9f32cae1b02fc55aa60/delocate/wheeltools.py#L28-L32 |
matthew-brett/delocate | delocate/wheeltools.py | rewrite_record | def rewrite_record(bdist_dir):
""" Rewrite RECORD file with hashes for all files in `wheel_sdir`
Copied from :method:`wheel.bdist_wheel.bdist_wheel.write_record`
Will also unsign wheel
Parameters
----------
bdist_dir : str
Path of unpacked wheel file
"""
info_dirs = glob.glob(... | python | def rewrite_record(bdist_dir):
""" Rewrite RECORD file with hashes for all files in `wheel_sdir`
Copied from :method:`wheel.bdist_wheel.bdist_wheel.write_record`
Will also unsign wheel
Parameters
----------
bdist_dir : str
Path of unpacked wheel file
"""
info_dirs = glob.glob(... | [
"def",
"rewrite_record",
"(",
"bdist_dir",
")",
":",
"info_dirs",
"=",
"glob",
".",
"glob",
"(",
"pjoin",
"(",
"bdist_dir",
",",
"'*.dist-info'",
")",
")",
"if",
"len",
"(",
"info_dirs",
")",
"!=",
"1",
":",
"raise",
"WheelToolsError",
"(",
"\"Should be ex... | Rewrite RECORD file with hashes for all files in `wheel_sdir`
Copied from :method:`wheel.bdist_wheel.bdist_wheel.write_record`
Will also unsign wheel
Parameters
----------
bdist_dir : str
Path of unpacked wheel file | [
"Rewrite",
"RECORD",
"file",
"with",
"hashes",
"for",
"all",
"files",
"in",
"wheel_sdir"
] | train | https://github.com/matthew-brett/delocate/blob/ed48de15fce31c3f52f1a9f32cae1b02fc55aa60/delocate/wheeltools.py#L35-L81 |
matthew-brett/delocate | delocate/wheeltools.py | add_platforms | def add_platforms(in_wheel, platforms, out_path=None, clobber=False):
""" Add platform tags `platforms` to `in_wheel` filename and WHEEL tags
Add any platform tags in `platforms` that are missing from `in_wheel`
filename.
Add any platform tags in `platforms` that are missing from `in_wheel`
``WHEE... | python | def add_platforms(in_wheel, platforms, out_path=None, clobber=False):
""" Add platform tags `platforms` to `in_wheel` filename and WHEEL tags
Add any platform tags in `platforms` that are missing from `in_wheel`
filename.
Add any platform tags in `platforms` that are missing from `in_wheel`
``WHEE... | [
"def",
"add_platforms",
"(",
"in_wheel",
",",
"platforms",
",",
"out_path",
"=",
"None",
",",
"clobber",
"=",
"False",
")",
":",
"in_wheel",
"=",
"abspath",
"(",
"in_wheel",
")",
"out_path",
"=",
"dirname",
"(",
"in_wheel",
")",
"if",
"out_path",
"is",
"... | Add platform tags `platforms` to `in_wheel` filename and WHEEL tags
Add any platform tags in `platforms` that are missing from `in_wheel`
filename.
Add any platform tags in `platforms` that are missing from `in_wheel`
``WHEEL`` file.
Parameters
----------
in_wheel : str
Filename o... | [
"Add",
"platform",
"tags",
"platforms",
"to",
"in_wheel",
"filename",
"and",
"WHEEL",
"tags"
] | train | https://github.com/matthew-brett/delocate/blob/ed48de15fce31c3f52f1a9f32cae1b02fc55aa60/delocate/wheeltools.py#L162-L222 |
wiheto/teneto | teneto/networkmeasures/temporal_betweenness_centrality.py | temporal_betweenness_centrality | def temporal_betweenness_centrality(tnet=None, paths=None, calc='time'):
'''
Returns temporal betweenness centrality per node.
Parameters
-----------
Input should be *either* tnet or paths.
data : array or dict
Temporal network input (graphlet or contact). nettype: 'bu', 'bd'.
c... | python | def temporal_betweenness_centrality(tnet=None, paths=None, calc='time'):
'''
Returns temporal betweenness centrality per node.
Parameters
-----------
Input should be *either* tnet or paths.
data : array or dict
Temporal network input (graphlet or contact). nettype: 'bu', 'bd'.
c... | [
"def",
"temporal_betweenness_centrality",
"(",
"tnet",
"=",
"None",
",",
"paths",
"=",
"None",
",",
"calc",
"=",
"'time'",
")",
":",
"if",
"tnet",
"is",
"not",
"None",
"and",
"paths",
"is",
"not",
"None",
":",
"raise",
"ValueError",
"(",
"'Only network or ... | Returns temporal betweenness centrality per node.
Parameters
-----------
Input should be *either* tnet or paths.
data : array or dict
Temporal network input (graphlet or contact). nettype: 'bu', 'bd'.
calc : str
either 'global' or 'time'
paths : pandas dataframe
O... | [
"Returns",
"temporal",
"betweenness",
"centrality",
"per",
"node",
"."
] | train | https://github.com/wiheto/teneto/blob/80d7a83a9adc1714589b020627c45bd5b66248ab/teneto/networkmeasures/temporal_betweenness_centrality.py#L9-L72 |
wiheto/teneto | teneto/networkmeasures/volatility.py | volatility | def volatility(tnet, distance_func_name='default', calc='global', communities=None, event_displacement=None):
r"""
Volatility of temporal networks.
Volatility is the average distance between consecutive time points of graphlets (difference is caclualted either globally or per edge).
Parameters
---... | python | def volatility(tnet, distance_func_name='default', calc='global', communities=None, event_displacement=None):
r"""
Volatility of temporal networks.
Volatility is the average distance between consecutive time points of graphlets (difference is caclualted either globally or per edge).
Parameters
---... | [
"def",
"volatility",
"(",
"tnet",
",",
"distance_func_name",
"=",
"'default'",
",",
"calc",
"=",
"'global'",
",",
"communities",
"=",
"None",
",",
"event_displacement",
"=",
"None",
")",
":",
"# Get input (C or G)",
"tnet",
",",
"netinfo",
"=",
"process_input",
... | r"""
Volatility of temporal networks.
Volatility is the average distance between consecutive time points of graphlets (difference is caclualted either globally or per edge).
Parameters
----------
tnet : array or dict
temporal network input (graphlet or contact). Nettype: 'bu','bd','wu','w... | [
"r",
"Volatility",
"of",
"temporal",
"networks",
"."
] | train | https://github.com/wiheto/teneto/blob/80d7a83a9adc1714589b020627c45bd5b66248ab/teneto/networkmeasures/volatility.py#L5-L188 |
wiheto/teneto | teneto/temporalcommunity/allegiance.py | allegiance | def allegiance(community):
"""
Computes the allegiance matrix with values representing the probability that
nodes i and j were assigned to the same community by time-varying clustering methods.
parameters
----------
community : array
array of community assignment of size node,time
... | python | def allegiance(community):
"""
Computes the allegiance matrix with values representing the probability that
nodes i and j were assigned to the same community by time-varying clustering methods.
parameters
----------
community : array
array of community assignment of size node,time
... | [
"def",
"allegiance",
"(",
"community",
")",
":",
"N",
"=",
"community",
".",
"shape",
"[",
"0",
"]",
"C",
"=",
"community",
".",
"shape",
"[",
"1",
"]",
"T",
"=",
"P",
"=",
"np",
".",
"zeros",
"(",
"[",
"N",
",",
"N",
"]",
")",
"for",
"t",
... | Computes the allegiance matrix with values representing the probability that
nodes i and j were assigned to the same community by time-varying clustering methods.
parameters
----------
community : array
array of community assignment of size node,time
returns
-------
P : array
... | [
"Computes",
"the",
"allegiance",
"matrix",
"with",
"values",
"representing",
"the",
"probability",
"that",
"nodes",
"i",
"and",
"j",
"were",
"assigned",
"to",
"the",
"same",
"community",
"by",
"time",
"-",
"varying",
"clustering",
"methods",
"."
] | train | https://github.com/wiheto/teneto/blob/80d7a83a9adc1714589b020627c45bd5b66248ab/teneto/temporalcommunity/allegiance.py#L3-L39 |
wiheto/teneto | teneto/generatenetwork/rand_poisson.py | rand_poisson | def rand_poisson(nnodes, ncontacts, lam=1, nettype='bu', netinfo=None, netrep='graphlet'):
"""
Generate a random network where intervals between contacts are distributed by a poisson distribution
Parameters
----------
nnodes : int
Number of nodes in networks
ncontacts : int or list
... | python | def rand_poisson(nnodes, ncontacts, lam=1, nettype='bu', netinfo=None, netrep='graphlet'):
"""
Generate a random network where intervals between contacts are distributed by a poisson distribution
Parameters
----------
nnodes : int
Number of nodes in networks
ncontacts : int or list
... | [
"def",
"rand_poisson",
"(",
"nnodes",
",",
"ncontacts",
",",
"lam",
"=",
"1",
",",
"nettype",
"=",
"'bu'",
",",
"netinfo",
"=",
"None",
",",
"netrep",
"=",
"'graphlet'",
")",
":",
"if",
"isinstance",
"(",
"ncontacts",
",",
"list",
")",
":",
"if",
"le... | Generate a random network where intervals between contacts are distributed by a poisson distribution
Parameters
----------
nnodes : int
Number of nodes in networks
ncontacts : int or list
Number of expected contacts (i.e. edges). If list, number of contacts for each node.
Any ... | [
"Generate",
"a",
"random",
"network",
"where",
"intervals",
"between",
"contacts",
"are",
"distributed",
"by",
"a",
"poisson",
"distribution"
] | train | https://github.com/wiheto/teneto/blob/80d7a83a9adc1714589b020627c45bd5b66248ab/teneto/generatenetwork/rand_poisson.py#L9-L92 |
wiheto/teneto | teneto/networkmeasures/temporal_participation_coeff.py | temporal_participation_coeff | def temporal_participation_coeff(tnet, communities=None, decay=None, removeneg=False):
r'''
Temporal participation coefficient is a measure of diversity of connections across communities for individual nodes.
Parameters
----------
tnet : array, dict
graphlet or contact sequence input. Only ... | python | def temporal_participation_coeff(tnet, communities=None, decay=None, removeneg=False):
r'''
Temporal participation coefficient is a measure of diversity of connections across communities for individual nodes.
Parameters
----------
tnet : array, dict
graphlet or contact sequence input. Only ... | [
"def",
"temporal_participation_coeff",
"(",
"tnet",
",",
"communities",
"=",
"None",
",",
"decay",
"=",
"None",
",",
"removeneg",
"=",
"False",
")",
":",
"if",
"communities",
"is",
"None",
":",
"if",
"isinstance",
"(",
"tnet",
",",
"dict",
")",
":",
"if"... | r'''
Temporal participation coefficient is a measure of diversity of connections across communities for individual nodes.
Parameters
----------
tnet : array, dict
graphlet or contact sequence input. Only positive matrices considered.
communities : array
community vector. Either 1D (... | [
"r",
"Temporal",
"participation",
"coefficient",
"is",
"a",
"measure",
"of",
"diversity",
"of",
"connections",
"across",
"communities",
"for",
"individual",
"nodes",
"."
] | train | https://github.com/wiheto/teneto/blob/80d7a83a9adc1714589b020627c45bd5b66248ab/teneto/networkmeasures/temporal_participation_coeff.py#L7-L162 |
wiheto/teneto | teneto/plot/graphlet_stack_plot.py | graphlet_stack_plot | def graphlet_stack_plot(netin, ax, q=10, cmap='Reds', gridcolor='k', borderwidth=2, bordercolor=None, Fs=1, timeunit='', t0=1, sharpen='yes', vminmax='minmax'):
r'''
Returns matplotlib axis handle for graphlet_stack_plot. This is a row of transformed connectivity matrices to look like a 3D stack.
Parameter... | python | def graphlet_stack_plot(netin, ax, q=10, cmap='Reds', gridcolor='k', borderwidth=2, bordercolor=None, Fs=1, timeunit='', t0=1, sharpen='yes', vminmax='minmax'):
r'''
Returns matplotlib axis handle for graphlet_stack_plot. This is a row of transformed connectivity matrices to look like a 3D stack.
Parameter... | [
"def",
"graphlet_stack_plot",
"(",
"netin",
",",
"ax",
",",
"q",
"=",
"10",
",",
"cmap",
"=",
"'Reds'",
",",
"gridcolor",
"=",
"'k'",
",",
"borderwidth",
"=",
"2",
",",
"bordercolor",
"=",
"None",
",",
"Fs",
"=",
"1",
",",
"timeunit",
"=",
"''",
",... | r'''
Returns matplotlib axis handle for graphlet_stack_plot. This is a row of transformed connectivity matrices to look like a 3D stack.
Parameters
----------
netin : array, dict
network input (graphlet or contact)
ax : matplotlib ax handles.
q : int
Quality. Increaseing this w... | [
"r",
"Returns",
"matplotlib",
"axis",
"handle",
"for",
"graphlet_stack_plot",
".",
"This",
"is",
"a",
"row",
"of",
"transformed",
"connectivity",
"matrices",
"to",
"look",
"like",
"a",
"3D",
"stack",
"."
] | train | https://github.com/wiheto/teneto/blob/80d7a83a9adc1714589b020627c45bd5b66248ab/teneto/plot/graphlet_stack_plot.py#L9-L271 |
wiheto/teneto | teneto/communitydetection/tctc.py | partition_inference | def partition_inference(tctc_mat, comp, tau, sigma, kappa):
r"""
Takes tctc trajectory matrix and returns dataframe where all multi-label communities are listed
Can take a little bit of time with large datasets and optimizaiton could remove some for loops.
"""
communityinfo = {}
communityinfo[... | python | def partition_inference(tctc_mat, comp, tau, sigma, kappa):
r"""
Takes tctc trajectory matrix and returns dataframe where all multi-label communities are listed
Can take a little bit of time with large datasets and optimizaiton could remove some for loops.
"""
communityinfo = {}
communityinfo[... | [
"def",
"partition_inference",
"(",
"tctc_mat",
",",
"comp",
",",
"tau",
",",
"sigma",
",",
"kappa",
")",
":",
"communityinfo",
"=",
"{",
"}",
"communityinfo",
"[",
"'community'",
"]",
"=",
"[",
"]",
"communityinfo",
"[",
"'start'",
"]",
"=",
"np",
".",
... | r"""
Takes tctc trajectory matrix and returns dataframe where all multi-label communities are listed
Can take a little bit of time with large datasets and optimizaiton could remove some for loops. | [
"r",
"Takes",
"tctc",
"trajectory",
"matrix",
"and",
"returns",
"dataframe",
"where",
"all",
"multi",
"-",
"label",
"communities",
"are",
"listed"
] | train | https://github.com/wiheto/teneto/blob/80d7a83a9adc1714589b020627c45bd5b66248ab/teneto/communitydetection/tctc.py#L8-L114 |
wiheto/teneto | teneto/communitydetection/tctc.py | tctc | def tctc(data, tau, epsilon, sigma, kappa=0, largedataset=False, rule='flock', noise=None, raw_signal='amplitude', output='array', tempdir=None, njobs=1, largestonly=False):
r"""
Runs TCTC community detection
Parameters
----------
data : array
Multiariate series with dimensions: "time, node... | python | def tctc(data, tau, epsilon, sigma, kappa=0, largedataset=False, rule='flock', noise=None, raw_signal='amplitude', output='array', tempdir=None, njobs=1, largestonly=False):
r"""
Runs TCTC community detection
Parameters
----------
data : array
Multiariate series with dimensions: "time, node... | [
"def",
"tctc",
"(",
"data",
",",
"tau",
",",
"epsilon",
",",
"sigma",
",",
"kappa",
"=",
"0",
",",
"largedataset",
"=",
"False",
",",
"rule",
"=",
"'flock'",
",",
"noise",
"=",
"None",
",",
"raw_signal",
"=",
"'amplitude'",
",",
"output",
"=",
"'arra... | r"""
Runs TCTC community detection
Parameters
----------
data : array
Multiariate series with dimensions: "time, node" that belong to a network.
tau : int
tau specifies the minimum number of time-points of each temporal community must last.
epsilon : float
epsilon specif... | [
"r",
"Runs",
"TCTC",
"community",
"detection"
] | train | https://github.com/wiheto/teneto/blob/80d7a83a9adc1714589b020627c45bd5b66248ab/teneto/communitydetection/tctc.py#L117-L271 |
wiheto/teneto | teneto/networkmeasures/temporal_efficiency.py | temporal_efficiency | def temporal_efficiency(tnet=None, paths=None, calc='global'):
r"""
Returns temporal efficiency estimate. BU networks only.
Parameters
----------
Input should be *either* tnet or paths.
data : array or dict
Temporal network input (graphlet or contact). nettype: 'bu', 'bd'.
paths ... | python | def temporal_efficiency(tnet=None, paths=None, calc='global'):
r"""
Returns temporal efficiency estimate. BU networks only.
Parameters
----------
Input should be *either* tnet or paths.
data : array or dict
Temporal network input (graphlet or contact). nettype: 'bu', 'bd'.
paths ... | [
"def",
"temporal_efficiency",
"(",
"tnet",
"=",
"None",
",",
"paths",
"=",
"None",
",",
"calc",
"=",
"'global'",
")",
":",
"if",
"tnet",
"is",
"not",
"None",
"and",
"paths",
"is",
"not",
"None",
":",
"raise",
"ValueError",
"(",
"'Only network or path input... | r"""
Returns temporal efficiency estimate. BU networks only.
Parameters
----------
Input should be *either* tnet or paths.
data : array or dict
Temporal network input (graphlet or contact). nettype: 'bu', 'bd'.
paths : pandas dataframe
Output of TenetoBIDS.networkmeasure.sho... | [
"r",
"Returns",
"temporal",
"efficiency",
"estimate",
".",
"BU",
"networks",
"only",
"."
] | train | https://github.com/wiheto/teneto/blob/80d7a83a9adc1714589b020627c45bd5b66248ab/teneto/networkmeasures/temporal_efficiency.py#L9-L60 |
wiheto/teneto | teneto/classes/network.py | TemporalNetwork.network_from_array | def network_from_array(self, array):
"""impo
Defines a network from an array.
Parameters
----------
array : array
3D numpy array.
"""
if len(array.shape) == 2:
array = np.array(array, ndmin=3).transpose([1, 2, 0])
teneto.utils.chec... | python | def network_from_array(self, array):
"""impo
Defines a network from an array.
Parameters
----------
array : array
3D numpy array.
"""
if len(array.shape) == 2:
array = np.array(array, ndmin=3).transpose([1, 2, 0])
teneto.utils.chec... | [
"def",
"network_from_array",
"(",
"self",
",",
"array",
")",
":",
"if",
"len",
"(",
"array",
".",
"shape",
")",
"==",
"2",
":",
"array",
"=",
"np",
".",
"array",
"(",
"array",
",",
"ndmin",
"=",
"3",
")",
".",
"transpose",
"(",
"[",
"1",
",",
"... | impo
Defines a network from an array.
Parameters
----------
array : array
3D numpy array. | [
"impo",
"Defines",
"a",
"network",
"from",
"an",
"array",
"."
] | train | https://github.com/wiheto/teneto/blob/80d7a83a9adc1714589b020627c45bd5b66248ab/teneto/classes/network.py#L179-L202 |
wiheto/teneto | teneto/classes/network.py | TemporalNetwork.network_from_df | def network_from_df(self, df):
"""
Defines a network from an array.
Parameters
----------
array : array
Pandas dataframe. Should have columns: \'i\', \'j\', \'t\' where i and j are node indicies and t is the temporal index.
If weighted, should also includ... | python | def network_from_df(self, df):
"""
Defines a network from an array.
Parameters
----------
array : array
Pandas dataframe. Should have columns: \'i\', \'j\', \'t\' where i and j are node indicies and t is the temporal index.
If weighted, should also includ... | [
"def",
"network_from_df",
"(",
"self",
",",
"df",
")",
":",
"teneto",
".",
"utils",
".",
"check_TemporalNetwork_input",
"(",
"df",
",",
"'df'",
")",
"self",
".",
"network",
"=",
"df",
"self",
".",
"_update_network",
"(",
")"
] | Defines a network from an array.
Parameters
----------
array : array
Pandas dataframe. Should have columns: \'i\', \'j\', \'t\' where i and j are node indicies and t is the temporal index.
If weighted, should also include \'weight\'. Each row is an edge. | [
"Defines",
"a",
"network",
"from",
"an",
"array",
"."
] | train | https://github.com/wiheto/teneto/blob/80d7a83a9adc1714589b020627c45bd5b66248ab/teneto/classes/network.py#L213-L225 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.