source_code stringlengths 45 7.24k | function_name stringlengths 1 111 | messages list |
|---|---|---|
def x(self) ->Iterator[Post]:
self._obtain_metadata()
yield from (Post(self._context, node, self if int(node['owner']['id']) ==
self.userid else None) for node in self._context.graphql_node_list(
'e31a871f7301132ceaab56507a66bbb7', {'id': self.userid},
'https://www.instagram.com/{0}/'.format(self.username), lambda d: d
['data']['user']['edge_user_to_photos_of_you'], self._rhx_gis))
| get_tagged_posts | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self) ->Iterator[Post]:\n self._obtain_metadata()\n yield from (Post(self._context, node, self if int(node['owner']['id']) ==\n self.userid else None) for node in self._context.graphql_node_list(\n 'e31a871f7301132ceaab56507a66bbb7', {'id': self.userid},\n 'https://www.instagram.com/{0}/'.format(self.username), lambda d: d\n ['data']['user']['edge_user_to_photos_of_you'], self._rhx_gis))\n",
"role": "user"
},
{
"content": "get_tagged_posts",
"role": "assistant"
}
] |
def x(self) ->Iterator['Profile']:
if not self._context.is_logged_in:
raise LoginRequiredException(
"--login required to get a profile's followers.")
self._obtain_metadata()
yield from (Profile(self._context, node) for node in self._context.
graphql_node_list('37479f2b8209594dde7facb0d904896a', {'id': str(
self.userid)}, 'https://www.instagram.com/' + self.username + '/',
lambda d: d['data']['user']['edge_followed_by'], self._rhx_gis))
| get_followers | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self) ->Iterator['Profile']:\n if not self._context.is_logged_in:\n raise LoginRequiredException(\n \"--login required to get a profile's followers.\")\n self._obtain_metadata()\n yield from (Profile(self._context, node) for node in self._context.\n graphql_node_list('37479f2b8209594dde7facb0d904896a', {'id': str(\n self.userid)}, 'https://www.instagram.com/' + self.username + '/', \n lambda d: d['data']['user']['edge_followed_by'], self._rhx_gis))\n",
"role": "user"
},
{
"content": "get_followers",
"role": "assistant"
}
] |
def x(self) ->Profile:
if not self._owner_profile:
self._owner_profile = Profile.from_id(self._context, self._node[
'owner']['id'])
return self._owner_profile
| owner_profile | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self) ->Profile:\n if not self._owner_profile:\n self._owner_profile = Profile.from_id(self._context, self._node[\n 'owner']['id'])\n return self._owner_profile\n",
"role": "user"
},
{
"content": "owner_profile",
"role": "assistant"
}
] |
def x(self) ->str:
if not self._unique_id:
id_list = [item.mediaid for item in self.get_items()]
id_list.sort()
self._unique_id = str().join([str(self.owner_id)] + list(map(str,
id_list)))
return self._unique_id
| unique_id | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self) ->str:\n if not self._unique_id:\n id_list = [item.mediaid for item in self.get_items()]\n id_list.sort()\n self._unique_id = str().join([str(self.owner_id)] + list(map(str,\n id_list)))\n return self._unique_id\n",
"role": "user"
},
{
"content": "unique_id",
"role": "assistant"
}
] |
def x(self) ->Optional[datetime]:
if self._node['seen']:
return datetime.fromtimestamp(self._node['seen'])
| last_seen_local | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self) ->Optional[datetime]:\n if self._node['seen']:\n return datetime.fromtimestamp(self._node['seen'])\n",
"role": "user"
},
{
"content": "last_seen_local",
"role": "assistant"
}
] |
def x(self) ->Optional[datetime]:
if self._node['seen']:
return datetime.utcfromtimestamp(self._node['seen'])
| last_seen_utc | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self) ->Optional[datetime]:\n if self._node['seen']:\n return datetime.utcfromtimestamp(self._node['seen'])\n",
"role": "user"
},
{
"content": "last_seen_utc",
"role": "assistant"
}
] |
def x(self) ->Iterator[StoryItem]:
yield from (StoryItem(self._context, item, self.owner_profile) for item in
reversed(self._node['items']))
| get_items | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self) ->Iterator[StoryItem]:\n yield from (StoryItem(self._context, item, self.owner_profile) for item in\n reversed(self._node['items']))\n",
"role": "user"
},
{
"content": "get_items",
"role": "assistant"
}
] |
def x(self) ->Iterator[StoryItem]:
self._fetch_items()
yield from (StoryItem(self._context, item, self.owner_profile) for item in
self._items)
| get_items | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self) ->Iterator[StoryItem]:\n self._fetch_items()\n yield from (StoryItem(self._context, item, self.owner_profile) for item in\n self._items)\n",
"role": "user"
},
{
"content": "get_items",
"role": "assistant"
}
] |
def x(session: requests.Session) ->requests.Session:
new = requests.Session()
new.cookies = requests.utils.cookiejar_from_dict(requests.utils.
dict_from_cookiejar(session.cookies))
new.headers = session.headers.copy()
return new
| copy_session | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(session: requests.Session) ->requests.Session:\n new = requests.Session()\n new.cookies = requests.utils.cookiejar_from_dict(requests.utils.\n dict_from_cookiejar(session.cookies))\n new.headers = session.headers.copy()\n return new\n",
"role": "user"
},
{
"content": "copy_session",
"role": "assistant"
}
] |
def x(self, *msg, sep='', end='\n', flush=False):
if not self.quiet:
print(*msg, sep=sep, end=end, flush=flush)
| log | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, *msg, sep='', end='\\n', flush=False):\n if not self.quiet:\n print(*msg, sep=sep, end=end, flush=flush)\n",
"role": "user"
},
{
"content": "log",
"role": "assistant"
}
] |
def x(self, msg, repeat_at_end=True):
print(msg, file=sys.stderr)
if repeat_at_end:
self.error_log.append(msg)
| error | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, msg, repeat_at_end=True):\n print(msg, file=sys.stderr)\n if repeat_at_end:\n self.error_log.append(msg)\n",
"role": "user"
},
{
"content": "error",
"role": "assistant"
}
] |
def x(self):
if self.error_log and not self.quiet:
print('\nErrors occured:', file=sys.stderr)
for err in self.error_log:
print(err, file=sys.stderr)
self._session.close()
| close | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n if self.error_log and not self.quiet:\n print('\\nErrors occured:', file=sys.stderr)\n for err in self.error_log:\n print(err, file=sys.stderr)\n self._session.close()\n",
"role": "user"
},
{
"content": "close",
"role": "assistant"
}
] |
def x(self, extra_info: Optional[str]=None):
try:
yield
except InstaloaderException as err:
if extra_info:
self.error('{}: {}'.format(extra_info, err))
else:
self.error('{}'.format(err))
if self.raise_all_errors:
raise
| error_catcher | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, extra_info: Optional[str]=None):\n try:\n yield\n except InstaloaderException as err:\n if extra_info:\n self.error('{}: {}'.format(extra_info, err))\n else:\n self.error('{}'.format(err))\n if self.raise_all_errors:\n raise\n",
"role": "user"
},
{
"content": "error_catcher",
"role": "assistant"
}
] |
def x(self, empty_session_only: bool=False) ->Dict[str, str]:
header = {'Accept-Encoding': 'gzip, deflate', 'Accept-Language':
'en-US,en;q=0.8', 'Connection': 'keep-alive', 'Content-Length': '0',
'Host': 'www.instagram.com', 'Origin': 'https://www.instagram.com',
'Referer': 'https://www.instagram.com/', 'User-Agent': self.
user_agent, 'X-Instagram-AJAX': '1', 'X-Requested-With':
'XMLHttpRequest'}
if empty_session_only:
del header['Host']
del header['Origin']
del header['Referer']
del header['X-Instagram-AJAX']
del header['X-Requested-With']
return header
| default_http_header | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, empty_session_only: bool=False) ->Dict[str, str]:\n header = {'Accept-Encoding': 'gzip, deflate', 'Accept-Language':\n 'en-US,en;q=0.8', 'Connection': 'keep-alive', 'Content-Length': '0',\n 'Host': 'www.instagram.com', 'Origin': 'https://www.instagram.com',\n 'Referer': 'https://www.instagram.com/', 'User-Agent': self.\n user_agent, 'X-Instagram-AJAX': '1', 'X-Requested-With':\n 'XMLHttpRequest'}\n if empty_session_only:\n del header['Host']\n del header['Origin']\n del header['Referer']\n del header['X-Instagram-AJAX']\n del header['X-Requested-With']\n return header\n",
"role": "user"
},
{
"content": "default_http_header",
"role": "assistant"
}
] |
def x(self) ->requests.Session:
session = requests.Session()
session.cookies.update({'sessionid': '', 'mid': '', 'ig_pr': '1',
'ig_vw': '1920', 'csrftoken': '', 's_network': '', 'ds_user_id': ''})
session.headers.update(self._default_http_header(empty_session_only=True))
return session
| get_anonymous_session | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self) ->requests.Session:\n session = requests.Session()\n session.cookies.update({'sessionid': '', 'mid': '', 'ig_pr': '1',\n 'ig_vw': '1920', 'csrftoken': '', 's_network': '', 'ds_user_id': ''})\n session.headers.update(self._default_http_header(empty_session_only=True))\n return session\n",
"role": "user"
},
{
"content": "get_anonymous_session",
"role": "assistant"
}
] |
def x(self, sessionfile):
pickle.dump(requests.utils.dict_from_cookiejar(self._session.cookies),
sessionfile)
| save_session_to_file | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, sessionfile):\n pickle.dump(requests.utils.dict_from_cookiejar(self._session.cookies),\n sessionfile)\n",
"role": "user"
},
{
"content": "save_session_to_file",
"role": "assistant"
}
] |
def x(self, username, sessionfile):
session = requests.Session()
session.cookies = requests.utils.cookiejar_from_dict(pickle.load(
sessionfile))
session.headers.update(self._default_http_header())
session.headers.update({'X-CSRFToken': session.cookies.get_dict()[
'csrftoken']})
self._session = session
self.username = username
| load_session_from_file | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, username, sessionfile):\n session = requests.Session()\n session.cookies = requests.utils.cookiejar_from_dict(pickle.load(\n sessionfile))\n session.headers.update(self._default_http_header())\n session.headers.update({'X-CSRFToken': session.cookies.get_dict()[\n 'csrftoken']})\n self._session = session\n self.username = username\n",
"role": "user"
},
{
"content": "load_session_from_file",
"role": "assistant"
}
] |
def x(self, user, passwd):
import http.client
http.client._MAXHEADERS = 200
session = requests.Session()
session.cookies.update({'sessionid': '', 'mid': '', 'ig_pr': '1',
'ig_vw': '1920', 'ig_cb': '1', 'csrftoken': '', 's_network': '',
'ds_user_id': ''})
session.headers.update(self._default_http_header())
session.get('https://www.instagram.com/web/__mid/')
csrf_token = session.cookies.get_dict()['csrftoken']
session.headers.update({'X-CSRFToken': csrf_token})
self.do_sleep()
login = session.post('https://www.instagram.com/accounts/login/ajax/',
data={'password': passwd, 'username': user}, allow_redirects=True)
try:
resp_json = login.json()
except json.decoder.JSONDecodeError:
raise ConnectionException('Login error: JSON decode fail, {} - {}.'
.format(login.status_code, login.reason))
if resp_json.get('two_factor_required'):
two_factor_session = copy_session(session)
two_factor_session.headers.update({'X-CSRFToken': csrf_token})
two_factor_session.cookies.update({'csrftoken': csrf_token})
self.two_factor_auth_pending = two_factor_session, user, resp_json[
'two_factor_info']['two_factor_identifier']
raise TwoFactorAuthRequiredException(
'Login error: two-factor authentication required.')
if resp_json.get('checkpoint_url'):
raise ConnectionException(
'Login: Checkpoint required. Point your browser to https://www.instagram.com{}, follow the instructions, then retry.'
.format(resp_json.get('checkpoint_url')))
if resp_json['status'] != 'ok':
if 'message' in resp_json:
raise ConnectionException('Login error: "{}" status, message "{}".'
.format(resp_json['status'], resp_json['message']))
else:
raise ConnectionException('Login error: "{}" status.'.format(
resp_json['status']))
if not resp_json['authenticated']:
if resp_json['user']:
raise BadCredentialsException('Login error: Wrong password.')
else:
raise InvalidArgumentException(
'Login error: User {} does not exist.'.format(user))
session.headers.update({'X-CSRFToken': login.cookies['csrftoken']})
self._session = session
self.username = user
| login | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, user, passwd):\n import http.client\n http.client._MAXHEADERS = 200\n session = requests.Session()\n session.cookies.update({'sessionid': '', 'mid': '', 'ig_pr': '1',\n 'ig_vw': '1920', 'ig_cb': '1', 'csrftoken': '', 's_network': '',\n 'ds_user_id': ''})\n session.headers.update(self._default_http_header())\n session.get('https://www.instagram.com/web/__mid/')\n csrf_token = session.cookies.get_dict()['csrftoken']\n session.headers.update({'X-CSRFToken': csrf_token})\n self.do_sleep()\n login = session.post('https://www.instagram.com/accounts/login/ajax/',\n data={'password': passwd, 'username': user}, allow_redirects=True)\n try:\n resp_json = login.json()\n except json.decoder.JSONDecodeError:\n raise ConnectionException('Login error: JSON decode fail, {} - {}.'\n .format(login.status_code, login.reason))\n if resp_json.get('two_factor_required'):\n two_factor_session = copy_session(session)\n two_factor_session.headers.update({'X-CSRFToken': csrf_token})\n two_factor_session.cookies.update({'csrftoken': csrf_token})\n self.two_factor_auth_pending = two_factor_session, user, resp_json[\n 'two_factor_info']['two_factor_identifier']\n raise TwoFactorAuthRequiredException(\n 'Login error: two-factor authentication required.')\n if resp_json.get('checkpoint_url'):\n raise ConnectionException(\n 'Login: Checkpoint required. Point your browser to https://www.instagram.com{}, follow the instructions, then retry.'\n .format(resp_json.get('checkpoint_url')))\n if resp_json['status'] != 'ok':\n if 'message' in resp_json:\n raise ConnectionException('Login error: \"{}\" status, message \"{}\".'\n .format(resp_json['status'], resp_json['message']))\n else:\n raise ConnectionException('Login error: \"{}\" status.'.format(\n resp_json['status']))\n if not resp_json['authenticated']:\n if resp_json['user']:\n raise BadCredentialsException('Login error: Wrong password.')\n else:\n raise InvalidArgumentException(\n 'Login error: User {} does not exist.'.format(user))\n session.headers.update({'X-CSRFToken': login.cookies['csrftoken']})\n self._session = session\n self.username = user\n",
"role": "user"
},
{
"content": "login",
"role": "assistant"
}
] |
def x(self, two_factor_code):
if not self.two_factor_auth_pending:
raise InvalidArgumentException('No two-factor authentication pending.')
session, user, two_factor_id = self.two_factor_auth_pending
login = session.post(
'https://www.instagram.com/accounts/login/ajax/two_factor/', data={
'username': user, 'verificationCode': two_factor_code, 'identifier':
two_factor_id}, allow_redirects=True)
resp_json = login.json()
if resp_json['status'] != 'ok':
if 'message' in resp_json:
raise BadCredentialsException('Login error: {}'.format(
resp_json['message']))
else:
raise BadCredentialsException('Login error: "{}" status.'.
format(resp_json['status']))
session.headers.update({'X-CSRFToken': login.cookies['csrftoken']})
self._session = session
self.username = user
self.two_factor_auth_pending = None
| two_factor_login | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, two_factor_code):\n if not self.two_factor_auth_pending:\n raise InvalidArgumentException('No two-factor authentication pending.')\n session, user, two_factor_id = self.two_factor_auth_pending\n login = session.post(\n 'https://www.instagram.com/accounts/login/ajax/two_factor/', data={\n 'username': user, 'verificationCode': two_factor_code, 'identifier':\n two_factor_id}, allow_redirects=True)\n resp_json = login.json()\n if resp_json['status'] != 'ok':\n if 'message' in resp_json:\n raise BadCredentialsException('Login error: {}'.format(\n resp_json['message']))\n else:\n raise BadCredentialsException('Login error: \"{}\" status.'.\n format(resp_json['status']))\n session.headers.update({'X-CSRFToken': login.cookies['csrftoken']})\n self._session = session\n self.username = user\n self.two_factor_auth_pending = None\n",
"role": "user"
},
{
"content": "two_factor_login",
"role": "assistant"
}
] |
def x(self, current_time: float):
windows = [10, 11, 15, 20, 30, 60]
print('GraphQL requests:', file=sys.stderr)
for query_hash, times in self._graphql_query_timestamps.items():
print(' {}'.format(query_hash), file=sys.stderr)
for window in windows:
reqs_in_sliding_window = sum(t > current_time - window * 60 for
t in times)
print(' last {} minutes: {} requests'.format(window,
reqs_in_sliding_window), file=sys.stderr)
| dump_query_timestamps | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, current_time: float):\n windows = [10, 11, 15, 20, 30, 60]\n print('GraphQL requests:', file=sys.stderr)\n for query_hash, times in self._graphql_query_timestamps.items():\n print(' {}'.format(query_hash), file=sys.stderr)\n for window in windows:\n reqs_in_sliding_window = sum(t > current_time - window * 60 for\n t in times)\n print(' last {} minutes: {} requests'.format(window,\n reqs_in_sliding_window), file=sys.stderr)\n",
"role": "user"
},
{
"content": "dump_query_timestamps",
"role": "assistant"
}
] |
def x(self, query_hash: str) ->int:
if self.is_logged_in:
max_reqs = {'1cb6ec562846122743b61e492c85999f': 20,
'33ba35852cb50da46f5b5e889df7d159': 20}
else:
max_reqs = {'1cb6ec562846122743b61e492c85999f': 200,
'33ba35852cb50da46f5b5e889df7d159': 200}
return max_reqs.get(query_hash) or min(max_reqs.values())
| graphql_request_count_per_sliding_window | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, query_hash: str) ->int:\n if self.is_logged_in:\n max_reqs = {'1cb6ec562846122743b61e492c85999f': 20,\n '33ba35852cb50da46f5b5e889df7d159': 20}\n else:\n max_reqs = {'1cb6ec562846122743b61e492c85999f': 200,\n '33ba35852cb50da46f5b5e889df7d159': 200}\n return max_reqs.get(query_hash) or min(max_reqs.values())\n",
"role": "user"
},
{
"content": "graphql_request_count_per_sliding_window",
"role": "assistant"
}
] |
def x(self, query_hash: str, current_time: float, untracked_queries: bool=False
) ->int:
sliding_window = 660
if query_hash not in self._graphql_query_timestamps:
self._graphql_query_timestamps[query_hash] = []
self._graphql_query_timestamps[query_hash] = list(filter(lambda t: t >
current_time - 60 * 60, self._graphql_query_timestamps[query_hash]))
reqs_in_sliding_window = list(filter(lambda t: t > current_time -
sliding_window, self._graphql_query_timestamps[query_hash]))
count_per_sliding_window = self._graphql_request_count_per_sliding_window(
query_hash)
if len(reqs_in_sliding_window
) < count_per_sliding_window and not untracked_queries:
return max(0, self._graphql_earliest_next_request_time - current_time)
next_request_time = min(reqs_in_sliding_window) + sliding_window + 6
if untracked_queries:
self._graphql_earliest_next_request_time = next_request_time
return round(max(next_request_time, self.
_graphql_earliest_next_request_time) - current_time)
| graphql_query_waittime | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, query_hash: str, current_time: float, untracked_queries: bool=False\n ) ->int:\n sliding_window = 660\n if query_hash not in self._graphql_query_timestamps:\n self._graphql_query_timestamps[query_hash] = []\n self._graphql_query_timestamps[query_hash] = list(filter(lambda t: t > \n current_time - 60 * 60, self._graphql_query_timestamps[query_hash]))\n reqs_in_sliding_window = list(filter(lambda t: t > current_time -\n sliding_window, self._graphql_query_timestamps[query_hash]))\n count_per_sliding_window = self._graphql_request_count_per_sliding_window(\n query_hash)\n if len(reqs_in_sliding_window\n ) < count_per_sliding_window and not untracked_queries:\n return max(0, self._graphql_earliest_next_request_time - current_time)\n next_request_time = min(reqs_in_sliding_window) + sliding_window + 6\n if untracked_queries:\n self._graphql_earliest_next_request_time = next_request_time\n return round(max(next_request_time, self.\n _graphql_earliest_next_request_time) - current_time)\n",
"role": "user"
},
{
"content": "graphql_query_waittime",
"role": "assistant"
}
] |
def x(self, query_hash: str, untracked_queries: bool=False):
if not untracked_queries:
waittime = self._graphql_query_waittime(query_hash, time.monotonic(
), untracked_queries)
assert waittime >= 0
if waittime > 10:
self.log(
"""
Too many queries in the last time. Need to wait {} seconds, until {:%H:%M}."""
.format(waittime, datetime.now() + timedelta(seconds=waittime))
)
time.sleep(waittime)
if query_hash not in self._graphql_query_timestamps:
self._graphql_query_timestamps[query_hash] = [time.monotonic()]
else:
self._graphql_query_timestamps[query_hash].append(time.monotonic())
else:
text_for_429 = (
'HTTP error code 429 was returned because too many queries occured in the last time. Please do not use Instagram in your browser or run multiple instances of Instaloader in parallel.'
)
print(textwrap.fill(text_for_429), file=sys.stderr)
current_time = time.monotonic()
waittime = self._graphql_query_waittime(query_hash, current_time,
untracked_queries)
assert waittime >= 0
if waittime > 10:
self.log('The request will be retried in {} seconds, at {:%H:%M}.'
.format(waittime, datetime.now() + timedelta(seconds=waittime))
)
self._dump_query_timestamps(current_time)
time.sleep(waittime)
| ratecontrol_graphql_query | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, query_hash: str, untracked_queries: bool=False):\n if not untracked_queries:\n waittime = self._graphql_query_waittime(query_hash, time.monotonic(\n ), untracked_queries)\n assert waittime >= 0\n if waittime > 10:\n self.log(\n \"\"\"\nToo many queries in the last time. Need to wait {} seconds, until {:%H:%M}.\"\"\"\n .format(waittime, datetime.now() + timedelta(seconds=waittime))\n )\n time.sleep(waittime)\n if query_hash not in self._graphql_query_timestamps:\n self._graphql_query_timestamps[query_hash] = [time.monotonic()]\n else:\n self._graphql_query_timestamps[query_hash].append(time.monotonic())\n else:\n text_for_429 = (\n 'HTTP error code 429 was returned because too many queries occured in the last time. Please do not use Instagram in your browser or run multiple instances of Instaloader in parallel.'\n )\n print(textwrap.fill(text_for_429), file=sys.stderr)\n current_time = time.monotonic()\n waittime = self._graphql_query_waittime(query_hash, current_time,\n untracked_queries)\n assert waittime >= 0\n if waittime > 10:\n self.log('The request will be retried in {} seconds, at {:%H:%M}.'\n .format(waittime, datetime.now() + timedelta(seconds=waittime))\n )\n self._dump_query_timestamps(current_time)\n time.sleep(waittime)\n",
"role": "user"
},
{
"content": "ratecontrol_graphql_query",
"role": "assistant"
}
] |
def x(self, path: str, params: Dict[str, Any], host: str=
'www.instagram.com', session: Optional[requests.Session]=None, _attempt=1
) ->Dict[str, Any]:
is_graphql_query = 'query_hash' in params and 'graphql/query' in path
sess = session if session else self._session
try:
self.do_sleep()
if is_graphql_query:
self._ratecontrol_graphql_query(params['query_hash'])
resp = sess.get('https://{0}/{1}'.format(host, path), params=params,
allow_redirects=False)
while resp.is_redirect:
redirect_url = resp.headers['location']
self.log('\nHTTP redirect from https://{0}/{1} to {2}'.format(
host, path, redirect_url))
if redirect_url.startswith('https://{}/'.format(host)):
resp = sess.get(redirect_url if redirect_url.endswith('/') else
redirect_url + '/', params=params, allow_redirects=False)
else:
break
if resp.status_code == 400:
raise QueryReturnedBadRequestException('400 Bad Request')
if resp.status_code == 404:
raise QueryReturnedNotFoundException('404 Not Found')
if resp.status_code == 429:
raise TooManyRequestsException('429 Too Many Requests')
if resp.status_code != 200:
raise ConnectionException('HTTP error code {}.'.format(resp.
status_code))
is_html_query = (not is_graphql_query and not '__a' in params and
host == 'www.instagram.com')
if is_html_query:
match = re.search('window\\._sharedData = (.*);</script>', resp
.text)
if match is None:
raise ConnectionException(
'Could not find "window._sharedData" in html response.')
return json.loads(match.group(1))
else:
resp_json = resp.json()
if 'status' in resp_json and resp_json['status'] != 'ok':
if 'message' in resp_json:
raise ConnectionException('Returned "{}" status, message "{}".'
.format(resp_json['status'], resp_json['message']))
else:
raise ConnectionException('Returned "{}" status.'.format(
resp_json['status']))
return resp_json
except (ConnectionException, json.decoder.JSONDecodeError, requests.
exceptions.RequestException) as err:
error_string = 'JSON Query to {}: {}'.format(path, err)
if _attempt == self.max_connection_attempts:
raise ConnectionException(error_string) from err
self.error(error_string + ' [retrying; skip with ^C]',
repeat_at_end=False)
try:
if isinstance(err, TooManyRequestsException):
self._ratecontrol_graphql_query(params['query_hash'],
untracked_queries=True)
return self.get_json(path=path, params=params, host=host,
session=sess, _attempt=_attempt + 1)
except KeyboardInterrupt:
self.error('[skipped by user]', repeat_at_end=False)
raise ConnectionException(error_string) from err
| get_json | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, path: str, params: Dict[str, Any], host: str=\n 'www.instagram.com', session: Optional[requests.Session]=None, _attempt=1\n ) ->Dict[str, Any]:\n is_graphql_query = 'query_hash' in params and 'graphql/query' in path\n sess = session if session else self._session\n try:\n self.do_sleep()\n if is_graphql_query:\n self._ratecontrol_graphql_query(params['query_hash'])\n resp = sess.get('https://{0}/{1}'.format(host, path), params=params,\n allow_redirects=False)\n while resp.is_redirect:\n redirect_url = resp.headers['location']\n self.log('\\nHTTP redirect from https://{0}/{1} to {2}'.format(\n host, path, redirect_url))\n if redirect_url.startswith('https://{}/'.format(host)):\n resp = sess.get(redirect_url if redirect_url.endswith('/') else\n redirect_url + '/', params=params, allow_redirects=False)\n else:\n break\n if resp.status_code == 400:\n raise QueryReturnedBadRequestException('400 Bad Request')\n if resp.status_code == 404:\n raise QueryReturnedNotFoundException('404 Not Found')\n if resp.status_code == 429:\n raise TooManyRequestsException('429 Too Many Requests')\n if resp.status_code != 200:\n raise ConnectionException('HTTP error code {}.'.format(resp.\n status_code))\n is_html_query = (not is_graphql_query and not '__a' in params and \n host == 'www.instagram.com')\n if is_html_query:\n match = re.search('window\\\\._sharedData = (.*);</script>', resp\n .text)\n if match is None:\n raise ConnectionException(\n 'Could not find \"window._sharedData\" in html response.')\n return json.loads(match.group(1))\n else:\n resp_json = resp.json()\n if 'status' in resp_json and resp_json['status'] != 'ok':\n if 'message' in resp_json:\n raise ConnectionException('Returned \"{}\" status, message \"{}\".'\n .format(resp_json['status'], resp_json['message']))\n else:\n raise ConnectionException('Returned \"{}\" status.'.format(\n resp_json['status']))\n return resp_json\n except (ConnectionException, json.decoder.JSONDecodeError, requests.\n exceptions.RequestException) as err:\n error_string = 'JSON Query to {}: {}'.format(path, err)\n if _attempt == self.max_connection_attempts:\n raise ConnectionException(error_string) from err\n self.error(error_string + ' [retrying; skip with ^C]',\n repeat_at_end=False)\n try:\n if isinstance(err, TooManyRequestsException):\n self._ratecontrol_graphql_query(params['query_hash'],\n untracked_queries=True)\n return self.get_json(path=path, params=params, host=host,\n session=sess, _attempt=_attempt + 1)\n except KeyboardInterrupt:\n self.error('[skipped by user]', repeat_at_end=False)\n raise ConnectionException(error_string) from err\n",
"role": "user"
},
{
"content": "get_json",
"role": "assistant"
}
] |
def x(self, query_hash: str, variables: Dict[str, Any], referer: Optional[
str]=None, rhx_gis: Optional[str]=None) ->Dict[str, Any]:
with copy_session(self._session) as tmpsession:
tmpsession.headers.update(self._default_http_header(
empty_session_only=True))
del tmpsession.headers['Connection']
del tmpsession.headers['Content-Length']
tmpsession.headers['authority'] = 'www.instagram.com'
tmpsession.headers['scheme'] = 'https'
tmpsession.headers['accept'] = '*/*'
if referer is not None:
tmpsession.headers['referer'] = urllib.parse.quote(referer)
variables_json = json.dumps(variables, separators=(',', ':'))
if rhx_gis:
values = '{}:{}'.format(rhx_gis, variables_json)
x_instagram_gis = hashlib.md5(values.encode()).hexdigest()
tmpsession.headers['x-instagram-gis'] = x_instagram_gis
resp_json = self.get_json('graphql/query', params={'query_hash':
query_hash, 'variables': variables_json}, session=tmpsession)
if 'status' not in resp_json:
self.error('GraphQL response did not contain a "status" field.')
return resp_json
| graphql_query | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, query_hash: str, variables: Dict[str, Any], referer: Optional[\n str]=None, rhx_gis: Optional[str]=None) ->Dict[str, Any]:\n with copy_session(self._session) as tmpsession:\n tmpsession.headers.update(self._default_http_header(\n empty_session_only=True))\n del tmpsession.headers['Connection']\n del tmpsession.headers['Content-Length']\n tmpsession.headers['authority'] = 'www.instagram.com'\n tmpsession.headers['scheme'] = 'https'\n tmpsession.headers['accept'] = '*/*'\n if referer is not None:\n tmpsession.headers['referer'] = urllib.parse.quote(referer)\n variables_json = json.dumps(variables, separators=(',', ':'))\n if rhx_gis:\n values = '{}:{}'.format(rhx_gis, variables_json)\n x_instagram_gis = hashlib.md5(values.encode()).hexdigest()\n tmpsession.headers['x-instagram-gis'] = x_instagram_gis\n resp_json = self.get_json('graphql/query', params={'query_hash':\n query_hash, 'variables': variables_json}, session=tmpsession)\n if 'status' not in resp_json:\n self.error('GraphQL response did not contain a \"status\" field.')\n return resp_json\n",
"role": "user"
},
{
"content": "graphql_query",
"role": "assistant"
}
] |
def x(self, query_hash: str, query_variables: Dict[str, Any], query_referer:
Optional[str], edge_extractor: Callable[[Dict[str, Any]], Dict[str, Any
]], rhx_gis: Optional[str]=None, first_data: Optional[Dict[str, Any]]=None
) ->Iterator[Dict[str, Any]]:
def x():
query_variables['first'] = self._graphql_page_length
try:
return edge_extractor(self.graphql_query(query_hash,
query_variables, query_referer, rhx_gis))
except QueryReturnedBadRequestException:
new_page_length = int(self._graphql_page_length / 2)
if new_page_length >= 12:
self._graphql_page_length = new_page_length
self.error(
'HTTP Error 400 (Bad Request) on GraphQL Query. Retrying with shorter page length.'
, repeat_at_end=False)
return _query()
else:
raise
if first_data:
data = first_data
else:
data = _query()
yield from (edge['node'] for edge in data['edges'])
while data['page_info']['has_next_page']:
query_variables['after'] = data['page_info']['end_cursor']
data = _query()
yield from (edge['node'] for edge in data['edges'])
| graphql_node_list | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, query_hash: str, query_variables: Dict[str, Any], query_referer:\n Optional[str], edge_extractor: Callable[[Dict[str, Any]], Dict[str, Any\n ]], rhx_gis: Optional[str]=None, first_data: Optional[Dict[str, Any]]=None\n ) ->Iterator[Dict[str, Any]]:\n\n def x():\n query_variables['first'] = self._graphql_page_length\n try:\n return edge_extractor(self.graphql_query(query_hash,\n query_variables, query_referer, rhx_gis))\n except QueryReturnedBadRequestException:\n new_page_length = int(self._graphql_page_length / 2)\n if new_page_length >= 12:\n self._graphql_page_length = new_page_length\n self.error(\n 'HTTP Error 400 (Bad Request) on GraphQL Query. Retrying with shorter page length.'\n , repeat_at_end=False)\n return _query()\n else:\n raise\n if first_data:\n data = first_data\n else:\n data = _query()\n yield from (edge['node'] for edge in data['edges'])\n while data['page_info']['has_next_page']:\n query_variables['after'] = data['page_info']['end_cursor']\n data = _query()\n yield from (edge['node'] for edge in data['edges'])\n",
"role": "user"
},
{
"content": "graphql_node_list",
"role": "assistant"
}
] |
def x(self, path: str, params: Dict[str, Any]) ->Dict[str, Any]:
with copy_session(self._session) as tempsession:
tempsession.headers['User-Agent'] = (
'Instagram 10.3.2 (iPhone7,2; iPhone OS 9_3_3; en_US; en-US; scale=2.00; 750x1334) AppleWebKit/420+'
)
for header in ['Host', 'Origin', 'X-Instagram-AJAX', 'X-Requested-With'
]:
tempsession.headers.pop(header, None)
return self.get_json(path, params, 'i.instagram.com', tempsession)
| get_iphone_json | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, path: str, params: Dict[str, Any]) ->Dict[str, Any]:\n with copy_session(self._session) as tempsession:\n tempsession.headers['User-Agent'] = (\n 'Instagram 10.3.2 (iPhone7,2; iPhone OS 9_3_3; en_US; en-US; scale=2.00; 750x1334) AppleWebKit/420+'\n )\n for header in ['Host', 'Origin', 'X-Instagram-AJAX', 'X-Requested-With'\n ]:\n tempsession.headers.pop(header, None)\n return self.get_json(path, params, 'i.instagram.com', tempsession)\n",
"role": "user"
},
{
"content": "get_iphone_json",
"role": "assistant"
}
] |
def x(self, resp: Union[bytes, requests.Response], filename: str) ->None:
self.log(filename, end=' ', flush=True)
with open(filename, 'wb') as file:
if isinstance(resp, requests.Response):
shutil.copyfileobj(resp.raw, file)
else:
file.write(resp)
| write_raw | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, resp: Union[bytes, requests.Response], filename: str) ->None:\n self.log(filename, end=' ', flush=True)\n with open(filename, 'wb') as file:\n if isinstance(resp, requests.Response):\n shutil.copyfileobj(resp.raw, file)\n else:\n file.write(resp)\n",
"role": "user"
},
{
"content": "write_raw",
"role": "assistant"
}
] |
def x(self, url: str, _attempt=1) ->requests.Response:
with self.get_anonymous_session() as anonymous_session:
resp = anonymous_session.get(url, stream=True)
if resp.status_code == 200:
resp.raw.decode_content = True
return resp
else:
if resp.status_code == 403:
raise QueryReturnedForbiddenException('403 when accessing {}.'.
format(url))
if resp.status_code == 404:
raise QueryReturnedNotFoundException('404 when accessing {}.'.
format(url))
raise ConnectionException('HTTP error code {}.'.format(resp.
status_code))
| get_raw | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, url: str, _attempt=1) ->requests.Response:\n with self.get_anonymous_session() as anonymous_session:\n resp = anonymous_session.get(url, stream=True)\n if resp.status_code == 200:\n resp.raw.decode_content = True\n return resp\n else:\n if resp.status_code == 403:\n raise QueryReturnedForbiddenException('403 when accessing {}.'.\n format(url))\n if resp.status_code == 404:\n raise QueryReturnedNotFoundException('404 when accessing {}.'.\n format(url))\n raise ConnectionException('HTTP error code {}.'.format(resp.\n status_code))\n",
"role": "user"
},
{
"content": "get_raw",
"role": "assistant"
}
] |
def x(self, url: str, filename: str) ->None:
self.write_raw(self.get_raw(url), filename)
| get_and_write_raw | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, url: str, filename: str) ->None:\n self.write_raw(self.get_raw(url), filename)\n",
"role": "user"
},
{
"content": "get_and_write_raw",
"role": "assistant"
}
] |
def x(self) ->Optional[str]:
if self.is_logged_in:
return None
if not self._root_rhx_gis:
self._root_rhx_gis = self.get_json('', {})['rhx_gis']
return self._root_rhx_gis
| root_rhx_gis | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self) ->Optional[str]:\n if self.is_logged_in:\n return None\n if not self._root_rhx_gis:\n self._root_rhx_gis = self.get_json('', {})['rhx_gis']\n return self._root_rhx_gis\n",
"role": "user"
},
{
"content": "root_rhx_gis",
"role": "assistant"
}
] |
def x(username: str) ->str:
dirname = tempfile.gettempdir() + '/' + '.instaloader-' + getpass.getuser()
filename = dirname + '/' + 'session-' + username
return filename.lower()
| get_default_session_filename | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(username: str) ->str:\n dirname = tempfile.gettempdir() + '/' + '.instaloader-' + getpass.getuser()\n filename = dirname + '/' + 'session-' + username\n return filename.lower()\n",
"role": "user"
},
{
"content": "get_default_session_filename",
"role": "assistant"
}
] |
def x(func: Callable) ->Callable:
@wraps(func)
def x(instaloader, *args, **kwargs):
if not instaloader.context.is_logged_in:
raise LoginRequiredException('--login=USERNAME required.')
return func(instaloader, *args, **kwargs)
call.__doc__ += (
':raises LoginRequiredException: If called without being logged in.\n')
return call
| requires_login | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(func: Callable) ->Callable:\n\n @wraps(func)\n def x(instaloader, *args, **kwargs):\n if not instaloader.context.is_logged_in:\n raise LoginRequiredException('--login=USERNAME required.')\n return func(instaloader, *args, **kwargs)\n call.__doc__ += (\n ':raises LoginRequiredException: If called without being logged in.\\n')\n return call\n",
"role": "user"
},
{
"content": "requires_login",
"role": "assistant"
}
] |
def x(func: Callable) ->Callable:
@wraps(func)
def x(instaloader, *args, **kwargs):
try:
return func(instaloader, *args, **kwargs)
except (urllib3.exceptions.HTTPError, requests.exceptions.
RequestException, ConnectionException) as err:
error_string = '{}({}): {}'.format(func.__name__, ', '.join([
repr(arg) for arg in args]), err)
if (kwargs.get('_attempt') or 1
) == instaloader.context.max_connection_attempts:
raise ConnectionException(error_string) from None
instaloader.context.error(error_string +
' [retrying; skip with ^C]', repeat_at_end=False)
try:
if kwargs.get('_attempt'):
kwargs['_attempt'] += 1
else:
kwargs['_attempt'] = 2
instaloader.context.do_sleep()
return call(instaloader, *args, **kwargs)
except KeyboardInterrupt:
instaloader.context.error('[skipped by user]',
repeat_at_end=False)
raise ConnectionException(error_string) from None
return call
| retry_on_connection_error | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(func: Callable) ->Callable:\n\n @wraps(func)\n def x(instaloader, *args, **kwargs):\n try:\n return func(instaloader, *args, **kwargs)\n except (urllib3.exceptions.HTTPError, requests.exceptions.\n RequestException, ConnectionException) as err:\n error_string = '{}({}): {}'.format(func.__name__, ', '.join([\n repr(arg) for arg in args]), err)\n if (kwargs.get('_attempt') or 1\n ) == instaloader.context.max_connection_attempts:\n raise ConnectionException(error_string) from None\n instaloader.context.error(error_string +\n ' [retrying; skip with ^C]', repeat_at_end=False)\n try:\n if kwargs.get('_attempt'):\n kwargs['_attempt'] += 1\n else:\n kwargs['_attempt'] = 2\n instaloader.context.do_sleep()\n return call(instaloader, *args, **kwargs)\n except KeyboardInterrupt:\n instaloader.context.error('[skipped by user]',\n repeat_at_end=False)\n raise ConnectionException(error_string) from None\n return call\n",
"role": "user"
},
{
"content": "retry_on_connection_error",
"role": "assistant"
}
] |
def x(self, key, args, kwargs):
if hasattr(self._item, key):
return getattr(self._item, key)
return super().get_value(key, args, kwargs)
| get_value | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, key, args, kwargs):\n if hasattr(self._item, key):\n return getattr(self._item, key)\n return super().get_value(key, args, kwargs)\n",
"role": "user"
},
{
"content": "get_value",
"role": "assistant"
}
] |
def x(self, value, format_spec):
if isinstance(value, datetime) and not format_spec:
return super().format_field(value, '%Y-%m-%d_%H-%M-%S')
if value is None:
return ''
return super().format_field(value, format_spec)
| format_field | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, value, format_spec):\n if isinstance(value, datetime) and not format_spec:\n return super().format_field(value, '%Y-%m-%d_%H-%M-%S')\n if value is None:\n return ''\n return super().format_field(value, format_spec)\n",
"role": "user"
},
{
"content": "format_field",
"role": "assistant"
}
] |
def x(self, format_string, args, kwargs):
ret = super().vformat(format_string, args, kwargs)
if platform.system() == 'Windows':
ret = ret.replace(':', '꞉').replace('<', '﹤').replace('>', '﹥'
).replace('"', '"')
ret = ret.replace('\\', '\').replace('|', '|').replace('?', '︖'
).replace('*', '*')
return ret
| vformat | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, format_string, args, kwargs):\n ret = super().vformat(format_string, args, kwargs)\n if platform.system() == 'Windows':\n ret = ret.replace(':', '꞉').replace('<', '﹤').replace('>', '﹥'\n ).replace('\"', '"')\n ret = ret.replace('\\\\', '\').replace('|', '|').replace('?', '︖'\n ).replace('*', '*')\n return ret\n",
"role": "user"
},
{
"content": "vformat",
"role": "assistant"
}
] |
def x(self):
new_loader = Instaloader(self.context.sleep, self.context.quiet, self.
context.user_agent, self.dirname_pattern, self.filename_pattern,
download_pictures=self.download_pictures, download_videos=self.
download_videos, download_video_thumbnails=self.
download_video_thumbnails, download_geotags=self.download_geotags,
download_comments=self.download_comments, save_metadata=self.
save_metadata, compress_json=self.compress_json,
post_metadata_txt_pattern=self.post_metadata_txt_pattern,
storyitem_metadata_txt_pattern=self.storyitem_metadata_txt_pattern,
max_connection_attempts=self.context.max_connection_attempts)
yield new_loader
self.context.error_log.extend(new_loader.context.error_log)
new_loader.context.error_log = []
new_loader.close()
| anonymous_copy | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n new_loader = Instaloader(self.context.sleep, self.context.quiet, self.\n context.user_agent, self.dirname_pattern, self.filename_pattern,\n download_pictures=self.download_pictures, download_videos=self.\n download_videos, download_video_thumbnails=self.\n download_video_thumbnails, download_geotags=self.download_geotags,\n download_comments=self.download_comments, save_metadata=self.\n save_metadata, compress_json=self.compress_json,\n post_metadata_txt_pattern=self.post_metadata_txt_pattern,\n storyitem_metadata_txt_pattern=self.storyitem_metadata_txt_pattern,\n max_connection_attempts=self.context.max_connection_attempts)\n yield new_loader\n self.context.error_log.extend(new_loader.context.error_log)\n new_loader.context.error_log = []\n new_loader.close()\n",
"role": "user"
},
{
"content": "anonymous_copy",
"role": "assistant"
}
] |
def x(self, filename: str, url: str, mtime: datetime, filename_suffix:
Optional[str]=None, _attempt: int=1) ->bool:
urlmatch = re.search('\\.[a-z0-9]*\\?', url)
file_extension = url[-3:] if urlmatch is None else urlmatch.group(0)[1:-1]
if filename_suffix is not None:
filename += '_' + filename_suffix
filename += '.' + file_extension
if self.commit_mode:
if self._committed and os.path.isfile(filename):
self.context.log(filename + ' exists', end=' ', flush=True)
return False
elif os.path.isfile(filename):
self.context.log(filename + ' exists', end=' ', flush=True)
return False
self.context.get_and_write_raw(url, filename)
os.utime(filename, (datetime.now().timestamp(), mtime.timestamp()))
return True
| download_pic | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, filename: str, url: str, mtime: datetime, filename_suffix:\n Optional[str]=None, _attempt: int=1) ->bool:\n urlmatch = re.search('\\\\.[a-z0-9]*\\\\?', url)\n file_extension = url[-3:] if urlmatch is None else urlmatch.group(0)[1:-1]\n if filename_suffix is not None:\n filename += '_' + filename_suffix\n filename += '.' + file_extension\n if self.commit_mode:\n if self._committed and os.path.isfile(filename):\n self.context.log(filename + ' exists', end=' ', flush=True)\n return False\n elif os.path.isfile(filename):\n self.context.log(filename + ' exists', end=' ', flush=True)\n return False\n self.context.get_and_write_raw(url, filename)\n os.utime(filename, (datetime.now().timestamp(), mtime.timestamp()))\n return True\n",
"role": "user"
},
{
"content": "download_pic",
"role": "assistant"
}
] |
def x(self, filename: str, structure: JsonExportable) ->None:
if self.compress_json:
filename += '.json.xz'
else:
filename += '.json'
save_structure_to_file(structure, filename)
if isinstance(structure, (Post, StoryItem)):
self.context.log('json', end=' ', flush=True)
| save_metadata_json | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, filename: str, structure: JsonExportable) ->None:\n if self.compress_json:\n filename += '.json.xz'\n else:\n filename += '.json'\n save_structure_to_file(structure, filename)\n if isinstance(structure, (Post, StoryItem)):\n self.context.log('json', end=' ', flush=True)\n",
"role": "user"
},
{
"content": "save_metadata_json",
"role": "assistant"
}
] |
def x(self, filename: str, mtime: datetime, caption: str) ->None:
def x(caption):
pcaption = caption.replace('\n', ' ').strip()
return '[' + (pcaption[:29] + u'…' if len(pcaption) > 31 else pcaption
) + ']'
filename += '.txt'
caption += '\n'
pcaption = _elliptify(caption)
caption = caption.encode('UTF-8')
with suppress(FileNotFoundError):
with open(filename, 'rb') as file:
file_caption = file.read()
if file_caption.replace(b'\r\n', b'\n') == caption.replace(b'\r\n',
b'\n'):
try:
self.context.log(pcaption + ' unchanged', end=' ', flush=True)
except UnicodeEncodeError:
self.context.log('txt unchanged', end=' ', flush=True)
return None
else:
def x(index):
return filename if index == 0 else '{0}_old_{2:02}{1}'.format(*
os.path.splitext(filename), index)
i = 0
while os.path.isfile(get_filename(i)):
i = i + 1
for index in range(i, 0, -1):
os.rename(get_filename(index - 1), get_filename(index))
try:
self.context.log(_elliptify(file_caption.decode('UTF-8')) +
' updated', end=' ', flush=True)
except UnicodeEncodeError:
self.context.log('txt updated', end=' ', flush=True)
try:
self.context.log(pcaption, end=' ', flush=True)
except UnicodeEncodeError:
self.context.log('txt', end=' ', flush=True)
with open(filename, 'wb') as text_file:
shutil.copyfileobj(BytesIO(caption), text_file)
os.utime(filename, (datetime.now().timestamp(), mtime.timestamp()))
| save_caption | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, filename: str, mtime: datetime, caption: str) ->None:\n\n def x(caption):\n pcaption = caption.replace('\\n', ' ').strip()\n return '[' + (pcaption[:29] + u'…' if len(pcaption) > 31 else pcaption\n ) + ']'\n filename += '.txt'\n caption += '\\n'\n pcaption = _elliptify(caption)\n caption = caption.encode('UTF-8')\n with suppress(FileNotFoundError):\n with open(filename, 'rb') as file:\n file_caption = file.read()\n if file_caption.replace(b'\\r\\n', b'\\n') == caption.replace(b'\\r\\n',\n b'\\n'):\n try:\n self.context.log(pcaption + ' unchanged', end=' ', flush=True)\n except UnicodeEncodeError:\n self.context.log('txt unchanged', end=' ', flush=True)\n return None\n else:\n\n def x(index):\n return filename if index == 0 else '{0}_old_{2:02}{1}'.format(*\n os.path.splitext(filename), index)\n i = 0\n while os.path.isfile(get_filename(i)):\n i = i + 1\n for index in range(i, 0, -1):\n os.rename(get_filename(index - 1), get_filename(index))\n try:\n self.context.log(_elliptify(file_caption.decode('UTF-8')) +\n ' updated', end=' ', flush=True)\n except UnicodeEncodeError:\n self.context.log('txt updated', end=' ', flush=True)\n try:\n self.context.log(pcaption, end=' ', flush=True)\n except UnicodeEncodeError:\n self.context.log('txt', end=' ', flush=True)\n with open(filename, 'wb') as text_file:\n shutil.copyfileobj(BytesIO(caption), text_file)\n os.utime(filename, (datetime.now().timestamp(), mtime.timestamp()))\n",
"role": "user"
},
{
"content": "save_caption",
"role": "assistant"
}
] |
def x(self, filename: str, location: PostLocation, mtime: datetime) ->None:
filename += '_location.txt'
location_string = (location.name + '\n' +
'https://maps.google.com/maps?q={0},{1}&ll={0},{1}\n'.format(
location.lat, location.lng))
with open(filename, 'wb') as text_file:
shutil.copyfileobj(BytesIO(location_string.encode()), text_file)
os.utime(filename, (datetime.now().timestamp(), mtime.timestamp()))
self.context.log('geo', end=' ', flush=True)
| save_location | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, filename: str, location: PostLocation, mtime: datetime) ->None:\n filename += '_location.txt'\n location_string = (location.name + '\\n' +\n 'https://maps.google.com/maps?q={0},{1}&ll={0},{1}\\n'.format(\n location.lat, location.lng))\n with open(filename, 'wb') as text_file:\n shutil.copyfileobj(BytesIO(location_string.encode()), text_file)\n os.utime(filename, (datetime.now().timestamp(), mtime.timestamp()))\n self.context.log('geo', end=' ', flush=True)\n",
"role": "user"
},
{
"content": "save_location",
"role": "assistant"
}
] |
def x(self, profile: Profile, _attempt: int=1) ->None:
def x(epoch: datetime) ->str:
return epoch.strftime('%Y-%m-%d_%H-%M-%S_UTC')
profile_pic_response = self.context.get_raw(profile.profile_pic_url)
if 'Last-Modified' in profile_pic_response.headers:
date_object = datetime.strptime(profile_pic_response.headers[
'Last-Modified'], '%a, %d %b %Y %H:%M:%S GMT')
profile_pic_bytes = None
profile_pic_identifier = _epoch_to_string(date_object)
else:
date_object = None
profile_pic_bytes = profile_pic_response.content
profile_pic_identifier = md5(profile_pic_bytes).hexdigest()[:16]
profile_pic_extension = 'jpg'
if format_string_contains_key(self.dirname_pattern, 'profile'
) or format_string_contains_key(self.dirname_pattern, 'target'):
filename = '{0}/{1}_profile_pic.{2}'.format(self.dirname_pattern.
format(profile=profile.username.lower(), target=profile.
username.lower()), profile_pic_identifier, profile_pic_extension)
else:
filename = '{0}/{1}_{2}_profile_pic.{3}'.format(self.
dirname_pattern.format(), profile.username.lower(),
profile_pic_identifier, profile_pic_extension)
content_length = profile_pic_response.headers.get('Content-Length', None)
if os.path.isfile(filename) and (not self.context.is_logged_in or
content_length is not None and os.path.getsize(filename) >= int(
content_length)):
self.context.log(filename + ' already exists')
return None
self.context.write_raw(profile_pic_bytes if profile_pic_bytes else
profile_pic_response, filename)
if date_object:
os.utime(filename, (datetime.now().timestamp(), date_object.
timestamp()))
self.context.log('')
| download_profilepic | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, profile: Profile, _attempt: int=1) ->None:\n\n def x(epoch: datetime) ->str:\n return epoch.strftime('%Y-%m-%d_%H-%M-%S_UTC')\n profile_pic_response = self.context.get_raw(profile.profile_pic_url)\n if 'Last-Modified' in profile_pic_response.headers:\n date_object = datetime.strptime(profile_pic_response.headers[\n 'Last-Modified'], '%a, %d %b %Y %H:%M:%S GMT')\n profile_pic_bytes = None\n profile_pic_identifier = _epoch_to_string(date_object)\n else:\n date_object = None\n profile_pic_bytes = profile_pic_response.content\n profile_pic_identifier = md5(profile_pic_bytes).hexdigest()[:16]\n profile_pic_extension = 'jpg'\n if format_string_contains_key(self.dirname_pattern, 'profile'\n ) or format_string_contains_key(self.dirname_pattern, 'target'):\n filename = '{0}/{1}_profile_pic.{2}'.format(self.dirname_pattern.\n format(profile=profile.username.lower(), target=profile.\n username.lower()), profile_pic_identifier, profile_pic_extension)\n else:\n filename = '{0}/{1}_{2}_profile_pic.{3}'.format(self.\n dirname_pattern.format(), profile.username.lower(),\n profile_pic_identifier, profile_pic_extension)\n content_length = profile_pic_response.headers.get('Content-Length', None)\n if os.path.isfile(filename) and (not self.context.is_logged_in or \n content_length is not None and os.path.getsize(filename) >= int(\n content_length)):\n self.context.log(filename + ' already exists')\n return None\n self.context.write_raw(profile_pic_bytes if profile_pic_bytes else\n profile_pic_response, filename)\n if date_object:\n os.utime(filename, (datetime.now().timestamp(), date_object.\n timestamp()))\n self.context.log('')\n",
"role": "user"
},
{
"content": "download_profilepic",
"role": "assistant"
}
] |
def x(self, filename: Optional[str]=None) ->None:
if filename is None:
filename = get_default_session_filename(self.context.username)
dirname = os.path.dirname(filename)
if dirname != '' and not os.path.exists(dirname):
os.makedirs(dirname)
os.chmod(dirname, 448)
with open(filename, 'wb') as sessionfile:
os.chmod(filename, 384)
self.context.save_session_to_file(sessionfile)
self.context.log('Saved session to %s.' % filename)
| save_session_to_file | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, filename: Optional[str]=None) ->None:\n if filename is None:\n filename = get_default_session_filename(self.context.username)\n dirname = os.path.dirname(filename)\n if dirname != '' and not os.path.exists(dirname):\n os.makedirs(dirname)\n os.chmod(dirname, 448)\n with open(filename, 'wb') as sessionfile:\n os.chmod(filename, 384)\n self.context.save_session_to_file(sessionfile)\n self.context.log('Saved session to %s.' % filename)\n",
"role": "user"
},
{
"content": "save_session_to_file",
"role": "assistant"
}
] |
def x(self, username: str, filename: Optional[str]=None) ->None:
if filename is None:
filename = get_default_session_filename(username)
with open(filename, 'rb') as sessionfile:
self.context.load_session_from_file(username, sessionfile)
self.context.log('Loaded session from %s.' % filename)
| load_session_from_file | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, username: str, filename: Optional[str]=None) ->None:\n if filename is None:\n filename = get_default_session_filename(username)\n with open(filename, 'rb') as sessionfile:\n self.context.load_session_from_file(username, sessionfile)\n self.context.log('Loaded session from %s.' % filename)\n",
"role": "user"
},
{
"content": "load_session_from_file",
"role": "assistant"
}
] |
def x(self, user: str, passwd: str) ->None:
self.context.login(user, passwd)
| login | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, user: str, passwd: str) ->None:\n self.context.login(user, passwd)\n",
"role": "user"
},
{
"content": "login",
"role": "assistant"
}
] |
def x(self, item: Union[Post, StoryItem], target: Optional[str]=None):
return _PostPathFormatter(item).format(self.filename_pattern, target=target
)
| format_filename | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, item: Union[Post, StoryItem], target: Optional[str]=None):\n return _PostPathFormatter(item).format(self.filename_pattern, target=target\n )\n",
"role": "user"
},
{
"content": "format_filename",
"role": "assistant"
}
] |
def x(self, post: Post, target: str) ->bool:
dirname = _PostPathFormatter(post).format(self.dirname_pattern, target=
target)
filename = dirname + '/' + self.format_filename(post, target=target)
os.makedirs(os.path.dirname(filename), exist_ok=True)
downloaded = True
self._committed = self.check_if_committed(filename)
if self.download_pictures:
if post.typename == 'GraphSidecar':
edge_number = 1
for sidecar_node in post.get_sidecar_nodes():
if (not sidecar_node.is_video or self.
download_video_thumbnails is True):
downloaded &= self.download_pic(filename=filename, url=
sidecar_node.display_url, mtime=post.date_local,
filename_suffix=str(edge_number))
if sidecar_node.is_video and self.download_videos is True:
downloaded &= self.download_pic(filename=filename, url=
sidecar_node.video_url, mtime=post.date_local,
filename_suffix=str(edge_number))
edge_number += 1
elif post.typename == 'GraphImage':
downloaded = self.download_pic(filename=filename, url=post.url,
mtime=post.date_local)
elif post.typename == 'GraphVideo':
if self.download_video_thumbnails is True:
downloaded = self.download_pic(filename=filename, url=post.
url, mtime=post.date_local)
else:
self.context.error('Warning: {0} has unknown typename: {1}'.
format(post, post.typename))
metadata_string = _ArbitraryItemFormatter(post).format(self.
post_metadata_txt_pattern).strip()
if metadata_string:
self.save_caption(filename=filename, mtime=post.date_local, caption
=metadata_string)
if post.is_video and self.download_videos is True:
downloaded &= self.download_pic(filename=filename, url=post.
video_url, mtime=post.date_local)
if self.download_geotags and post.location:
self.save_location(filename, post.location, post.date_local)
if self.download_comments is True:
self.update_comments(filename=filename, post=post)
if self.save_metadata is not False:
self.save_metadata_json(filename, post)
self.context.log()
return downloaded
| download_post | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, post: Post, target: str) ->bool:\n dirname = _PostPathFormatter(post).format(self.dirname_pattern, target=\n target)\n filename = dirname + '/' + self.format_filename(post, target=target)\n os.makedirs(os.path.dirname(filename), exist_ok=True)\n downloaded = True\n self._committed = self.check_if_committed(filename)\n if self.download_pictures:\n if post.typename == 'GraphSidecar':\n edge_number = 1\n for sidecar_node in post.get_sidecar_nodes():\n if (not sidecar_node.is_video or self.\n download_video_thumbnails is True):\n downloaded &= self.download_pic(filename=filename, url=\n sidecar_node.display_url, mtime=post.date_local,\n filename_suffix=str(edge_number))\n if sidecar_node.is_video and self.download_videos is True:\n downloaded &= self.download_pic(filename=filename, url=\n sidecar_node.video_url, mtime=post.date_local,\n filename_suffix=str(edge_number))\n edge_number += 1\n elif post.typename == 'GraphImage':\n downloaded = self.download_pic(filename=filename, url=post.url,\n mtime=post.date_local)\n elif post.typename == 'GraphVideo':\n if self.download_video_thumbnails is True:\n downloaded = self.download_pic(filename=filename, url=post.\n url, mtime=post.date_local)\n else:\n self.context.error('Warning: {0} has unknown typename: {1}'.\n format(post, post.typename))\n metadata_string = _ArbitraryItemFormatter(post).format(self.\n post_metadata_txt_pattern).strip()\n if metadata_string:\n self.save_caption(filename=filename, mtime=post.date_local, caption\n =metadata_string)\n if post.is_video and self.download_videos is True:\n downloaded &= self.download_pic(filename=filename, url=post.\n video_url, mtime=post.date_local)\n if self.download_geotags and post.location:\n self.save_location(filename, post.location, post.date_local)\n if self.download_comments is True:\n self.update_comments(filename=filename, post=post)\n if self.save_metadata is not False:\n self.save_metadata_json(filename, post)\n self.context.log()\n return downloaded\n",
"role": "user"
},
{
"content": "download_post",
"role": "assistant"
}
] |
def x(self, userids: Optional[List[int]]=None) ->Iterator[Story]:
if not userids:
data = self.context.graphql_query('d15efd8c0c5b23f0ef71f18bf363c704',
{'only_stories': True})['data']['user']
if data is None:
raise BadResponseException('Bad stories reel JSON.')
userids = list(edge['node']['id'] for edge in data[
'feed_reels_tray']['edge_reels_tray_to_reel']['edges'])
def x():
userids_per_query = 100
for i in range(0, len(userids), userids_per_query):
yield userids[i:i + userids_per_query]
for userid_chunk in _userid_chunks():
stories = self.context.graphql_query('bf41e22b1c4ba4c9f31b844ebb7d9056'
, {'reel_ids': userid_chunk, 'precomposed_overlay': False})['data']
yield from (Story(self.context, media) for media in stories[
'reels_media'])
| get_stories | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, userids: Optional[List[int]]=None) ->Iterator[Story]:\n if not userids:\n data = self.context.graphql_query('d15efd8c0c5b23f0ef71f18bf363c704',\n {'only_stories': True})['data']['user']\n if data is None:\n raise BadResponseException('Bad stories reel JSON.')\n userids = list(edge['node']['id'] for edge in data[\n 'feed_reels_tray']['edge_reels_tray_to_reel']['edges'])\n\n def x():\n userids_per_query = 100\n for i in range(0, len(userids), userids_per_query):\n yield userids[i:i + userids_per_query]\n for userid_chunk in _userid_chunks():\n stories = self.context.graphql_query('bf41e22b1c4ba4c9f31b844ebb7d9056'\n , {'reel_ids': userid_chunk, 'precomposed_overlay': False})['data']\n yield from (Story(self.context, media) for media in stories[\n 'reels_media'])\n",
"role": "user"
},
{
"content": "get_stories",
"role": "assistant"
}
] |
def x(self, userids: Optional[List[Union[int, Profile]]]=None, fast_update:
bool=False, filename_target: Optional[str]=':stories', storyitem_filter:
Optional[Callable[[StoryItem], bool]]=None) ->None:
if not userids:
self.context.log('Retrieving all visible stories...')
else:
userids = [(p if isinstance(p, int) else p.userid) for p in userids]
for user_story in self.get_stories(userids):
name = user_story.owner_username
self.context.log('Retrieving stories from profile {}.'.format(name))
totalcount = user_story.itemcount
count = 1
for item in user_story.get_items():
if storyitem_filter is not None and not storyitem_filter(item):
self.context.log('<{} skipped>'.format(item), flush=True)
continue
self.context.log('[%3i/%3i] ' % (count, totalcount), end='',
flush=True)
count += 1
with self.context.error_catcher('Download story from user {}'.
format(name)):
downloaded = self.download_storyitem(item, filename_target if
filename_target else name)
if fast_update and not downloaded:
break
| download_stories | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, userids: Optional[List[Union[int, Profile]]]=None, fast_update:\n bool=False, filename_target: Optional[str]=':stories', storyitem_filter:\n Optional[Callable[[StoryItem], bool]]=None) ->None:\n if not userids:\n self.context.log('Retrieving all visible stories...')\n else:\n userids = [(p if isinstance(p, int) else p.userid) for p in userids]\n for user_story in self.get_stories(userids):\n name = user_story.owner_username\n self.context.log('Retrieving stories from profile {}.'.format(name))\n totalcount = user_story.itemcount\n count = 1\n for item in user_story.get_items():\n if storyitem_filter is not None and not storyitem_filter(item):\n self.context.log('<{} skipped>'.format(item), flush=True)\n continue\n self.context.log('[%3i/%3i] ' % (count, totalcount), end='',\n flush=True)\n count += 1\n with self.context.error_catcher('Download story from user {}'.\n format(name)):\n downloaded = self.download_storyitem(item, filename_target if\n filename_target else name)\n if fast_update and not downloaded:\n break\n",
"role": "user"
},
{
"content": "download_stories",
"role": "assistant"
}
] |
def x(self, item: StoryItem, target: str) ->bool:
date_local = item.date_local
dirname = _PostPathFormatter(item).format(self.dirname_pattern, target=
target)
filename = dirname + '/' + self.format_filename(item, target=target)
os.makedirs(os.path.dirname(filename), exist_ok=True)
downloaded = False
if not item.is_video or self.download_video_thumbnails is True:
url = item.url
downloaded = self.download_pic(filename=filename, url=url, mtime=
date_local)
if item.is_video and self.download_videos is True:
downloaded |= self.download_pic(filename=filename, url=item.
video_url, mtime=date_local)
metadata_string = _ArbitraryItemFormatter(item).format(self.
storyitem_metadata_txt_pattern).strip()
if metadata_string:
self.save_caption(filename=filename, mtime=item.date_local, caption
=metadata_string)
if self.save_metadata is not False:
self.save_metadata_json(filename, item)
self.context.log()
return downloaded
| download_storyitem | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, item: StoryItem, target: str) ->bool:\n date_local = item.date_local\n dirname = _PostPathFormatter(item).format(self.dirname_pattern, target=\n target)\n filename = dirname + '/' + self.format_filename(item, target=target)\n os.makedirs(os.path.dirname(filename), exist_ok=True)\n downloaded = False\n if not item.is_video or self.download_video_thumbnails is True:\n url = item.url\n downloaded = self.download_pic(filename=filename, url=url, mtime=\n date_local)\n if item.is_video and self.download_videos is True:\n downloaded |= self.download_pic(filename=filename, url=item.\n video_url, mtime=date_local)\n metadata_string = _ArbitraryItemFormatter(item).format(self.\n storyitem_metadata_txt_pattern).strip()\n if metadata_string:\n self.save_caption(filename=filename, mtime=item.date_local, caption\n =metadata_string)\n if self.save_metadata is not False:\n self.save_metadata_json(filename, item)\n self.context.log()\n return downloaded\n",
"role": "user"
},
{
"content": "download_storyitem",
"role": "assistant"
}
] |
def x(self, user: Union[int, Profile]) ->Iterator[Highlight]:
userid = user if isinstance(user, int) else user.userid
data = self.context.graphql_query('7c16654f22c819fb63d1183034a5162f', {
'user_id': userid, 'include_chaining': False, 'include_reel': False,
'include_suggested_users': False, 'include_logged_out_extras':
False, 'include_highlight_reels': True})['data']['user'][
'edge_highlight_reels']
if data is None:
raise BadResponseException('Bad highlights reel JSON.')
yield from (Highlight(self.context, edge['node'], user if isinstance(
user, Profile) else None) for edge in data['edges'])
| get_highlights | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, user: Union[int, Profile]) ->Iterator[Highlight]:\n userid = user if isinstance(user, int) else user.userid\n data = self.context.graphql_query('7c16654f22c819fb63d1183034a5162f', {\n 'user_id': userid, 'include_chaining': False, 'include_reel': False,\n 'include_suggested_users': False, 'include_logged_out_extras': \n False, 'include_highlight_reels': True})['data']['user'][\n 'edge_highlight_reels']\n if data is None:\n raise BadResponseException('Bad highlights reel JSON.')\n yield from (Highlight(self.context, edge['node'], user if isinstance(\n user, Profile) else None) for edge in data['edges'])\n",
"role": "user"
},
{
"content": "get_highlights",
"role": "assistant"
}
] |
def x(self, user: Union[int, Profile], fast_update: bool=False,
filename_target: Optional[str]=None, storyitem_filter: Optional[
Callable[[StoryItem], bool]]=None) ->None:
for user_highlight in self.get_highlights(user):
name = user_highlight.owner_username
self.context.log('Retrieving highlights "{}" from profile {}'.
format(user_highlight.title, name))
totalcount = user_highlight.itemcount
count = 1
for item in user_highlight.get_items():
if storyitem_filter is not None and not storyitem_filter(item):
self.context.log('<{} skipped>'.format(item), flush=True)
continue
self.context.log('[%3i/%3i] ' % (count, totalcount), end='',
flush=True)
count += 1
with self.context.error_catcher(
'Download highlights "{}" from user {}'.format(
user_highlight.title, name)):
downloaded = self.download_storyitem(item, filename_target if
filename_target else '{}/{}'.format(name,
user_highlight.title))
if fast_update and not downloaded:
break
| download_highlights | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, user: Union[int, Profile], fast_update: bool=False,\n filename_target: Optional[str]=None, storyitem_filter: Optional[\n Callable[[StoryItem], bool]]=None) ->None:\n for user_highlight in self.get_highlights(user):\n name = user_highlight.owner_username\n self.context.log('Retrieving highlights \"{}\" from profile {}'.\n format(user_highlight.title, name))\n totalcount = user_highlight.itemcount\n count = 1\n for item in user_highlight.get_items():\n if storyitem_filter is not None and not storyitem_filter(item):\n self.context.log('<{} skipped>'.format(item), flush=True)\n continue\n self.context.log('[%3i/%3i] ' % (count, totalcount), end='',\n flush=True)\n count += 1\n with self.context.error_catcher(\n 'Download highlights \"{}\" from user {}'.format(\n user_highlight.title, name)):\n downloaded = self.download_storyitem(item, filename_target if\n filename_target else '{}/{}'.format(name,\n user_highlight.title))\n if fast_update and not downloaded:\n break\n",
"role": "user"
},
{
"content": "download_highlights",
"role": "assistant"
}
] |
def x(self) ->Iterator[Post]:
data = self.context.graphql_query('d6f4427fbe92d846298cf93df0b937d3', {})[
'data']
while True:
feed = data['user']['edge_web_feed_timeline']
yield from (Post(self.context, edge['node']) for edge in feed[
'edges'] if not edge['node']['__typename'] ==
'GraphSuggestedUserFeedUnit')
if not feed['page_info']['has_next_page']:
break
data = self.context.graphql_query('d6f4427fbe92d846298cf93df0b937d3',
{'fetch_media_item_count': 12, 'fetch_media_item_cursor': feed[
'page_info']['end_cursor'], 'fetch_comment_count': 4,
'fetch_like': 10, 'has_stories': False})['data']
| get_feed_posts | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self) ->Iterator[Post]:\n data = self.context.graphql_query('d6f4427fbe92d846298cf93df0b937d3', {})[\n 'data']\n while True:\n feed = data['user']['edge_web_feed_timeline']\n yield from (Post(self.context, edge['node']) for edge in feed[\n 'edges'] if not edge['node']['__typename'] ==\n 'GraphSuggestedUserFeedUnit')\n if not feed['page_info']['has_next_page']:\n break\n data = self.context.graphql_query('d6f4427fbe92d846298cf93df0b937d3',\n {'fetch_media_item_count': 12, 'fetch_media_item_cursor': feed[\n 'page_info']['end_cursor'], 'fetch_comment_count': 4,\n 'fetch_like': 10, 'has_stories': False})['data']\n",
"role": "user"
},
{
"content": "get_feed_posts",
"role": "assistant"
}
] |
def x(self, max_count: int=None, fast_update: bool=False, post_filter:
Optional[Callable[[Post], bool]]=None) ->None:
self.context.log('Retrieving pictures from your feed...')
count = 1
for post in self.get_feed_posts():
if max_count is not None and count > max_count:
break
name = post.owner_username
if post_filter is not None and not post_filter(post):
self.context.log('<pic by %s skipped>' % name, flush=True)
continue
self.context.log('[%3i] %s ' % (count, name), end='', flush=True)
count += 1
with self.context.error_catcher('Download feed'):
downloaded = self.download_post(post, target=':feed')
if fast_update and not downloaded:
break
| download_feed_posts | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, max_count: int=None, fast_update: bool=False, post_filter:\n Optional[Callable[[Post], bool]]=None) ->None:\n self.context.log('Retrieving pictures from your feed...')\n count = 1\n for post in self.get_feed_posts():\n if max_count is not None and count > max_count:\n break\n name = post.owner_username\n if post_filter is not None and not post_filter(post):\n self.context.log('<pic by %s skipped>' % name, flush=True)\n continue\n self.context.log('[%3i] %s ' % (count, name), end='', flush=True)\n count += 1\n with self.context.error_catcher('Download feed'):\n downloaded = self.download_post(post, target=':feed')\n if fast_update and not downloaded:\n break\n",
"role": "user"
},
{
"content": "download_feed_posts",
"role": "assistant"
}
] |
def x(self, max_count: int=None, fast_update: bool=False, post_filter:
Optional[Callable[[Post], bool]]=None) ->None:
self.context.log('Retrieving saved posts...')
count = 1
for post in Profile.from_username(self.context, self.context.username
).get_saved_posts():
if max_count is not None and count > max_count:
break
if post_filter is not None and not post_filter(post):
self.context.log('<{} skipped>'.format(post), flush=True)
continue
self.context.log('[{:>3}] '.format(count), end=str(), flush=True)
count += 1
with self.context.error_catcher('Download saved posts'):
downloaded = self.download_post(post, target=':saved')
if fast_update and not downloaded:
break
| download_saved_posts | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, max_count: int=None, fast_update: bool=False, post_filter:\n Optional[Callable[[Post], bool]]=None) ->None:\n self.context.log('Retrieving saved posts...')\n count = 1\n for post in Profile.from_username(self.context, self.context.username\n ).get_saved_posts():\n if max_count is not None and count > max_count:\n break\n if post_filter is not None and not post_filter(post):\n self.context.log('<{} skipped>'.format(post), flush=True)\n continue\n self.context.log('[{:>3}] '.format(count), end=str(), flush=True)\n count += 1\n with self.context.error_catcher('Download saved posts'):\n downloaded = self.download_post(post, target=':saved')\n if fast_update and not downloaded:\n break\n",
"role": "user"
},
{
"content": "download_saved_posts",
"role": "assistant"
}
] |
def x(self, location: str) ->Iterator[Post]:
has_next_page = True
end_cursor = None
while has_next_page:
if end_cursor:
params = {'__a': 1, 'max_id': end_cursor}
else:
params = {'__a': 1}
location_data = self.context.get_json('explore/locations/{0}/'.
format(location), params)['graphql']['location'][
'edge_location_to_media']
yield from (Post(self.context, edge['node']) for edge in
location_data['edges'])
has_next_page = location_data['page_info']['has_next_page']
end_cursor = location_data['page_info']['end_cursor']
| get_location_posts | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, location: str) ->Iterator[Post]:\n has_next_page = True\n end_cursor = None\n while has_next_page:\n if end_cursor:\n params = {'__a': 1, 'max_id': end_cursor}\n else:\n params = {'__a': 1}\n location_data = self.context.get_json('explore/locations/{0}/'.\n format(location), params)['graphql']['location'][\n 'edge_location_to_media']\n yield from (Post(self.context, edge['node']) for edge in\n location_data['edges'])\n has_next_page = location_data['page_info']['has_next_page']\n end_cursor = location_data['page_info']['end_cursor']\n",
"role": "user"
},
{
"content": "get_location_posts",
"role": "assistant"
}
] |
def x(self, location: str, max_count: Optional[int]=None, post_filter:
Optional[Callable[[Post], bool]]=None, fast_update: bool=False) ->None:
self.context.log('Retrieving pictures for location {}...'.format(location))
count = 1
for post in self.get_location_posts(location):
if max_count is not None and count > max_count:
break
self.context.log('[{0:3d}] %{1} '.format(count, location), end='',
flush=True)
if post_filter is not None and not post_filter(post):
self.context.log('<skipped>')
continue
count += 1
with self.context.error_catcher('Download location {}'.format(location)
):
downloaded = self.download_post(post, target='%' + location)
if fast_update and not downloaded:
break
| download_location | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, location: str, max_count: Optional[int]=None, post_filter:\n Optional[Callable[[Post], bool]]=None, fast_update: bool=False) ->None:\n self.context.log('Retrieving pictures for location {}...'.format(location))\n count = 1\n for post in self.get_location_posts(location):\n if max_count is not None and count > max_count:\n break\n self.context.log('[{0:3d}] %{1} '.format(count, location), end='',\n flush=True)\n if post_filter is not None and not post_filter(post):\n self.context.log('<skipped>')\n continue\n count += 1\n with self.context.error_catcher('Download location {}'.format(location)\n ):\n downloaded = self.download_post(post, target='%' + location)\n if fast_update and not downloaded:\n break\n",
"role": "user"
},
{
"content": "download_location",
"role": "assistant"
}
] |
def x(self) ->Iterator[Post]:
data = self.context.get_json('explore/', {})
yield from (Post(self.context, node) for node in self.context.
graphql_node_list('df0dcc250c2b18d9fd27c5581ef33c7c', {},
'https://www.instagram.com/explore/', lambda d: d['data']['user'][
'edge_web_discover_media'], data['rhx_gis']))
| get_explore_posts | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self) ->Iterator[Post]:\n data = self.context.get_json('explore/', {})\n yield from (Post(self.context, node) for node in self.context.\n graphql_node_list('df0dcc250c2b18d9fd27c5581ef33c7c', {},\n 'https://www.instagram.com/explore/', lambda d: d['data']['user'][\n 'edge_web_discover_media'], data['rhx_gis']))\n",
"role": "user"
},
{
"content": "get_explore_posts",
"role": "assistant"
}
] |
def x(self, hashtag: str) ->Iterator[Post]:
has_next_page = True
end_cursor = None
while has_next_page:
if end_cursor:
params = {'__a': 1, 'max_id': end_cursor}
else:
params = {'__a': 1}
hashtag_data = self.context.get_json('explore/tags/{0}/'.format(
hashtag), params)['graphql']['hashtag']['edge_hashtag_to_media']
yield from (Post(self.context, edge['node']) for edge in
hashtag_data['edges'])
has_next_page = hashtag_data['page_info']['has_next_page']
end_cursor = hashtag_data['page_info']['end_cursor']
| get_hashtag_posts | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, hashtag: str) ->Iterator[Post]:\n has_next_page = True\n end_cursor = None\n while has_next_page:\n if end_cursor:\n params = {'__a': 1, 'max_id': end_cursor}\n else:\n params = {'__a': 1}\n hashtag_data = self.context.get_json('explore/tags/{0}/'.format(\n hashtag), params)['graphql']['hashtag']['edge_hashtag_to_media']\n yield from (Post(self.context, edge['node']) for edge in\n hashtag_data['edges'])\n has_next_page = hashtag_data['page_info']['has_next_page']\n end_cursor = hashtag_data['page_info']['end_cursor']\n",
"role": "user"
},
{
"content": "get_hashtag_posts",
"role": "assistant"
}
] |
def x(self, hashtag: str, max_count: Optional[int]=None, post_filter:
Optional[Callable[[Post], bool]]=None, fast_update: bool=False) ->None:
hashtag = hashtag.lower()
self.context.log('Retrieving pictures with hashtag {}...'.format(hashtag))
count = 1
for post in self.get_hashtag_posts(hashtag):
if max_count is not None and count > max_count:
break
self.context.log('[{0:3d}] #{1} '.format(count, hashtag), end='',
flush=True)
if post_filter is not None and not post_filter(post):
self.context.log('<skipped>')
continue
count += 1
with self.context.error_catcher('Download hashtag #{}'.format(hashtag)
):
downloaded = self.download_post(post, target='#' + hashtag)
if fast_update and not downloaded:
break
| download_hashtag | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, hashtag: str, max_count: Optional[int]=None, post_filter:\n Optional[Callable[[Post], bool]]=None, fast_update: bool=False) ->None:\n hashtag = hashtag.lower()\n self.context.log('Retrieving pictures with hashtag {}...'.format(hashtag))\n count = 1\n for post in self.get_hashtag_posts(hashtag):\n if max_count is not None and count > max_count:\n break\n self.context.log('[{0:3d}] #{1} '.format(count, hashtag), end='',\n flush=True)\n if post_filter is not None and not post_filter(post):\n self.context.log('<skipped>')\n continue\n count += 1\n with self.context.error_catcher('Download hashtag #{}'.format(hashtag)\n ):\n downloaded = self.download_post(post, target='#' + hashtag)\n if fast_update and not downloaded:\n break\n",
"role": "user"
},
{
"content": "download_hashtag",
"role": "assistant"
}
] |
def x(self, profile: Profile, fast_update: bool=False, target: Optional[str
]=None, post_filter: Optional[Callable[[Post], bool]]=None) ->None:
if target is None:
target = profile.username + '/:tagged'
self.context.log('Retrieving tagged posts for profile {}.'.format(
profile.username))
count = 1
for post in profile.get_tagged_posts():
self.context.log('[%3i/???] ' % count, end='', flush=True)
count += 1
if post_filter is not None and not post_filter(post):
self.context.log('<{} skipped>'.format(post))
with self.context.error_catcher('Download tagged {}'.format(profile
.username)):
downloaded = self.download_post(post, target)
if fast_update and not downloaded:
break
| download_tagged | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, profile: Profile, fast_update: bool=False, target: Optional[str\n ]=None, post_filter: Optional[Callable[[Post], bool]]=None) ->None:\n if target is None:\n target = profile.username + '/:tagged'\n self.context.log('Retrieving tagged posts for profile {}.'.format(\n profile.username))\n count = 1\n for post in profile.get_tagged_posts():\n self.context.log('[%3i/???] ' % count, end='', flush=True)\n count += 1\n if post_filter is not None and not post_filter(post):\n self.context.log('<{} skipped>'.format(post))\n with self.context.error_catcher('Download tagged {}'.format(profile\n .username)):\n downloaded = self.download_post(post, target)\n if fast_update and not downloaded:\n break\n",
"role": "user"
},
{
"content": "download_tagged",
"role": "assistant"
}
] |
def x(self, profile: Profile):
os.makedirs(self.dirname_pattern.format(profile=profile.username,
target=profile.username), exist_ok=True)
with open(self._get_id_filename(profile.username), 'w') as text_file:
text_file.write(str(profile.userid) + '\n')
self.context.log('Stored ID {0} for profile {1}.'.format(profile.
userid, profile.username))
| save_profile_id | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, profile: Profile):\n os.makedirs(self.dirname_pattern.format(profile=profile.username,\n target=profile.username), exist_ok=True)\n with open(self._get_id_filename(profile.username), 'w') as text_file:\n text_file.write(str(profile.userid) + '\\n')\n self.context.log('Stored ID {0} for profile {1}.'.format(profile.\n userid, profile.username))\n",
"role": "user"
},
{
"content": "save_profile_id",
"role": "assistant"
}
] |
def x(self, profile_name: str) ->Profile:
profile = None
with suppress(ProfileNotExistsException):
profile = Profile.from_username(self.context, profile_name)
profile_exists = profile is not None
id_filename = self._get_id_filename(profile_name)
try:
with open(id_filename, 'rb') as id_file:
profile_id = int(id_file.read())
if not profile_exists or profile_id != profile.userid:
if profile_exists:
self.context.log(
'Profile {0} does not match the stored unique ID {1}.'.
format(profile_name, profile_id))
else:
self.context.log(
'Trying to find profile {0} using its unique ID {1}.'.
format(profile_name, profile_id))
profile_from_id = Profile.from_id(self.context, profile_id)
newname = profile_from_id.username
self.context.log('Profile {0} has changed its name to {1}.'.
format(profile_name, newname))
if format_string_contains_key(self.dirname_pattern, 'profile'
) or format_string_contains_key(self.dirname_pattern, 'target'
):
os.rename(self.dirname_pattern.format(profile=profile_name.
lower(), target=profile_name.lower()), self.
dirname_pattern.format(profile=newname.lower(), target=
newname.lower()))
else:
os.rename('{0}/{1}_id'.format(self.dirname_pattern.format(),
profile_name.lower()), '{0}/{1}_id'.format(self.
dirname_pattern.format(), newname.lower()))
return profile_from_id
return profile
except (FileNotFoundError, ValueError):
pass
if profile_exists:
self.save_profile_id(profile)
return profile
raise ProfileNotExistsException('Profile {0} does not exist.'.format(
profile_name))
| check_profile_id | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, profile_name: str) ->Profile:\n profile = None\n with suppress(ProfileNotExistsException):\n profile = Profile.from_username(self.context, profile_name)\n profile_exists = profile is not None\n id_filename = self._get_id_filename(profile_name)\n try:\n with open(id_filename, 'rb') as id_file:\n profile_id = int(id_file.read())\n if not profile_exists or profile_id != profile.userid:\n if profile_exists:\n self.context.log(\n 'Profile {0} does not match the stored unique ID {1}.'.\n format(profile_name, profile_id))\n else:\n self.context.log(\n 'Trying to find profile {0} using its unique ID {1}.'.\n format(profile_name, profile_id))\n profile_from_id = Profile.from_id(self.context, profile_id)\n newname = profile_from_id.username\n self.context.log('Profile {0} has changed its name to {1}.'.\n format(profile_name, newname))\n if format_string_contains_key(self.dirname_pattern, 'profile'\n ) or format_string_contains_key(self.dirname_pattern, 'target'\n ):\n os.rename(self.dirname_pattern.format(profile=profile_name.\n lower(), target=profile_name.lower()), self.\n dirname_pattern.format(profile=newname.lower(), target=\n newname.lower()))\n else:\n os.rename('{0}/{1}_id'.format(self.dirname_pattern.format(),\n profile_name.lower()), '{0}/{1}_id'.format(self.\n dirname_pattern.format(), newname.lower()))\n return profile_from_id\n return profile\n except (FileNotFoundError, ValueError):\n pass\n if profile_exists:\n self.save_profile_id(profile)\n return profile\n raise ProfileNotExistsException('Profile {0} does not exist.'.format(\n profile_name))\n",
"role": "user"
},
{
"content": "check_profile_id",
"role": "assistant"
}
] |
def x(self, profiles: Set[Profile], profile_pic: bool=True, posts: bool=
True, tagged: bool=False, highlights: bool=False, stories: bool=False,
fast_update: bool=False, post_filter: Optional[Callable[[Post], bool]]=
None, storyitem_filter: Optional[Callable[[Post], bool]]=None,
raise_errors: bool=False):
def x(_str):
yield
error_handler = (_error_raiser if raise_errors else self.context.
error_catcher)
for profile in profiles:
with error_handler(profile.username):
profile_name = profile.username
if profile_pic:
with self.context.error_catcher(
'Download profile picture of {}'.format(profile_name)):
self.download_profilepic(profile)
if self.save_metadata:
json_filename = '{0}/{1}_{2}'.format(self.dirname_pattern.
format(profile=profile_name, target=profile_name),
profile_name, profile.userid)
self.save_metadata_json(json_filename, profile)
if profile.is_private and (tagged or highlights or posts):
if not self.context.is_logged_in:
raise LoginRequiredException('--login=USERNAME required.')
if (not profile.followed_by_viewer and self.context.
username != profile.username):
raise PrivateProfileNotFollowedException(
'Private but not followed.')
if tagged:
with self.context.error_catcher('Download tagged of {}'.
format(profile_name)):
self.download_tagged(profile, fast_update=fast_update,
post_filter=post_filter)
if highlights:
with self.context.error_catcher('Download highlights of {}'
.format(profile_name)):
self.download_highlights(profile, fast_update=
fast_update, storyitem_filter=storyitem_filter)
if posts:
self.context.log('Retrieving posts from profile {}.'.format
(profile_name))
totalcount = profile.mediacount
count = 1
for post in profile.get_posts():
self.context.log('[%3i/%3i] ' % (count, totalcount),
end='', flush=True)
count += 1
if post_filter is not None and not post_filter(post):
self.context.log('<skipped>')
continue
with self.context.error_catcher('Download {} of {}'.
format(post, profile_name)):
post_changed = False
while True:
try:
downloaded = self.download_post(post,
target=profile_name)
break
except PostChangedException:
post_changed = True
continue
if fast_update and not downloaded and not post_changed:
break
if stories and profiles:
with self.context.error_catcher('Download stories'):
self.context.log('Downloading stories')
self.download_stories(userids=list(profiles), fast_update=
fast_update, filename_target=None, storyitem_filter=
storyitem_filter)
| download_profiles | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, profiles: Set[Profile], profile_pic: bool=True, posts: bool=\n True, tagged: bool=False, highlights: bool=False, stories: bool=False,\n fast_update: bool=False, post_filter: Optional[Callable[[Post], bool]]=\n None, storyitem_filter: Optional[Callable[[Post], bool]]=None,\n raise_errors: bool=False):\n\n def x(_str):\n yield\n error_handler = (_error_raiser if raise_errors else self.context.\n error_catcher)\n for profile in profiles:\n with error_handler(profile.username):\n profile_name = profile.username\n if profile_pic:\n with self.context.error_catcher(\n 'Download profile picture of {}'.format(profile_name)):\n self.download_profilepic(profile)\n if self.save_metadata:\n json_filename = '{0}/{1}_{2}'.format(self.dirname_pattern.\n format(profile=profile_name, target=profile_name),\n profile_name, profile.userid)\n self.save_metadata_json(json_filename, profile)\n if profile.is_private and (tagged or highlights or posts):\n if not self.context.is_logged_in:\n raise LoginRequiredException('--login=USERNAME required.')\n if (not profile.followed_by_viewer and self.context.\n username != profile.username):\n raise PrivateProfileNotFollowedException(\n 'Private but not followed.')\n if tagged:\n with self.context.error_catcher('Download tagged of {}'.\n format(profile_name)):\n self.download_tagged(profile, fast_update=fast_update,\n post_filter=post_filter)\n if highlights:\n with self.context.error_catcher('Download highlights of {}'\n .format(profile_name)):\n self.download_highlights(profile, fast_update=\n fast_update, storyitem_filter=storyitem_filter)\n if posts:\n self.context.log('Retrieving posts from profile {}.'.format\n (profile_name))\n totalcount = profile.mediacount\n count = 1\n for post in profile.get_posts():\n self.context.log('[%3i/%3i] ' % (count, totalcount),\n end='', flush=True)\n count += 1\n if post_filter is not None and not post_filter(post):\n self.context.log('<skipped>')\n continue\n with self.context.error_catcher('Download {} of {}'.\n format(post, profile_name)):\n post_changed = False\n while True:\n try:\n downloaded = self.download_post(post,\n target=profile_name)\n break\n except PostChangedException:\n post_changed = True\n continue\n if fast_update and not downloaded and not post_changed:\n break\n if stories and profiles:\n with self.context.error_catcher('Download stories'):\n self.context.log('Downloading stories')\n self.download_stories(userids=list(profiles), fast_update=\n fast_update, filename_target=None, storyitem_filter=\n storyitem_filter)\n",
"role": "user"
},
{
"content": "download_profiles",
"role": "assistant"
}
] |
def x(self, profile_name: Union[str, Profile], profile_pic: bool=True,
profile_pic_only: bool=False, fast_update: bool=False, download_stories:
bool=False, download_stories_only: bool=False, download_tagged: bool=
False, download_tagged_only: bool=False, post_filter: Optional[Callable
[[Post], bool]]=None, storyitem_filter: Optional[Callable[[StoryItem],
bool]]=None) ->None:
if isinstance(profile_name, str):
profile = self.check_profile_id(profile_name.lower())
else:
profile = profile_name
profile_name = profile.username
if self.save_metadata is not False:
json_filename = '{0}/{1}_{2}'.format(self.dirname_pattern.format(
profile=profile_name, target=profile_name), profile_name,
profile.userid)
self.save_metadata_json(json_filename, profile)
if (self.context.is_logged_in and profile.has_blocked_viewer and not
profile.is_private):
raise ProfileNotExistsException('Profile {} has blocked you'.format
(profile_name))
if profile_pic or profile_pic_only:
with self.context.error_catcher('Download profile picture of {}'.
format(profile_name)):
self.download_profilepic(profile)
if profile_pic_only:
return
if profile.is_private:
if not self.context.is_logged_in:
raise LoginRequiredException('profile %s requires login' %
profile_name)
if (not profile.followed_by_viewer and self.context.username !=
profile.username):
raise PrivateProfileNotFollowedException(
'Profile %s: private but not followed.' % profile_name)
elif self.context.is_logged_in and not (download_stories or
download_stories_only):
self.context.log('profile %s could also be downloaded anonymously.' %
profile_name)
if download_stories or download_stories_only:
if profile.has_viewable_story:
with self.context.error_catcher('Download stories of {}'.format
(profile_name)):
self.download_stories(userids=[profile.userid],
filename_target=profile_name, fast_update=fast_update,
storyitem_filter=storyitem_filter)
else:
self.context.log('{} does not have any stories.'.format(
profile_name))
if download_stories_only:
return
if download_tagged or download_tagged_only:
with self.context.error_catcher('Download tagged of {}'.format(
profile_name)):
self.download_tagged(profile, fast_update=fast_update,
post_filter=post_filter)
if download_tagged_only:
return
self.context.log('Retrieving posts from profile {}.'.format(profile_name))
totalcount = profile.mediacount
count = 1
for post in profile.get_posts():
self.context.log('[%3i/%3i] ' % (count, totalcount), end='', flush=True
)
count += 1
if post_filter is not None and not post_filter(post):
self.context.log('<skipped>')
continue
with self.context.error_catcher('Download profile {}'.format(
profile_name)):
downloaded = self.download_post(post, target=profile_name)
if fast_update and not downloaded:
break
| download_profile | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, profile_name: Union[str, Profile], profile_pic: bool=True,\n profile_pic_only: bool=False, fast_update: bool=False, download_stories:\n bool=False, download_stories_only: bool=False, download_tagged: bool=\n False, download_tagged_only: bool=False, post_filter: Optional[Callable\n [[Post], bool]]=None, storyitem_filter: Optional[Callable[[StoryItem],\n bool]]=None) ->None:\n if isinstance(profile_name, str):\n profile = self.check_profile_id(profile_name.lower())\n else:\n profile = profile_name\n profile_name = profile.username\n if self.save_metadata is not False:\n json_filename = '{0}/{1}_{2}'.format(self.dirname_pattern.format(\n profile=profile_name, target=profile_name), profile_name,\n profile.userid)\n self.save_metadata_json(json_filename, profile)\n if (self.context.is_logged_in and profile.has_blocked_viewer and not\n profile.is_private):\n raise ProfileNotExistsException('Profile {} has blocked you'.format\n (profile_name))\n if profile_pic or profile_pic_only:\n with self.context.error_catcher('Download profile picture of {}'.\n format(profile_name)):\n self.download_profilepic(profile)\n if profile_pic_only:\n return\n if profile.is_private:\n if not self.context.is_logged_in:\n raise LoginRequiredException('profile %s requires login' %\n profile_name)\n if (not profile.followed_by_viewer and self.context.username !=\n profile.username):\n raise PrivateProfileNotFollowedException(\n 'Profile %s: private but not followed.' % profile_name)\n elif self.context.is_logged_in and not (download_stories or\n download_stories_only):\n self.context.log('profile %s could also be downloaded anonymously.' %\n profile_name)\n if download_stories or download_stories_only:\n if profile.has_viewable_story:\n with self.context.error_catcher('Download stories of {}'.format\n (profile_name)):\n self.download_stories(userids=[profile.userid],\n filename_target=profile_name, fast_update=fast_update,\n storyitem_filter=storyitem_filter)\n else:\n self.context.log('{} does not have any stories.'.format(\n profile_name))\n if download_stories_only:\n return\n if download_tagged or download_tagged_only:\n with self.context.error_catcher('Download tagged of {}'.format(\n profile_name)):\n self.download_tagged(profile, fast_update=fast_update,\n post_filter=post_filter)\n if download_tagged_only:\n return\n self.context.log('Retrieving posts from profile {}.'.format(profile_name))\n totalcount = profile.mediacount\n count = 1\n for post in profile.get_posts():\n self.context.log('[%3i/%3i] ' % (count, totalcount), end='', flush=True\n )\n count += 1\n if post_filter is not None and not post_filter(post):\n self.context.log('<skipped>')\n continue\n with self.context.error_catcher('Download profile {}'.format(\n profile_name)):\n downloaded = self.download_post(post, target=profile_name)\n if fast_update and not downloaded:\n break\n",
"role": "user"
},
{
"content": "download_profile",
"role": "assistant"
}
] |
def x(self, filename: str) ->bool:
if os.path.isfile(filename + '.json.xz'):
filename += '.json.xz'
elif os.path.isfile(filename + '.json'):
filename += '.json'
else:
return False
try:
load_structure_from_file(self.context, filename)
return True
except (FileNotFoundError, lzma.LZMAError, json.decoder.JSONDecodeError):
return False
| check_if_committed | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, filename: str) ->bool:\n if os.path.isfile(filename + '.json.xz'):\n filename += '.json.xz'\n elif os.path.isfile(filename + '.json'):\n filename += '.json'\n else:\n return False\n try:\n load_structure_from_file(self.context, filename)\n return True\n except (FileNotFoundError, lzma.LZMAError, json.decoder.JSONDecodeError):\n return False\n",
"role": "user"
},
{
"content": "check_if_committed",
"role": "assistant"
}
] |
def x(self, username: str) ->None:
if self.context.quiet:
raise LoginRequiredException(
'Quiet mode requires given password or valid session file.')
try:
password = None
while password is None:
password = getpass.getpass(prompt=
'Enter Instagram password for %s: ' % username)
try:
self.login(username, password)
except BadCredentialsException as err:
print(err, file=sys.stderr)
password = None
except TwoFactorAuthRequiredException:
while True:
try:
code = input('Enter 2FA verification code: ')
self.two_factor_login(code)
break
except BadCredentialsException:
pass
| interactive_login | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, username: str) ->None:\n if self.context.quiet:\n raise LoginRequiredException(\n 'Quiet mode requires given password or valid session file.')\n try:\n password = None\n while password is None:\n password = getpass.getpass(prompt=\n 'Enter Instagram password for %s: ' % username)\n try:\n self.login(username, password)\n except BadCredentialsException as err:\n print(err, file=sys.stderr)\n password = None\n except TwoFactorAuthRequiredException:\n while True:\n try:\n code = input('Enter 2FA verification code: ')\n self.two_factor_login(code)\n break\n except BadCredentialsException:\n pass\n",
"role": "user"
},
{
"content": "interactive_login",
"role": "assistant"
}
] |
def x(filter_str: str, item_type: type):
class TransformFilterAst(ast.NodeTransformer):
def x(self, node: ast.Name):
if not isinstance(node.ctx, ast.Load):
raise InvalidArgumentException(
'Invalid filter: Modifying variables ({}) not allowed.'
.format(node.id))
if node.id == 'datetime':
return node
if not hasattr(item_type, node.id):
raise InvalidArgumentException(
'Invalid filter: {} not a {} attribute.'.format(node.id,
item_type.__name__))
new_node = ast.Attribute(ast.copy_location(ast.Name('item', ast
.Load()), node), node.id, ast.copy_location(ast.Load(), node))
return ast.copy_location(new_node, node)
input_filename = '<command line filter parameter>'
compiled_filter = compile(TransformFilterAst().visit(ast.parse(
filter_str, filename=input_filename, mode='eval')), filename=
input_filename, mode='eval')
def x(item) ->bool:
return bool(eval(compiled_filter, {'item': item, 'datetime':
datetime.datetime}))
return filterfunc
| filterstr_to_filterfunc | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(filter_str: str, item_type: type):\n\n\n class TransformFilterAst(ast.NodeTransformer):\n\n def x(self, node: ast.Name):\n if not isinstance(node.ctx, ast.Load):\n raise InvalidArgumentException(\n 'Invalid filter: Modifying variables ({}) not allowed.'\n .format(node.id))\n if node.id == 'datetime':\n return node\n if not hasattr(item_type, node.id):\n raise InvalidArgumentException(\n 'Invalid filter: {} not a {} attribute.'.format(node.id,\n item_type.__name__))\n new_node = ast.Attribute(ast.copy_location(ast.Name('item', ast\n .Load()), node), node.id, ast.copy_location(ast.Load(), node))\n return ast.copy_location(new_node, node)\n input_filename = '<command line filter parameter>'\n compiled_filter = compile(TransformFilterAst().visit(ast.parse(\n filter_str, filename=input_filename, mode='eval')), filename=\n input_filename, mode='eval')\n\n def x(item) ->bool:\n return bool(eval(compiled_filter, {'item': item, 'datetime':\n datetime.datetime}))\n return filterfunc\n",
"role": "user"
},
{
"content": "filterstr_to_filterfunc",
"role": "assistant"
}
] |
def x(t, alpha=1, beta=1):
return beta * (1 - np.exp(-alpha * np.maximum(0, t - 10 * beta)))
| exp_comp_ | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(t, alpha=1, beta=1):\n return beta * (1 - np.exp(-alpha * np.maximum(0, t - 10 * beta)))\n",
"role": "user"
},
{
"content": "exp_comp_",
"role": "assistant"
}
] |
def x(t, alpha=1, beta=1):
return beta * np.log(alpha * np.maximum(0, t - 10 * beta) + 1)
| log_ | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(t, alpha=1, beta=1):\n return beta * np.log(alpha * np.maximum(0, t - 10 * beta) + 1)\n",
"role": "user"
},
{
"content": "log_",
"role": "assistant"
}
] |
def x(lifetimes, max_, min_=0):
n = lifetimes.shape[0]
u = min_ + (max_ - min_) * random.rand(n)
observations = np.minimum(u, lifetimes)
return observations, lifetimes == observations
| right_censor_lifetimes | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(lifetimes, max_, min_=0):\n n = lifetimes.shape[0]\n u = min_ + (max_ - min_) * random.rand(n)\n observations = np.minimum(u, lifetimes)\n return observations, lifetimes == observations\n",
"role": "user"
},
{
"content": "right_censor_lifetimes",
"role": "assistant"
}
] |
def x(n, d, n_binary=0, p=0.5):
assert n_binary >= 0 and n_binary <= d, 'binary must be between 0 and d'
covariates = np.zeros((n, d + 1))
covariates[:, :d - n_binary] = random.exponential(1, size=(n, d - n_binary)
)
covariates[:, d - n_binary:-1] = random.binomial(1, p, size=(n, n_binary))
covariates[:, -1] = np.ones(n)
return covariates
| generate_covariates | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(n, d, n_binary=0, p=0.5):\n assert n_binary >= 0 and n_binary <= d, 'binary must be between 0 and d'\n covariates = np.zeros((n, d + 1))\n covariates[:, :d - n_binary] = random.exponential(1, size=(n, d - n_binary)\n )\n covariates[:, d - n_binary:-1] = random.binomial(1, p, size=(n, n_binary))\n covariates[:, -1] = np.ones(n)\n return covariates\n",
"role": "user"
},
{
"content": "generate_covariates",
"role": "assistant"
}
] |
def x(d, timelines, constant=True, independent=0):
return time_varying_coefficients(d, timelines, constant, independent=
independent, randgen=random.normal)
| constant_coefficients | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(d, timelines, constant=True, independent=0):\n return time_varying_coefficients(d, timelines, constant, independent=\n independent, randgen=random.normal)\n",
"role": "user"
},
{
"content": "constant_coefficients",
"role": "assistant"
}
] |
def x(d, timelines, constant=False, independent=0, randgen=random.exponential):
t = timelines.shape[0]
try:
a = np.arange(d)
random.shuffle(a)
independent = a[:independent]
except IndexError:
pass
n_funcs = len(FUNCS)
coefficients = np.zeros((t, d))
data_generators = []
for i in range(d):
f = FUNCS[random.randint(0, n_funcs)] if not constant else constant_
if i in independent:
beta = 0
else:
beta = randgen((1 - constant) * 0.5 / d)
coefficients[:, i] = f(timelines, alpha=randgen(2000.0 / t), beta=beta)
data_generators.append(f.__doc__)
df_coefficients = pd.DataFrame(coefficients, columns=data_generators,
index=timelines)
return df_coefficients
| time_varying_coefficients | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(d, timelines, constant=False, independent=0, randgen=random.exponential):\n t = timelines.shape[0]\n try:\n a = np.arange(d)\n random.shuffle(a)\n independent = a[:independent]\n except IndexError:\n pass\n n_funcs = len(FUNCS)\n coefficients = np.zeros((t, d))\n data_generators = []\n for i in range(d):\n f = FUNCS[random.randint(0, n_funcs)] if not constant else constant_\n if i in independent:\n beta = 0\n else:\n beta = randgen((1 - constant) * 0.5 / d)\n coefficients[:, i] = f(timelines, alpha=randgen(2000.0 / t), beta=beta)\n data_generators.append(f.__doc__)\n df_coefficients = pd.DataFrame(coefficients, columns=data_generators,\n index=timelines)\n return df_coefficients\n",
"role": "user"
},
{
"content": "time_varying_coefficients",
"role": "assistant"
}
] |
def x(n, d, timelines, constant=False, independent=0, n_binary=0, model='aalen'
):
covariates = generate_covariates(n, d, n_binary=n_binary)
if model == 'aalen':
coefficients = time_varying_coefficients(d + 1, timelines,
independent=independent, constant=constant)
hazard_rates = np.dot(covariates, coefficients.T)
return pd.DataFrame(hazard_rates.T, index=timelines
), coefficients, pd.DataFrame(covariates)
if model == 'cox':
covariates = covariates[:, :-1]
coefficients = constant_coefficients(d, timelines, independent)
baseline = time_varying_coefficients(1, timelines)
hazard_rates = np.exp(np.dot(covariates, coefficients.T)) * baseline[
baseline.columns[0]].values
coefficients['baseline: ' + baseline.columns[0]] = baseline.values
return pd.DataFrame(hazard_rates.T, index=timelines
), coefficients, pd.DataFrame(covariates)
raise Exception
| generate_hazard_rates | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(n, d, timelines, constant=False, independent=0, n_binary=0, model='aalen'\n ):\n covariates = generate_covariates(n, d, n_binary=n_binary)\n if model == 'aalen':\n coefficients = time_varying_coefficients(d + 1, timelines,\n independent=independent, constant=constant)\n hazard_rates = np.dot(covariates, coefficients.T)\n return pd.DataFrame(hazard_rates.T, index=timelines\n ), coefficients, pd.DataFrame(covariates)\n if model == 'cox':\n covariates = covariates[:, :-1]\n coefficients = constant_coefficients(d, timelines, independent)\n baseline = time_varying_coefficients(1, timelines)\n hazard_rates = np.exp(np.dot(covariates, coefficients.T)) * baseline[\n baseline.columns[0]].values\n coefficients['baseline: ' + baseline.columns[0]] = baseline.values\n return pd.DataFrame(hazard_rates.T, index=timelines\n ), coefficients, pd.DataFrame(covariates)\n raise Exception\n",
"role": "user"
},
{
"content": "generate_hazard_rates",
"role": "assistant"
}
] |
def x(hazard_rates, timelines, size=1, censor=None):
n = hazard_rates.shape[1]
survival_times = np.empty((n, size))
cumulative_hazards = cumulative_integral(hazard_rates.values, timelines).T
for i in range(size):
u = random.rand(n, 1)
e = -np.log(u)
v = e - cumulative_hazards < 0
cross = v.argmax(1)
survival_times[:, i] = timelines[cross]
survival_times[cross == 0, i] = np.inf
if censor is not None:
if isinstance(censor, bool):
T = timelines.max()
rv = T * random.uniform(size=survival_times.shape)
else:
rv = censor
observed = np.less_equal(survival_times, rv)
survival_times = np.minimum(rv, survival_times)
return survival_times.T, observed.T
else:
return survival_times
| generate_random_lifetimes | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(hazard_rates, timelines, size=1, censor=None):\n n = hazard_rates.shape[1]\n survival_times = np.empty((n, size))\n cumulative_hazards = cumulative_integral(hazard_rates.values, timelines).T\n for i in range(size):\n u = random.rand(n, 1)\n e = -np.log(u)\n v = e - cumulative_hazards < 0\n cross = v.argmax(1)\n survival_times[:, i] = timelines[cross]\n survival_times[cross == 0, i] = np.inf\n if censor is not None:\n if isinstance(censor, bool):\n T = timelines.max()\n rv = T * random.uniform(size=survival_times.shape)\n else:\n rv = censor\n observed = np.less_equal(survival_times, rv)\n survival_times = np.minimum(rv, survival_times)\n return survival_times.T, observed.T\n else:\n return survival_times\n",
"role": "user"
},
{
"content": "generate_random_lifetimes",
"role": "assistant"
}
] |
def x(hazard_rates, timelines):
cumulative_hazards = cumulative_integral(hazard_rates.values, timelines)
return pd.DataFrame(np.exp(-cumulative_hazards), index=timelines)
| construct_survival_curves | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(hazard_rates, timelines):\n cumulative_hazards = cumulative_integral(hazard_rates.values, timelines)\n return pd.DataFrame(np.exp(-cumulative_hazards), index=timelines)\n",
"role": "user"
},
{
"content": "construct_survival_curves",
"role": "assistant"
}
] |
def x(self, df, duration_col, event_col=None, weights_col=None,
show_progress=False):
self._time_fit_was_called = datetime.utcnow().strftime('%Y-%m-%d %H:%M:%S'
) + ' UTC'
self._censoring_type = CensoringType.RIGHT
df = df.copy()
self.duration_col = duration_col
self.event_col = event_col
self.weights_col = weights_col
self._n_examples = df.shape[0]
X, T, E, weights = self._preprocess_dataframe(df)
self.durations = T.copy()
self.event_observed = E.copy()
self.weights = weights.copy()
self._norm_std = X.std(0)
if self.fit_intercept:
self._norm_std['_intercept'] = 1.0
else:
self._norm_std[self._norm_std < 1e-08] = 1.0
self.hazards_, self.cumulative_hazards_, self.cumulative_variance_ = (self
._fit_model(normalize(X, 0, self._norm_std), T, E, weights,
show_progress))
self.hazards_ /= self._norm_std
self.cumulative_hazards_ /= self._norm_std
self.cumulative_variance_ /= self._norm_std
self.confidence_intervals_ = self._compute_confidence_intervals()
self._index = self.hazards_.index
self._predicted_hazards_ = self.predict_cumulative_hazard(X).iloc[-1
].values.ravel()
return self
| fit | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, df, duration_col, event_col=None, weights_col=None,\n show_progress=False):\n self._time_fit_was_called = datetime.utcnow().strftime('%Y-%m-%d %H:%M:%S'\n ) + ' UTC'\n self._censoring_type = CensoringType.RIGHT\n df = df.copy()\n self.duration_col = duration_col\n self.event_col = event_col\n self.weights_col = weights_col\n self._n_examples = df.shape[0]\n X, T, E, weights = self._preprocess_dataframe(df)\n self.durations = T.copy()\n self.event_observed = E.copy()\n self.weights = weights.copy()\n self._norm_std = X.std(0)\n if self.fit_intercept:\n self._norm_std['_intercept'] = 1.0\n else:\n self._norm_std[self._norm_std < 1e-08] = 1.0\n self.hazards_, self.cumulative_hazards_, self.cumulative_variance_ = (self\n ._fit_model(normalize(X, 0, self._norm_std), T, E, weights,\n show_progress))\n self.hazards_ /= self._norm_std\n self.cumulative_hazards_ /= self._norm_std\n self.cumulative_variance_ /= self._norm_std\n self.confidence_intervals_ = self._compute_confidence_intervals()\n self._index = self.hazards_.index\n self._predicted_hazards_ = self.predict_cumulative_hazard(X).iloc[-1\n ].values.ravel()\n return self\n",
"role": "user"
},
{
"content": "fit",
"role": "assistant"
}
] |
def x(self, X):
n, _ = X.shape
cols = _get_index(X)
if isinstance(X, pd.DataFrame):
order = self.cumulative_hazards_.columns
order = order.drop('_intercept') if self.fit_intercept else order
X_ = X[order].values
else:
X_ = X
X_ = X_ if not self.fit_intercept else np.c_[X_, np.ones((n, 1))]
timeline = self._index
individual_cumulative_hazards_ = pd.DataFrame(np.dot(self.
cumulative_hazards_, X_.T), index=timeline, columns=cols)
return individual_cumulative_hazards_
| predict_cumulative_hazard | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, X):\n n, _ = X.shape\n cols = _get_index(X)\n if isinstance(X, pd.DataFrame):\n order = self.cumulative_hazards_.columns\n order = order.drop('_intercept') if self.fit_intercept else order\n X_ = X[order].values\n else:\n X_ = X\n X_ = X_ if not self.fit_intercept else np.c_[X_, np.ones((n, 1))]\n timeline = self._index\n individual_cumulative_hazards_ = pd.DataFrame(np.dot(self.\n cumulative_hazards_, X_.T), index=timeline, columns=cols)\n return individual_cumulative_hazards_\n",
"role": "user"
},
{
"content": "predict_cumulative_hazard",
"role": "assistant"
}
] |
def x(self, X):
index = _get_index(X)
t = self._index
return pd.DataFrame(trapz(self.predict_survival_function(X)[index].
values.T, t), index=index)
| predict_expectation | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, X):\n index = _get_index(X)\n t = self._index\n return pd.DataFrame(trapz(self.predict_survival_function(X)[index].\n values.T, t), index=index)\n",
"role": "user"
},
{
"content": "predict_expectation",
"role": "assistant"
}
] |
def x(self, columns=None, loc=None, iloc=None, **kwargs):
from matplotlib import pyplot as plt
assert loc is None or iloc is None, 'Cannot set both loc and iloc in call to .plot'
def x(ax, x, y, y_upper, y_lower, **kwargs):
base_line, = ax.plot(x, y, drawstyle='steps-post', **kwargs)
ax.fill_between(x, y_lower, y2=y_upper, alpha=0.25, color=base_line
.get_color(), linewidth=1.0, step='post')
def x(loc, iloc):
get_method = 'loc' if loc is not None else 'iloc'
if iloc is None and loc is None:
user_submitted_ix = slice(0, None)
else:
user_submitted_ix = loc if loc is not None else iloc
return lambda df: getattr(df, get_method)[user_submitted_ix]
subset_df = create_df_slicer(loc, iloc)
if not columns:
columns = self.cumulative_hazards_.columns
else:
columns = _to_list(columns)
set_kwargs_ax(kwargs)
ax = kwargs.pop('ax')
x = subset_df(self.cumulative_hazards_).index.values.astype(float)
for column in columns:
y = subset_df(self.cumulative_hazards_[column]).values
index = subset_df(self.cumulative_hazards_[column]).index
y_upper = subset_df(self.confidence_intervals_[column].loc[
'upper-bound']).values
y_lower = subset_df(self.confidence_intervals_[column].loc[
'lower-bound']).values
shaded_plot(ax, x, y, y_upper, y_lower, label=column, **kwargs)
plt.hlines(0, index.min() - 1, index.max(), color='k', linestyles='--',
alpha=0.5)
ax.legend()
return ax
| plot | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, columns=None, loc=None, iloc=None, **kwargs):\n from matplotlib import pyplot as plt\n assert loc is None or iloc is None, 'Cannot set both loc and iloc in call to .plot'\n\n def x(ax, x, y, y_upper, y_lower, **kwargs):\n base_line, = ax.plot(x, y, drawstyle='steps-post', **kwargs)\n ax.fill_between(x, y_lower, y2=y_upper, alpha=0.25, color=base_line\n .get_color(), linewidth=1.0, step='post')\n\n def x(loc, iloc):\n get_method = 'loc' if loc is not None else 'iloc'\n if iloc is None and loc is None:\n user_submitted_ix = slice(0, None)\n else:\n user_submitted_ix = loc if loc is not None else iloc\n return lambda df: getattr(df, get_method)[user_submitted_ix]\n subset_df = create_df_slicer(loc, iloc)\n if not columns:\n columns = self.cumulative_hazards_.columns\n else:\n columns = _to_list(columns)\n set_kwargs_ax(kwargs)\n ax = kwargs.pop('ax')\n x = subset_df(self.cumulative_hazards_).index.values.astype(float)\n for column in columns:\n y = subset_df(self.cumulative_hazards_[column]).values\n index = subset_df(self.cumulative_hazards_[column]).index\n y_upper = subset_df(self.confidence_intervals_[column].loc[\n 'upper-bound']).values\n y_lower = subset_df(self.confidence_intervals_[column].loc[\n 'lower-bound']).values\n shaded_plot(ax, x, y, y_upper, y_lower, label=column, **kwargs)\n plt.hlines(0, index.min() - 1, index.max(), color='k', linestyles='--',\n alpha=0.5)\n ax.legend()\n return ax\n",
"role": "user"
},
{
"content": "plot",
"role": "assistant"
}
] |
def x(self, bandwidth=1):
timeline = self._index.values
return pd.DataFrame(np.dot(epanechnikov_kernel(timeline[:, None],
timeline, bandwidth), self.hazards_.values), columns=self.hazards_.
columns, index=timeline)
| smoothed_hazards_ | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, bandwidth=1):\n timeline = self._index.values\n return pd.DataFrame(np.dot(epanechnikov_kernel(timeline[:, None],\n timeline, bandwidth), self.hazards_.values), columns=self.hazards_.\n columns, index=timeline)\n",
"role": "user"
},
{
"content": "smoothed_hazards_",
"role": "assistant"
}
] |
def x(self):
if hasattr(self, '_predicted_hazards_'):
self._concordance_score_ = concordance_index(self.durations, -self.
_predicted_hazards_, self.event_observed)
del self._predicted_hazards_
return self._concordance_score_
return self._concordance_score_
| score_ | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n if hasattr(self, '_predicted_hazards_'):\n self._concordance_score_ = concordance_index(self.durations, -self.\n _predicted_hazards_, self.event_observed)\n del self._predicted_hazards_\n return self._concordance_score_\n return self._concordance_score_\n",
"role": "user"
},
{
"content": "score_",
"role": "assistant"
}
] |
def x(self):
df = pd.DataFrame(index=self.cumulative_hazards_.columns)
betas, se = self._compute_slopes()
df['slope(coef)'] = betas
df['se(slope(coef))'] = se
return df
| summary | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n df = pd.DataFrame(index=self.cumulative_hazards_.columns)\n betas, se = self._compute_slopes()\n df['slope(coef)'] = betas\n df['se(slope(coef))'] = se\n return df\n",
"role": "user"
},
{
"content": "summary",
"role": "assistant"
}
] |
def x(self, decimals=2, **kwargs):
justify = string_justify(18)
print(self)
print("{} = '{}'".format(justify('duration col'), self.duration_col))
print("{} = '{}'".format(justify('event col'), self.event_col))
if self.weights_col:
print("{} = '{}'".format(justify('weights col'), self.weights_col))
if self.coef_penalizer > 0:
print("{} = '{}'".format(justify('coef penalizer'), self.
coef_penalizer))
if self.smoothing_penalizer > 0:
print("{} = '{}'".format(justify('smoothing penalizer'), self.
smoothing_penalizer))
print('{} = {}'.format(justify('number of subjects'), self._n_examples))
print('{} = {}'.format(justify('number of events'), self.event_observed
.sum()))
print('{} = {}'.format(justify('time fit was run'), self.
_time_fit_was_called))
for k, v in kwargs.items():
print('{} = {}\n'.format(justify(k), v))
print(end='\n')
print('---')
df = self.summary
print(df.to_string(float_format=format_floats(decimals), formatters={
'p': format_p_value(decimals), 'exp(coef)': format_exp_floats(
decimals)}))
print('---')
print('Concordance = {:.{prec}f}'.format(self.score_, prec=decimals))
| print_summary | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, decimals=2, **kwargs):\n justify = string_justify(18)\n print(self)\n print(\"{} = '{}'\".format(justify('duration col'), self.duration_col))\n print(\"{} = '{}'\".format(justify('event col'), self.event_col))\n if self.weights_col:\n print(\"{} = '{}'\".format(justify('weights col'), self.weights_col))\n if self.coef_penalizer > 0:\n print(\"{} = '{}'\".format(justify('coef penalizer'), self.\n coef_penalizer))\n if self.smoothing_penalizer > 0:\n print(\"{} = '{}'\".format(justify('smoothing penalizer'), self.\n smoothing_penalizer))\n print('{} = {}'.format(justify('number of subjects'), self._n_examples))\n print('{} = {}'.format(justify('number of events'), self.event_observed\n .sum()))\n print('{} = {}'.format(justify('time fit was run'), self.\n _time_fit_was_called))\n for k, v in kwargs.items():\n print('{} = {}\\n'.format(justify(k), v))\n print(end='\\n')\n print('---')\n df = self.summary\n print(df.to_string(float_format=format_floats(decimals), formatters={\n 'p': format_p_value(decimals), 'exp(coef)': format_exp_floats(\n decimals)}))\n print('---')\n print('Concordance = {:.{prec}f}'.format(self.score_, prec=decimals))\n",
"role": "user"
},
{
"content": "print_summary",
"role": "assistant"
}
] |
def x(self, X, ancillary_X=None, p=0.5):
lambda_, rho_ = self._prep_inputs_for_prediction_and_return_scores(X,
ancillary_X)
return pd.DataFrame(lambda_ * np.power(-np.log(p), 1 / rho_), index=
_get_index(X))
| predict_percentile | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, X, ancillary_X=None, p=0.5):\n lambda_, rho_ = self._prep_inputs_for_prediction_and_return_scores(X,\n ancillary_X)\n return pd.DataFrame(lambda_ * np.power(-np.log(p), 1 / rho_), index=\n _get_index(X))\n",
"role": "user"
},
{
"content": "predict_percentile",
"role": "assistant"
}
] |
def x(self, X, ancillary_X=None):
lambda_, rho_ = self._prep_inputs_for_prediction_and_return_scores(X,
ancillary_X)
return pd.DataFrame(lambda_ * gamma(1 + 1 / rho_), index=_get_index(X))
| predict_expectation | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, X, ancillary_X=None):\n lambda_, rho_ = self._prep_inputs_for_prediction_and_return_scores(X,\n ancillary_X)\n return pd.DataFrame(lambda_ * gamma(1 + 1 / rho_), index=_get_index(X))\n",
"role": "user"
},
{
"content": "predict_expectation",
"role": "assistant"
}
] |
def x(self, X, times=None, ancillary_X=None):
times = coalesce(times, self.timeline, np.unique(self.durations))
lambda_, rho_ = self._prep_inputs_for_prediction_and_return_scores(X,
ancillary_X)
return pd.DataFrame(np.outer(times, 1 / lambda_) ** rho_, columns=
_get_index(X), index=times)
| predict_cumulative_hazard | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, X, times=None, ancillary_X=None):\n times = coalesce(times, self.timeline, np.unique(self.durations))\n lambda_, rho_ = self._prep_inputs_for_prediction_and_return_scores(X,\n ancillary_X)\n return pd.DataFrame(np.outer(times, 1 / lambda_) ** rho_, columns=\n _get_index(X), index=times)\n",
"role": "user"
},
{
"content": "predict_cumulative_hazard",
"role": "assistant"
}
] |
def x(self, durations, event_observed, event_of_interest, timeline=None,
entry=None, label='AJ_estimate', alpha=None, ci_labels=None, weights=None):
self._censoring_type = CensoringType.RIGHT
ties = self._check_for_duplicates(durations=durations, events=
event_observed)
if ties:
warnings.warn(dedent(
"""Tied event times were detected. The Aalen-Johansen estimator cannot handle tied event times.
To resolve ties, data is randomly jittered."""
), Warning)
durations = self._jitter(durations=pd.Series(durations), event=pd.
Series(event_observed), jitter_level=self._jitter_level, seed=
self._seed)
alpha = alpha if alpha else self.alpha
event_of_interest = int(event_of_interest)
cmprisk_label = 'CIF_' + str(event_of_interest)
self.label_cmprisk = 'observed_' + str(event_of_interest)
km = KaplanMeierFitter().fit(durations, event_observed=event_observed,
timeline=timeline, entry=entry, weights=weights)
aj = km.event_table
aj['overall_survival'] = km.survival_function_
aj['lagged_overall_survival'] = aj['overall_survival'].shift()
event_spec = pd.Series(event_observed) == event_of_interest
self.durations, self.event_observed, *_, event_table = _preprocess_inputs(
durations=durations, event_observed=event_spec, timeline=timeline,
entry=entry, weights=weights)
event_spec_times = event_table['observed']
event_spec_times = event_spec_times.rename(self.label_cmprisk)
aj = pd.concat([aj, event_spec_times], axis=1).reset_index()
aj[cmprisk_label] = (aj[self.label_cmprisk] / aj['at_risk'] * aj[
'lagged_overall_survival']).cumsum()
aj.loc[0, cmprisk_label] = 0
aj = aj.set_index('event_at')
self._estimation_method = 'cumulative_density_'
self._estimate_name = 'cumulative_density_'
self._predict_label = label
self._update_docstrings()
self._label = label
self.cumulative_density_ = pd.DataFrame(aj[cmprisk_label])
self.event_table = aj[['removed', 'observed', self.label_cmprisk,
'censored', 'entrance', 'at_risk']]
if self._calc_var:
self.variance_, self.confidence_interval_ = self._bounds(aj[
'lagged_overall_survival'], alpha=alpha, ci_labels=ci_labels)
else:
self.variance_, self.confidence_interval_ = None, None
return self
| fit | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, durations, event_observed, event_of_interest, timeline=None,\n entry=None, label='AJ_estimate', alpha=None, ci_labels=None, weights=None):\n self._censoring_type = CensoringType.RIGHT\n ties = self._check_for_duplicates(durations=durations, events=\n event_observed)\n if ties:\n warnings.warn(dedent(\n \"\"\"Tied event times were detected. The Aalen-Johansen estimator cannot handle tied event times.\n To resolve ties, data is randomly jittered.\"\"\"\n ), Warning)\n durations = self._jitter(durations=pd.Series(durations), event=pd.\n Series(event_observed), jitter_level=self._jitter_level, seed=\n self._seed)\n alpha = alpha if alpha else self.alpha\n event_of_interest = int(event_of_interest)\n cmprisk_label = 'CIF_' + str(event_of_interest)\n self.label_cmprisk = 'observed_' + str(event_of_interest)\n km = KaplanMeierFitter().fit(durations, event_observed=event_observed,\n timeline=timeline, entry=entry, weights=weights)\n aj = km.event_table\n aj['overall_survival'] = km.survival_function_\n aj['lagged_overall_survival'] = aj['overall_survival'].shift()\n event_spec = pd.Series(event_observed) == event_of_interest\n self.durations, self.event_observed, *_, event_table = _preprocess_inputs(\n durations=durations, event_observed=event_spec, timeline=timeline,\n entry=entry, weights=weights)\n event_spec_times = event_table['observed']\n event_spec_times = event_spec_times.rename(self.label_cmprisk)\n aj = pd.concat([aj, event_spec_times], axis=1).reset_index()\n aj[cmprisk_label] = (aj[self.label_cmprisk] / aj['at_risk'] * aj[\n 'lagged_overall_survival']).cumsum()\n aj.loc[0, cmprisk_label] = 0\n aj = aj.set_index('event_at')\n self._estimation_method = 'cumulative_density_'\n self._estimate_name = 'cumulative_density_'\n self._predict_label = label\n self._update_docstrings()\n self._label = label\n self.cumulative_density_ = pd.DataFrame(aj[cmprisk_label])\n self.event_table = aj[['removed', 'observed', self.label_cmprisk,\n 'censored', 'entrance', 'at_risk']]\n if self._calc_var:\n self.variance_, self.confidence_interval_ = self._bounds(aj[\n 'lagged_overall_survival'], alpha=alpha, ci_labels=ci_labels)\n else:\n self.variance_, self.confidence_interval_ = None, None\n return self\n",
"role": "user"
},
{
"content": "fit",
"role": "assistant"
}
] |
def x(self, durations, event, jitter_level, seed=None):
np.random.seed(seed)
if jitter_level <= 0:
raise ValueError(
'The jitter level is less than zero, please select a jitter value greater than 0'
)
event_times = durations[event != 0].copy()
n = event_times.shape[0]
shift = np.random.uniform(low=-1, high=1, size=n) * jitter_level
event_times += shift
durations_jitter = event_times.align(durations)[0].fillna(durations)
if self._check_for_duplicates(durations=durations_jitter, events=event):
return self._jitter(durations=durations_jitter, event=event,
jitter_level=jitter_level, seed=seed)
return durations_jitter
| jitter | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, durations, event, jitter_level, seed=None):\n np.random.seed(seed)\n if jitter_level <= 0:\n raise ValueError(\n 'The jitter level is less than zero, please select a jitter value greater than 0'\n )\n event_times = durations[event != 0].copy()\n n = event_times.shape[0]\n shift = np.random.uniform(low=-1, high=1, size=n) * jitter_level\n event_times += shift\n durations_jitter = event_times.align(durations)[0].fillna(durations)\n if self._check_for_duplicates(durations=durations_jitter, events=event):\n return self._jitter(durations=durations_jitter, event=event,\n jitter_level=jitter_level, seed=seed)\n return durations_jitter\n",
"role": "user"
},
{
"content": "jitter",
"role": "assistant"
}
] |
def x(self, lagged_survival, alpha, ci_labels):
ci = 1 - alpha
df = self.event_table.copy()
df['Ft'] = self.cumulative_density_
df['lagS'] = lagged_survival.fillna(1)
if ci_labels is None:
ci_labels = ['%s_upper_%g' % (self._predict_label, ci),
'%s_lower_%g' % (self._predict_label, ci)]
assert len(ci_labels) == 2, 'ci_labels should be a length 2 array.'
all_vars = []
for _, r in df.iterrows():
sf = df.loc[df.index <= r.name].copy()
F_t = float(r['Ft'])
first_term = np.sum((F_t - sf['Ft']) ** 2 * sf['observed'] / sf[
'at_risk'] / (sf['at_risk'] - sf['observed']))
second_term = np.sum(sf['lagS'] ** 2 / sf['at_risk'] * sf[self.
label_cmprisk] / sf['at_risk'] * (sf['at_risk'] - sf[self.
label_cmprisk]) / sf['at_risk'])
third_term = np.sum((F_t - sf['Ft']) / sf['at_risk'] * sf['lagS'] *
sf[self.label_cmprisk] / sf['at_risk'])
variance = first_term + second_term - 2 * third_term
all_vars.append(variance)
df['variance'] = all_vars
df['F_transformed'] = np.log(-np.log(df['Ft']))
df['se_transformed'] = np.sqrt(df['variance']) / (df['Ft'] * np.
absolute(np.log(df['Ft'])))
zalpha = inv_normal_cdf(1 - alpha / 2)
df[ci_labels[0]] = np.exp(-np.exp(df['F_transformed'] + zalpha * df[
'se_transformed']))
df[ci_labels[1]] = np.exp(-np.exp(df['F_transformed'] - zalpha * df[
'se_transformed']))
return df['variance'], df[ci_labels]
| bounds | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, lagged_survival, alpha, ci_labels):\n ci = 1 - alpha\n df = self.event_table.copy()\n df['Ft'] = self.cumulative_density_\n df['lagS'] = lagged_survival.fillna(1)\n if ci_labels is None:\n ci_labels = ['%s_upper_%g' % (self._predict_label, ci), \n '%s_lower_%g' % (self._predict_label, ci)]\n assert len(ci_labels) == 2, 'ci_labels should be a length 2 array.'\n all_vars = []\n for _, r in df.iterrows():\n sf = df.loc[df.index <= r.name].copy()\n F_t = float(r['Ft'])\n first_term = np.sum((F_t - sf['Ft']) ** 2 * sf['observed'] / sf[\n 'at_risk'] / (sf['at_risk'] - sf['observed']))\n second_term = np.sum(sf['lagS'] ** 2 / sf['at_risk'] * sf[self.\n label_cmprisk] / sf['at_risk'] * (sf['at_risk'] - sf[self.\n label_cmprisk]) / sf['at_risk'])\n third_term = np.sum((F_t - sf['Ft']) / sf['at_risk'] * sf['lagS'] *\n sf[self.label_cmprisk] / sf['at_risk'])\n variance = first_term + second_term - 2 * third_term\n all_vars.append(variance)\n df['variance'] = all_vars\n df['F_transformed'] = np.log(-np.log(df['Ft']))\n df['se_transformed'] = np.sqrt(df['variance']) / (df['Ft'] * np.\n absolute(np.log(df['Ft'])))\n zalpha = inv_normal_cdf(1 - alpha / 2)\n df[ci_labels[0]] = np.exp(-np.exp(df['F_transformed'] + zalpha * df[\n 'se_transformed']))\n df[ci_labels[1]] = np.exp(-np.exp(df['F_transformed'] - zalpha * df[\n 'se_transformed']))\n return df['variance'], df[ci_labels]\n",
"role": "user"
},
{
"content": "bounds",
"role": "assistant"
}
] |
def x(durations, events):
df = pd.DataFrame({'t': durations, 'e': events})
dup_times = df.loc[df['e'] != 0, 't'].duplicated(keep=False)
dup_events = df.loc[df['e'] != 0, ['t', 'e']].duplicated(keep=False)
return (dup_times & ~dup_events).any()
| check_for_duplicates | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(durations, events):\n df = pd.DataFrame({'t': durations, 'e': events})\n dup_times = df.loc[df['e'] != 0, 't'].duplicated(keep=False)\n dup_events = df.loc[df['e'] != 0, ['t', 'e']].duplicated(keep=False)\n return (dup_times & ~dup_events).any()\n",
"role": "user"
},
{
"content": "check_for_duplicates",
"role": "assistant"
}
] |
def x(power, ratio_of_participants, p_exp, p_con, postulated_hazard_ratio,
alpha=0.05):
def x(p):
return stats.norm.ppf(p)
m = 1.0 / ratio_of_participants * ((ratio_of_participants *
postulated_hazard_ratio + 1.0) / (postulated_hazard_ratio - 1.0)
) ** 2 * (z(1.0 - alpha / 2.0) + z(power)) ** 2
n_exp = m * ratio_of_participants / (ratio_of_participants * p_exp + p_con)
n_con = m / (ratio_of_participants * p_exp + p_con)
return int(np.ceil(n_exp)), int(np.ceil(n_con))
| sample_size_necessary_under_cph | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(power, ratio_of_participants, p_exp, p_con, postulated_hazard_ratio,\n alpha=0.05):\n\n def x(p):\n return stats.norm.ppf(p)\n m = 1.0 / ratio_of_participants * ((ratio_of_participants *\n postulated_hazard_ratio + 1.0) / (postulated_hazard_ratio - 1.0)\n ) ** 2 * (z(1.0 - alpha / 2.0) + z(power)) ** 2\n n_exp = m * ratio_of_participants / (ratio_of_participants * p_exp + p_con)\n n_con = m / (ratio_of_participants * p_exp + p_con)\n return int(np.ceil(n_exp)), int(np.ceil(n_con))\n",
"role": "user"
},
{
"content": "sample_size_necessary_under_cph",
"role": "assistant"
}
] |
def x(n_exp, n_con, p_exp, p_con, postulated_hazard_ratio, alpha=0.05):
def x(p):
return stats.norm.ppf(p)
m = n_exp * p_exp + n_con * p_con
k = float(n_exp) / float(n_con)
return stats.norm.cdf(np.sqrt(k * m) * abs(postulated_hazard_ratio - 1) /
(k * postulated_hazard_ratio + 1) - z(1 - alpha / 2.0))
| power_under_cph | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(n_exp, n_con, p_exp, p_con, postulated_hazard_ratio, alpha=0.05):\n\n def x(p):\n return stats.norm.ppf(p)\n m = n_exp * p_exp + n_con * p_con\n k = float(n_exp) / float(n_con)\n return stats.norm.cdf(np.sqrt(k * m) * abs(postulated_hazard_ratio - 1) /\n (k * postulated_hazard_ratio + 1) - z(1 - alpha / 2.0))\n",
"role": "user"
},
{
"content": "power_under_cph",
"role": "assistant"
}
] |
def x(x, y, f=2.0 / 3.0, iterations=3):
n = len(x)
r = int(ceil(f * n))
h = [np.sort(np.abs(x - x[i]))[r] for i in range(n)]
w = np.clip(np.abs((x[:, None] - x[None, :]) / h), 0.0, 1.0)
w = (1 - w ** 3) ** 3
yest = np.zeros(n)
delta = np.ones(n)
for _ in range(iterations):
for i in range(n):
weights = delta * w[:, i]
b = np.array([np.sum(weights * y), np.sum(weights * y * x)])
A = np.array([[np.sum(weights), np.sum(weights * x)], [np.sum(
weights * x), np.sum(weights * x * x)]])
beta = linalg.solve(A, b, assume_a='pos', check_finite=False)
yest[i] = beta[0] + beta[1] * x[i]
residuals = y - yest
s = np.median(np.abs(residuals))
delta = np.clip(residuals / (6.0 * s), -1, 1)
delta = (1 - delta ** 2) ** 2
return yest
| lowess | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(x, y, f=2.0 / 3.0, iterations=3):\n n = len(x)\n r = int(ceil(f * n))\n h = [np.sort(np.abs(x - x[i]))[r] for i in range(n)]\n w = np.clip(np.abs((x[:, None] - x[None, :]) / h), 0.0, 1.0)\n w = (1 - w ** 3) ** 3\n yest = np.zeros(n)\n delta = np.ones(n)\n for _ in range(iterations):\n for i in range(n):\n weights = delta * w[:, i]\n b = np.array([np.sum(weights * y), np.sum(weights * y * x)])\n A = np.array([[np.sum(weights), np.sum(weights * x)], [np.sum(\n weights * x), np.sum(weights * x * x)]])\n beta = linalg.solve(A, b, assume_a='pos', check_finite=False)\n yest[i] = beta[0] + beta[1] * x[i]\n residuals = y - yest\n s = np.median(np.abs(residuals))\n delta = np.clip(residuals / (6.0 * s), -1, 1)\n delta = (1 - delta ** 2) ** 2\n return yest\n",
"role": "user"
},
{
"content": "lowess",
"role": "assistant"
}
] |
def x(q, survival_functions, cdf=False):
q = pd.Series(q)
if not ((q <= 1).all() and (0 <= q).all()):
raise ValueError('q must be between 0 and 1')
survival_functions = pd.DataFrame(survival_functions)
if survival_functions.shape[1] == 1 and q.shape == (1,):
q = q[0]
return survival_functions.apply(lambda s: qth_survival_time(q, s,
cdf=cdf)).iloc[0]
else:
d = {_q: survival_functions.apply(lambda s: qth_survival_time(_q, s,
cdf=cdf)) for _q in q}
survival_times = pd.DataFrame(d).T
if q.duplicated().any():
survival_times = survival_times.loc[q]
return survival_times
| qth_survival_times | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(q, survival_functions, cdf=False):\n q = pd.Series(q)\n if not ((q <= 1).all() and (0 <= q).all()):\n raise ValueError('q must be between 0 and 1')\n survival_functions = pd.DataFrame(survival_functions)\n if survival_functions.shape[1] == 1 and q.shape == (1,):\n q = q[0]\n return survival_functions.apply(lambda s: qth_survival_time(q, s,\n cdf=cdf)).iloc[0]\n else:\n d = {_q: survival_functions.apply(lambda s: qth_survival_time(_q, s,\n cdf=cdf)) for _q in q}\n survival_times = pd.DataFrame(d).T\n if q.duplicated().any():\n survival_times = survival_times.loc[q]\n return survival_times\n",
"role": "user"
},
{
"content": "qth_survival_times",
"role": "assistant"
}
] |
def x(q, survival_function, cdf=False):
if type(survival_function) is pd.DataFrame:
if survival_function.shape[1] > 1:
raise ValueError(
'Expecting a dataframe (or series) with a single column. Provide that or use utils.qth_survival_times.'
)
survival_function = survival_function.T.squeeze()
if cdf:
if survival_function.iloc[0] > q:
return -np.inf
v = survival_function.index[survival_function.searchsorted([q])[0]]
else:
if survival_function.iloc[-1] > q:
return np.inf
v = survival_function.index[(-survival_function).searchsorted([-q])[0]]
return v
| qth_survival_time | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(q, survival_function, cdf=False):\n if type(survival_function) is pd.DataFrame:\n if survival_function.shape[1] > 1:\n raise ValueError(\n 'Expecting a dataframe (or series) with a single column. Provide that or use utils.qth_survival_times.'\n )\n survival_function = survival_function.T.squeeze()\n if cdf:\n if survival_function.iloc[0] > q:\n return -np.inf\n v = survival_function.index[survival_function.searchsorted([q])[0]]\n else:\n if survival_function.iloc[-1] > q:\n return np.inf\n v = survival_function.index[(-survival_function).searchsorted([-q])[0]]\n return v\n",
"role": "user"
},
{
"content": "qth_survival_time",
"role": "assistant"
}
] |
def x(groups, durations, event_observed, birth_times=None, limit=-1):
n = np.max(groups.shape)
assert n == np.max(durations.shape) == np.max(event_observed.shape
), 'inputs must be of the same length.'
if birth_times is None:
birth_times = np.zeros(np.max(durations.shape))
birth_times[:] = np.min(durations)
assert n == np.max(birth_times.shape), 'inputs must be of the same length.'
groups, durations, event_observed, birth_times = [pd.Series(np.asarray(
vector).reshape(n)) for vector in [groups, durations,
event_observed, birth_times]]
unique_groups = groups.unique()
for i, group in enumerate(unique_groups):
ix = groups == group
T = durations[ix]
C = event_observed[ix]
B = birth_times[ix]
group_name = str(group)
columns = [(event_name + ':' + group_name) for event_name in [
'removed', 'observed', 'censored', 'entrance', 'at_risk']]
if i == 0:
survival_table = survival_table_from_events(T, C, B, columns=
columns)
else:
survival_table = survival_table.join(survival_table_from_events
(T, C, B, columns=columns), how='outer')
survival_table = survival_table.fillna(0)
if int(limit) != -1:
survival_table = survival_table.loc[:limit]
return unique_groups, survival_table.filter(like='removed:'
), survival_table.filter(like='observed:'), survival_table.filter(like
='censored:')
| group_survival_table_from_events | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(groups, durations, event_observed, birth_times=None, limit=-1):\n n = np.max(groups.shape)\n assert n == np.max(durations.shape) == np.max(event_observed.shape\n ), 'inputs must be of the same length.'\n if birth_times is None:\n birth_times = np.zeros(np.max(durations.shape))\n birth_times[:] = np.min(durations)\n assert n == np.max(birth_times.shape), 'inputs must be of the same length.'\n groups, durations, event_observed, birth_times = [pd.Series(np.asarray(\n vector).reshape(n)) for vector in [groups, durations,\n event_observed, birth_times]]\n unique_groups = groups.unique()\n for i, group in enumerate(unique_groups):\n ix = groups == group\n T = durations[ix]\n C = event_observed[ix]\n B = birth_times[ix]\n group_name = str(group)\n columns = [(event_name + ':' + group_name) for event_name in [\n 'removed', 'observed', 'censored', 'entrance', 'at_risk']]\n if i == 0:\n survival_table = survival_table_from_events(T, C, B, columns=\n columns)\n else:\n survival_table = survival_table.join(survival_table_from_events\n (T, C, B, columns=columns), how='outer')\n survival_table = survival_table.fillna(0)\n if int(limit) != -1:\n survival_table = survival_table.loc[:limit]\n return unique_groups, survival_table.filter(like='removed:'\n ), survival_table.filter(like='observed:'), survival_table.filter(like\n ='censored:')\n",
"role": "user"
},
{
"content": "group_survival_table_from_events",
"role": "assistant"
}
] |
def x(death_times, event_observed, birth_times=None, columns=['removed',
'observed', 'censored', 'entrance', 'at_risk'], weights=None, collapse=
False, intervals=None):
removed, observed, censored, entrance, at_risk = columns
death_times = np.asarray(death_times)
if birth_times is None:
birth_times = min(0, death_times.min()) * np.ones(death_times.shape[0])
else:
birth_times = np.asarray(birth_times)
if np.any(birth_times > death_times):
raise ValueError('birth time must be less than time of death.')
if weights is None:
weights = 1
df = pd.DataFrame(death_times, columns=['event_at'])
df[removed] = np.asarray(weights)
df[observed] = np.asarray(weights) * np.asarray(event_observed).astype(bool
)
death_table = df.groupby('event_at').sum()
death_table[censored] = (death_table[removed] - death_table[observed]
).astype(int)
births = pd.DataFrame(birth_times, columns=['event_at'])
births[entrance] = np.asarray(weights)
births_table = births.groupby('event_at').sum()
event_table = death_table.join(births_table, how='outer', sort=True
).fillna(0)
event_table[at_risk] = event_table[entrance].cumsum() - event_table[removed
].cumsum().shift(1).fillna(0)
if collapse or intervals is not None:
event_table = _group_event_table_by_intervals(event_table, intervals)
if (np.asarray(weights).astype(int) != weights).any():
return event_table.astype(float)
return event_table.astype(int)
| survival_table_from_events | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(death_times, event_observed, birth_times=None, columns=['removed',\n 'observed', 'censored', 'entrance', 'at_risk'], weights=None, collapse=\n False, intervals=None):\n removed, observed, censored, entrance, at_risk = columns\n death_times = np.asarray(death_times)\n if birth_times is None:\n birth_times = min(0, death_times.min()) * np.ones(death_times.shape[0])\n else:\n birth_times = np.asarray(birth_times)\n if np.any(birth_times > death_times):\n raise ValueError('birth time must be less than time of death.')\n if weights is None:\n weights = 1\n df = pd.DataFrame(death_times, columns=['event_at'])\n df[removed] = np.asarray(weights)\n df[observed] = np.asarray(weights) * np.asarray(event_observed).astype(bool\n )\n death_table = df.groupby('event_at').sum()\n death_table[censored] = (death_table[removed] - death_table[observed]\n ).astype(int)\n births = pd.DataFrame(birth_times, columns=['event_at'])\n births[entrance] = np.asarray(weights)\n births_table = births.groupby('event_at').sum()\n event_table = death_table.join(births_table, how='outer', sort=True\n ).fillna(0)\n event_table[at_risk] = event_table[entrance].cumsum() - event_table[removed\n ].cumsum().shift(1).fillna(0)\n if collapse or intervals is not None:\n event_table = _group_event_table_by_intervals(event_table, intervals)\n if (np.asarray(weights).astype(int) != weights).any():\n return event_table.astype(float)\n return event_table.astype(int)\n",
"role": "user"
},
{
"content": "survival_table_from_events",
"role": "assistant"
}
] |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.