sentence1
stringlengths
52
3.87M
sentence2
stringlengths
1
47.2k
label
stringclasses
1 value
def _buttonbox(msg, title, choices, root=None, timeout=None): """ Display a msg, a title, and a set of buttons. The buttons are defined by the members of the choices list. Return the text of the button that the user selected. @arg msg: the msg to be displayed. @arg title: the window title @arg choices: a list or tuple of the choices to be displayed """ global boxRoot, __replyButtonText, __widgetTexts, buttonsFrame # Initialize __replyButtonText to the first choice. # This is what will be used if the window is closed by the close button. __replyButtonText = choices[0] if root: root.withdraw() boxRoot = tk.Toplevel(master=root) boxRoot.withdraw() else: boxRoot = tk.Tk() boxRoot.withdraw() boxRoot.title(title) boxRoot.iconname('Dialog') boxRoot.geometry(rootWindowPosition) boxRoot.minsize(400, 100) # ------------- define the messageFrame --------------------------------- messageFrame = tk.Frame(master=boxRoot) messageFrame.pack(side=tk.TOP, fill=tk.BOTH) # ------------- define the buttonsFrame --------------------------------- buttonsFrame = tk.Frame(master=boxRoot) buttonsFrame.pack(side=tk.TOP, fill=tk.BOTH) # -------------------- place the widgets in the frames ----------------------- messageWidget = tk.Message(messageFrame, text=msg, width=400) messageWidget.configure(font=(PROPORTIONAL_FONT_FAMILY, PROPORTIONAL_FONT_SIZE)) messageWidget.pack(side=tk.TOP, expand=tk.YES, fill=tk.X, padx='3m', pady='3m') __put_buttons_in_buttonframe(choices) # -------------- the action begins ----------- # put the focus on the first button __firstWidget.focus_force() boxRoot.deiconify() if timeout is not None: boxRoot.after(timeout, timeoutBoxRoot) boxRoot.mainloop() try: boxRoot.destroy() except tk.TclError: if __replyButtonText != TIMEOUT_TEXT: __replyButtonText = None if root: root.deiconify() return __replyButtonText
Display a msg, a title, and a set of buttons. The buttons are defined by the members of the choices list. Return the text of the button that the user selected. @arg msg: the msg to be displayed. @arg title: the window title @arg choices: a list or tuple of the choices to be displayed
entailment
def __put_buttons_in_buttonframe(choices): """Put the buttons in the buttons frame""" global __widgetTexts, __firstWidget, buttonsFrame __firstWidget = None __widgetTexts = {} i = 0 for buttonText in choices: tempButton = tk.Button(buttonsFrame, takefocus=1, text=buttonText) _bindArrows(tempButton) tempButton.pack(expand=tk.YES, side=tk.LEFT, padx='1m', pady='1m', ipadx='2m', ipady='1m') # remember the text associated with this widget __widgetTexts[tempButton] = buttonText # remember the first widget, so we can put the focus there if i == 0: __firstWidget = tempButton i = 1 # for the commandButton, bind activation events to the activation event handler commandButton = tempButton handler = __buttonEvent for selectionEvent in STANDARD_SELECTION_EVENTS: commandButton.bind('<%s>' % selectionEvent, handler) if CANCEL_TEXT in choices: commandButton.bind('<Escape>', __cancelButtonEvent)
Put the buttons in the buttons frame
entailment
def __fillablebox(msg, title='', default='', mask=None, root=None, timeout=None): """ Show a box in which a user can enter some text. You may optionally specify some default text, which will appear in the enterbox when it is displayed. Returns the text that the user entered, or None if he cancels the operation. """ global boxRoot, __enterboxText, __enterboxDefaultText global cancelButton, entryWidget, okButton if title == None: title == '' if default == None: default = '' __enterboxDefaultText = default __enterboxText = __enterboxDefaultText if root: root.withdraw() boxRoot = tk.Toplevel(master=root) boxRoot.withdraw() else: boxRoot = tk.Tk() boxRoot.withdraw() boxRoot.title(title) boxRoot.iconname('Dialog') boxRoot.geometry(rootWindowPosition) boxRoot.bind('<Escape>', __enterboxCancel) # ------------- define the messageFrame --------------------------------- messageFrame = tk.Frame(master=boxRoot) messageFrame.pack(side=tk.TOP, fill=tk.BOTH) # ------------- define the buttonsFrame --------------------------------- buttonsFrame = tk.Frame(master=boxRoot) buttonsFrame.pack(side=tk.TOP, fill=tk.BOTH) # ------------- define the entryFrame --------------------------------- entryFrame = tk.Frame(master=boxRoot) entryFrame.pack(side=tk.TOP, fill=tk.BOTH) # ------------- define the buttonsFrame --------------------------------- buttonsFrame = tk.Frame(master=boxRoot) buttonsFrame.pack(side=tk.TOP, fill=tk.BOTH) #-------------------- the msg widget ---------------------------- messageWidget = tk.Message(messageFrame, width='4.5i', text=msg) messageWidget.configure(font=(PROPORTIONAL_FONT_FAMILY, PROPORTIONAL_FONT_SIZE)) messageWidget.pack(side=tk.RIGHT, expand=1, fill=tk.BOTH, padx='3m', pady='3m') # --------- entryWidget ---------------------------------------------- entryWidget = tk.Entry(entryFrame, width=40) _bindArrows(entryWidget, skipArrowKeys=True) entryWidget.configure(font=(PROPORTIONAL_FONT_FAMILY, TEXT_ENTRY_FONT_SIZE)) if mask: entryWidget.configure(show=mask) entryWidget.pack(side=tk.LEFT, padx='3m') entryWidget.bind('<Return>', __enterboxGetText) entryWidget.bind('<Escape>', __enterboxCancel) # put text into the entryWidget and have it pre-highlighted if __enterboxDefaultText != '': entryWidget.insert(0,__enterboxDefaultText) entryWidget.select_range(0, tk.END) # ------------------ ok button ------------------------------- okButton = tk.Button(buttonsFrame, takefocus=1, text=OK_TEXT) _bindArrows(okButton) okButton.pack(expand=1, side=tk.LEFT, padx='3m', pady='3m', ipadx='2m', ipady='1m') # for the commandButton, bind activation events to the activation event handler commandButton = okButton handler = __enterboxGetText for selectionEvent in STANDARD_SELECTION_EVENTS: commandButton.bind('<%s>' % selectionEvent, handler) # ------------------ cancel button ------------------------------- cancelButton = tk.Button(buttonsFrame, takefocus=1, text=CANCEL_TEXT) _bindArrows(cancelButton) cancelButton.pack(expand=1, side=tk.RIGHT, padx='3m', pady='3m', ipadx='2m', ipady='1m') # for the commandButton, bind activation events to the activation event handler commandButton = cancelButton handler = __enterboxCancel for selectionEvent in STANDARD_SELECTION_EVENTS: commandButton.bind('<%s>' % selectionEvent, handler) # ------------------- time for action! ----------------- entryWidget.focus_force() # put the focus on the entryWidget boxRoot.deiconify() if timeout is not None: boxRoot.after(timeout, timeoutBoxRoot) boxRoot.mainloop() # run it! # -------- after the run has completed ---------------------------------- if root: root.deiconify() try: boxRoot.destroy() # button_click didn't destroy boxRoot, so we do it now except tk.TclError: if __enterboxText != TIMEOUT_TEXT: return None return __enterboxText
Show a box in which a user can enter some text. You may optionally specify some default text, which will appear in the enterbox when it is displayed. Returns the text that the user entered, or None if he cancels the operation.
entailment
def parse(self, response): """ Parse the login xml response :param response: the login response from the RETS server :return: None """ self.headers = response.headers if 'xml' in self.headers.get('Content-Type'): # Got an XML response, likely an error code. xml = xmltodict.parse(response.text) self.analyze_reply_code(xml_response_dict=xml) lines = response.text.split('\r\n') if len(lines) < 3: lines = response.text.split('\n') for line in lines: line = line.strip() name, value = self.read_line(line) if name: if name in self.valid_transactions or re.match(pattern='/^X\-/', string=name): self.capabilities[name] = value else: self.details[name] = value
Parse the login xml response :param response: the login response from the RETS server :return: None
entailment
def read_line(line): """Reads lines of XML and delimits, strips, and returns.""" name, value = '', '' if '=' in line: name, value = line.split('=', 1) return [name.strip(), value.strip()]
Reads lines of XML and delimits, strips, and returns.
entailment
def generator(self, response): """ Takes a response socket connection and iteratively parses and yields the results as python dictionaries. :param response: a Requests response object with stream=True :return: """ delim = '\t' # Default to tab delimited columns = [] response.raw.decode_content = True events = ET.iterparse(BytesIO(response.content)) results = [] for event, elem in events: # Analyze search record data if "DATA" == elem.tag: data_dict = {column: data for column, data in zip(columns, elem.text.split(delim)) if column != ''} self.parsed_rows += 1 # Rows parsed with all requests results.append(data_dict) # Handle reply code elif "RETS" == elem.tag: reply_code = elem.get('ReplyCode') reply_text = elem.get('ReplyText') if reply_code == '20201': # RETS Response 20201 - No Records Found # Generator should continue and return nothing continue elif reply_code != '0': msg = "RETS Error {0!s}: {1!s}".format(reply_code, reply_text) raise RETSException(msg) # Analyze delimiter elif "DELIMITER" == elem.tag: val = elem.get("value") delim = chr(int(val)) # Analyze columns elif "COLUMNS" == elem.tag: columns = elem.text.split(delim) # handle max rows elif "MAXROWS" == elem.tag: logger.debug("MAXROWS Tag reached in XML") logger.debug("Received {0!s} results from this search".format(self.parsed_rows)) raise MaxrowException(results) else: # This is a tag we don't process (like COUNT) continue elem.clear() return results
Takes a response socket connection and iteratively parses and yields the results as python dictionaries. :param response: a Requests response object with stream=True :return:
entailment
def flatten_urlinfo(urlinfo, shorter_keys=True): """ Takes a urlinfo object and returns a flat dictionary.""" def flatten(value, prefix=""): if is_string(value): _result[prefix[1:]] = value return try: len(value) except (AttributeError, TypeError): # a leaf _result[prefix[1:]] = value return try: items = value.items() except AttributeError: # an iterable, but not a dict last_prefix = prefix.split(".")[-1] if shorter_keys: prefix = "." + last_prefix if last_prefix == "Country": for v in value: country = v.pop("@Code") flatten(v, ".".join([prefix, country])) elif last_prefix in ["RelatedLink", "CategoryData"]: for i, v in enumerate(value): flatten(v, ".".join([prefix, str(i)])) elif value[0].get("TimeRange"): for v in value: time_range = ".".join(tuple(v.pop("TimeRange").items())[0]) # python 3 odict_items don't support indexing if v.get("DataUrl"): time_range = ".".join([v.pop("DataUrl"), time_range]) flatten(v, ".".join([prefix, time_range])) else: msg = prefix + " contains a list we don't know how to flatten." raise NotImplementedError(msg) else: # a dict, go one level deeper for k, v in items: flatten(v, ".".join([prefix, k])) _result = {} info = xmltodict.parse(str(urlinfo)) flatten(info["aws:UrlInfoResponse"]["Response"]["UrlInfoResult"]["Alexa"]) _result["OutputTimestamp"] = datetime.datetime.utcnow().strftime('%Y%m%dT%H%M%SZ') return _result
Takes a urlinfo object and returns a flat dictionary.
entailment
def create_v4_signature(self, request_params): ''' Create URI and signature headers based on AWS V4 signing process. Refer to https://docs.aws.amazon.com/AlexaWebInfoService/latest/ApiReferenceArticle.html for request params. :param request_params: dictionary of request parameters :return: URL and header to be passed to requests.get ''' method = 'GET' service = 'awis' host = 'awis.us-west-1.amazonaws.com' region = 'us-west-1' endpoint = 'https://awis.amazonaws.com/api' request_parameters = urlencode([(key, request_params[key]) for key in sorted(request_params.keys())]) # Key derivation functions. See: # http://docs.aws.amazon.com/general/latest/gr/signature-v4-examples.html#signature-v4-examples-python def sign(key, msg): return hmac.new(key, msg.encode('utf-8'), hashlib.sha256).digest() def getSignatureKey(key, dateStamp, regionName, serviceName): kDate = sign(('AWS4' + key).encode('utf-8'), dateStamp) kRegion = sign(kDate, regionName) kService = sign(kRegion, serviceName) kSigning = sign(kService, 'aws4_request') return kSigning # Create a date for headers and the credential string t = datetime.datetime.utcnow() amzdate = t.strftime('%Y%m%dT%H%M%SZ') datestamp = t.strftime('%Y%m%d') # Date w/o time, used in credential scope # Create canonical request canonical_uri = '/api' canonical_querystring = request_parameters canonical_headers = 'host:' + host + '\n' + 'x-amz-date:' + amzdate + '\n' signed_headers = 'host;x-amz-date' payload_hash = hashlib.sha256(''.encode('utf8')).hexdigest() canonical_request = method + '\n' + canonical_uri + '\n' + canonical_querystring + '\n' + canonical_headers + '\n' + signed_headers + '\n' + payload_hash # Create string to sign algorithm = 'AWS4-HMAC-SHA256' credential_scope = datestamp + '/' + region + '/' + service + '/' + 'aws4_request' string_to_sign = algorithm + '\n' + amzdate + '\n' + credential_scope + '\n' + hashlib.sha256(canonical_request.encode('utf8')).hexdigest() # Calculate signature signing_key = getSignatureKey(self.secret_access_key, datestamp, region, service) # Sign the string_to_sign using the signing_key signature = hmac.new(signing_key, (string_to_sign).encode('utf-8'), hashlib.sha256).hexdigest() # Add signing information to the request authorization_header = algorithm + ' ' + 'Credential=' + self.access_id + '/' + credential_scope + ', ' + 'SignedHeaders=' + signed_headers + ', ' + 'Signature=' + signature headers = {'X-Amz-Date':amzdate, 'Authorization':authorization_header, 'Content-Type': 'application/xml', 'Accept': 'application/xml'} # Create request url request_url = endpoint + '?' + canonical_querystring return request_url, headers
Create URI and signature headers based on AWS V4 signing process. Refer to https://docs.aws.amazon.com/AlexaWebInfoService/latest/ApiReferenceArticle.html for request params. :param request_params: dictionary of request parameters :return: URL and header to be passed to requests.get
entailment
def urlinfo(self, domain, response_group = URLINFO_RESPONSE_GROUPS): ''' Provide information about supplied domain as specified by the response group :param domain: Any valid URL :param response_group: Any valid urlinfo response group :return: Traffic and/or content data of the domain in XML format ''' params = { 'Action': "UrlInfo", 'Url': domain, 'ResponseGroup': response_group } url, headers = self.create_v4_signature(params) return self.return_output(url, headers)
Provide information about supplied domain as specified by the response group :param domain: Any valid URL :param response_group: Any valid urlinfo response group :return: Traffic and/or content data of the domain in XML format
entailment
def traffichistory(self, domain, response_group=TRAFFICINFO_RESPONSE_GROUPS, myrange=31, start=20070801): ''' Provide traffic history of supplied domain :param domain: Any valid URL :param response_group: Any valid traffic history response group :return: Traffic and/or content data of the domain in XML format ''' params = { 'Action': "TrafficHistory", 'Url': domain, 'ResponseGroup': response_group, 'Range': myrange, 'Start': start, } url, headers = self.create_v4_signature(params) return self.return_output(url, headers)
Provide traffic history of supplied domain :param domain: Any valid URL :param response_group: Any valid traffic history response group :return: Traffic and/or content data of the domain in XML format
entailment
def cat_browse(self, domain, path, response_group=CATEGORYBROWSE_RESPONSE_GROUPS, descriptions='True'): ''' Provide category browse information of specified domain :param domain: Any valid URL :param path: Valid category path :param response_group: Any valid traffic history response group :return: Traffic and/or content data of the domain in XML format ''' params = { 'Action': "CategoryListings", 'ResponseGroup': 'Listings', 'Path': quote(path), 'Descriptions': descriptions } url, headers = self.create_v4_signature(params) return self.return_output(url, headers)
Provide category browse information of specified domain :param domain: Any valid URL :param path: Valid category path :param response_group: Any valid traffic history response group :return: Traffic and/or content data of the domain in XML format
entailment
def add_capability(self, name, uri): """ Add a capability of the RETS board :param name: The name of the capability :param uri: The capability URI given by the RETS board :return: None """ parse_results = urlparse(uri) if parse_results.hostname is None: # relative URL given, so build this into an absolute URL login_url = self.capabilities.get('Login') if not login_url: logger.error("There is no login URL stored, so additional capabilities cannot be added.") raise ValueError("Cannot automatically determine absolute path for {0!s} given.".format(uri)) parts = urlparse(login_url) port = ':{}'.format(parts.port) if parts.port else '' uri = parts.scheme + '://' + parts.hostname + port + '/' + uri.lstrip('/') self.capabilities[name] = uri
Add a capability of the RETS board :param name: The name of the capability :param uri: The capability URI given by the RETS board :return: None
entailment
def login(self): """ Login to the RETS board and return an instance of Bulletin :return: Bulletin instance """ response = self._request('Login') parser = OneXLogin() parser.parse(response) self.session_id = response.cookies.get(self.session_id_cookie_name, '') if parser.headers.get('RETS-Version') is not None: self.version = str(parser.headers.get('RETS-Version')) self.client.headers['RETS-Version'] = self.version for k, v in parser.capabilities.items(): self.add_capability(k, v) if self.capabilities.get('Action'): self._request('Action') return True
Login to the RETS board and return an instance of Bulletin :return: Bulletin instance
entailment
def get_resource_metadata(self, resource=None): """ Get resource metadata :param resource: The name of the resource to get metadata for :return: list """ result = self._make_metadata_request(meta_id=0, metadata_type='METADATA-RESOURCE') if resource: result = next((item for item in result if item['ResourceID'] == resource), None) return result
Get resource metadata :param resource: The name of the resource to get metadata for :return: list
entailment
def get_table_metadata(self, resource, resource_class): """ Get metadata for a given resource: class :param resource: The name of the resource :param resource_class: The name of the class to get metadata from :return: list """ return self._make_metadata_request(meta_id=resource + ':' + resource_class, metadata_type='METADATA-TABLE')
Get metadata for a given resource: class :param resource: The name of the resource :param resource_class: The name of the class to get metadata from :return: list
entailment
def get_lookup_values(self, resource, lookup_name): """ Get possible lookup values for a given field :param resource: The name of the resource :param lookup_name: The name of the the field to get lookup values for :return: list """ return self._make_metadata_request(meta_id=resource + ':' + lookup_name, metadata_type='METADATA-LOOKUP_TYPE')
Get possible lookup values for a given field :param resource: The name of the resource :param lookup_name: The name of the the field to get lookup values for :return: list
entailment
def _make_metadata_request(self, meta_id, metadata_type=None): """ Get the Metadata. The Session initializes with 'COMPACT-DECODED' as the format type. If that returns a DTD error then we change to the 'STANDARD-XML' format and try again. :param meta_id: The name of the resource, class, or lookup to get metadata for :param metadata_type: The RETS metadata type :return: list """ # If this metadata _request has already happened, returned the saved result. key = '{0!s}:{1!s}'.format(metadata_type, meta_id) if key in self.metadata_responses and self.cache_metadata: response = self.metadata_responses[key] else: response = self._request( capability='GetMetadata', options={ 'query': { 'Type': metadata_type, 'ID': meta_id, 'Format': self.metadata_format } } ) self.metadata_responses[key] = response if self.metadata_format == 'COMPACT-DECODED': parser = CompactMetadata() else: parser = StandardXMLetadata() try: return parser.parse(response=response, metadata_type=metadata_type) except RETSException as e: # Remove response from cache self.metadata_responses.pop(key, None) # If the server responds with an invalid parameter for COMPACT-DECODED, try STANDARD-XML if self.metadata_format != 'STANDARD-XML' and e.reply_code in ['20513', '20514']: self.metadata_responses.pop(key, None) self.metadata_format = 'STANDARD-XML' return self._make_metadata_request(meta_id=meta_id, metadata_type=metadata_type) raise RETSException(e.reply_text, e.reply_code)
Get the Metadata. The Session initializes with 'COMPACT-DECODED' as the format type. If that returns a DTD error then we change to the 'STANDARD-XML' format and try again. :param meta_id: The name of the resource, class, or lookup to get metadata for :param metadata_type: The RETS metadata type :return: list
entailment
def get_preferred_object(self, resource, object_type, content_id, location=0): """ Get the first object from a Resource :param resource: The name of the resource :param object_type: The type of object to fetch :param content_id: The unique id of the item to get objects for :param location: The path to get Objects from :return: Object """ collection = self.get_object(resource=resource, object_type=object_type, content_ids=content_id, object_ids='0', location=location) return collection[0]
Get the first object from a Resource :param resource: The name of the resource :param object_type: The type of object to fetch :param content_id: The unique id of the item to get objects for :param location: The path to get Objects from :return: Object
entailment
def get_object(self, resource, object_type, content_ids, object_ids='*', location=0): """ Get a list of Objects from a resource :param resource: The resource to get objects from :param object_type: The type of object to fetch :param content_ids: The unique id of the item to get objects for :param object_ids: ids of the objects to download :param location: The path to get Objects from :return: list """ object_helper = GetObject() request_ids = object_helper.ids(content_ids=content_ids, object_ids=object_ids) response = self._request( capability='GetObject', options={ 'query': { "Resource": resource, "Type": object_type, "ID": ','.join(request_ids), "Location": location } } ) if 'multipart' in response.headers.get('Content-Type'): parser = MultipleObjectParser() collection = parser.parse_image_response(response) else: parser = SingleObjectParser() collection = [parser.parse_image_response(response)] return collection
Get a list of Objects from a resource :param resource: The resource to get objects from :param object_type: The type of object to fetch :param content_ids: The unique id of the item to get objects for :param object_ids: ids of the objects to download :param location: The path to get Objects from :return: list
entailment
def search(self, resource, resource_class, search_filter=None, dmql_query=None, limit=9999999, offset=0, optional_parameters=None, auto_offset=True, query_type='DMQL2', standard_names=0, response_format='COMPACT-DECODED'): """ Preform a search on the RETS board :param resource: The resource that contains the class to search :param resource_class: The class to search :param search_filter: The query as a dict :param dmql_query: The query in dmql format :param limit: Limit search values count :param offset: Offset for RETS request. Useful when RETS limits number of results or transactions :param optional_parameters: Values for option paramters :param auto_offset: Should the search be allowed to trigger subsequent searches. :param query_type: DMQL or DMQL2 depending on the rets server. :param standard_names: 1 to use standard names, 0 to use system names :param response_format: COMPACT-DECODED, COMPACT, or STANDARD-XML :return: dict """ if (search_filter and dmql_query) or (not search_filter and not dmql_query): raise ValueError("You may specify either a search_filter or dmql_query") search_helper = DMQLHelper() if dmql_query: dmql_query = search_helper.dmql(query=dmql_query) else: dmql_query = search_helper.filter_to_dmql(filter_dict=search_filter) parameters = { 'SearchType': resource, 'Class': resource_class, 'Query': dmql_query, 'QueryType': query_type, 'Count': 1, 'Format': response_format, 'StandardNames': standard_names, } if not optional_parameters: optional_parameters = {} parameters.update(optional_parameters) # if the Select parameter given is an array, format it as it needs to be if 'Select' in parameters and isinstance(parameters.get('Select'), list): parameters['Select'] = ','.join(parameters['Select']) if limit: parameters['Limit'] = limit if offset: parameters['Offset'] = offset search_cursor = OneXSearchCursor() response = self._request( capability='Search', options={ 'query': parameters, }, stream=True ) try: return search_cursor.generator(response=response) except MaxrowException as max_exception: # Recursive searching if automatically performing offsets for the client if auto_offset and limit > len(max_exception.rows_returned): new_limit = limit - len(max_exception.rows_returned) # have not returned results to the desired limit new_offset = offset + len(max_exception.rows_returned) # adjust offset results = self.search(resource=resource, resource_class=resource_class, search_filter=None, dmql_query=dmql_query, offset=new_offset, limit=new_limit, optional_parameters=optional_parameters, auto_offset=auto_offset) previous_results = max_exception.rows_returned return previous_results + results return max_exception.rows_returned
Preform a search on the RETS board :param resource: The resource that contains the class to search :param resource_class: The class to search :param search_filter: The query as a dict :param dmql_query: The query in dmql format :param limit: Limit search values count :param offset: Offset for RETS request. Useful when RETS limits number of results or transactions :param optional_parameters: Values for option paramters :param auto_offset: Should the search be allowed to trigger subsequent searches. :param query_type: DMQL or DMQL2 depending on the rets server. :param standard_names: 1 to use standard names, 0 to use system names :param response_format: COMPACT-DECODED, COMPACT, or STANDARD-XML :return: dict
entailment
def _request(self, capability, options=None, stream=False): """ Make a _request to the RETS server :param capability: The name of the capability to use to get the URI :param options: Options to put into the _request :return: Response """ if options is None: options = {} options.update({ 'headers': self.client.headers.copy() }) url = self.capabilities.get(capability) if not url: msg = "{0!s} tried but no valid endpoints was found. Did you forget to Login?".format(capability) raise NotLoggedIn(msg) if self.user_agent_password: ua_digest = self._user_agent_digest_hash() options['headers']['RETS-UA-Authorization'] = 'Digest {0!s}'.format(ua_digest) if self.use_post_method and capability != 'Action': # Action Requests should always be GET query = options.get('query') response = self.client.post(url, data=query, headers=options['headers'], stream=stream) else: if 'query' in options: url += '?' + '&'.join('{0!s}={1!s}'.format(k, quote(str(v))) for k, v in options['query'].items()) response = self.client.get(url, headers=options['headers'], stream=stream) if response.status_code in [400, 401]: if capability == 'Login': m = "Could not log into the RETS server with the provided credentials." else: m = "The RETS server returned a 401 status code. You must be logged in to make this request." raise NotLoggedIn(m) elif response.status_code == 404 and self.use_post_method: raise HTTPException("Got a 404 when making a POST request. Try setting use_post_method=False when " "initializing the Session.") return response
Make a _request to the RETS server :param capability: The name of the capability to use to get the URI :param options: Options to put into the _request :return: Response
entailment
def _user_agent_digest_hash(self): """ Hash the user agent and user agent password Section 3.10 of https://www.nar.realtor/retsorg.nsf/retsproto1.7d6.pdf :return: md5 """ if not self.version: raise MissingVersion("A version is required for user agent auth. The RETS server should set this" "automatically but it has not. Please instantiate the session with a version argument" "to provide the version.") version_number = self.version.strip('RETS/') user_str = '{0!s}:{1!s}'.format(self.user_agent, self.user_agent_password).encode('utf-8') a1 = hashlib.md5(user_str).hexdigest() session_id = self.session_id if self.session_id is not None else '' digest_str = '{0!s}::{1!s}:{2!s}'.format(a1, session_id, version_number).encode('utf-8') digest = hashlib.md5(digest_str).hexdigest() return digest
Hash the user agent and user agent password Section 3.10 of https://www.nar.realtor/retsorg.nsf/retsproto1.7d6.pdf :return: md5
entailment
def response_minify(self, response): """ minify response html to decrease traffic """ if response.content_type == u'text/html; charset=utf-8': endpoint = request.endpoint or '' view_func = current_app.view_functions.get(endpoint, None) name = ( '%s.%s' % (view_func.__module__, view_func.__name__) if view_func else '' ) if name in self._exempt_routes: return response response.direct_passthrough = False response.set_data( self._html_minify.minify(response.get_data(as_text=True)) ) return response return response
minify response html to decrease traffic
entailment
def exempt(self, obj): """ decorator to mark a view as exempt from htmlmin. """ name = '%s.%s' % (obj.__module__, obj.__name__) @wraps(obj) def __inner(*a, **k): return obj(*a, **k) self._exempt_routes.add(name) return __inner
decorator to mark a view as exempt from htmlmin.
entailment
def dmql(query): """Client supplied raw DMQL, ensure quote wrap.""" if isinstance(query, dict): raise ValueError("You supplied a dictionary to the dmql_query parameter, but a string is required." " Did you mean to pass this to the search_filter parameter? ") # automatically surround the given query with parentheses if it doesn't have them already if len(query) > 0 and query != "*" and query[0] != '(' and query[-1] != ')': query = '({})'.format(query) return query
Client supplied raw DMQL, ensure quote wrap.
entailment
def filter_to_dmql(filter_dict): """Converts the filter dictionary into DMQL""" if not isinstance(filter_dict, (dict, collections.OrderedDict)): raise TypeError('Expected a dictionary type buy got {} instead.'.format(type(filter_dict))) def is_date_time_type(val): """Returns True if the value is a datetime""" return isinstance(val, (datetime.datetime, datetime.date, datetime.time)) def evaluate_datetime(val): """Converts the datetime object into the RETS expected format""" date_format = '%Y-%m-%d' time_format = '%H:%M:%S' datetime_format = '{}T{}'.format(date_format, time_format) if isinstance(val, datetime.datetime): evaluated = val.strftime(datetime_format) elif isinstance(val, datetime.date): evaluated = val.strftime(date_format) elif isinstance(val, datetime.time): evaluated = val.strftime(time_format) else: evaluated = val return evaluated def evaluate_operators(key_dict): """Turns the custom filter operators into the expected RETS query""" allowed_operators = ['$gte', '$lte', '$contains', '$begins', '$ends', '$in', '$nin', '$neq'] # If key not in allowed_operators, assume it is a field name with the and operation. if not all(op in allowed_operators for op in key_dict.keys()): raise ValueError("You have supplied an invalid operator. " "Please provide one of the following {}".format(allowed_operators)) # We can have a single operator key, or the combination of gte/lte keys = key_dict.keys() string = '' # Search between two numbers or two dates if len(keys) == 2 and all(k in ['$gte', '$lte'] for k in keys): if all(is_date_time_type(key_dict[v]) for v in keys): # comparing dates string = '{}-{}'.format(evaluate_datetime(key_dict['$gte']), evaluate_datetime(key_dict['$lte'])) else: # comparing numbers try: float(key_dict['$gte']) float(key_dict['$lte']) except ValueError: raise ValueError("$gte and $lte expect numeric or datetime values") string = '{:.2f}-{:.2f}'.format(key_dict['$gte'], key_dict['$lte']) # Using a single operator key elif len(keys) == 1: if '$gte' in key_dict: if is_date_time_type(key_dict['$gte']): string = '{}+'.format(evaluate_datetime(key_dict['$gte'])) else: try: float(key_dict['$gte']) except ValueError: raise ValueError("$gte expects a numeric value or a datetime object") string = '{:.2f}+'.format(key_dict['$gte']) elif '$lte' in key_dict: if is_date_time_type(key_dict['$lte']): string = '{}-'.format(evaluate_datetime(key_dict['$lte'])) else: try: float(key_dict['$lte']) except ValueError: raise ValueError("$lte expects a numeric value or a datetime object") string = '{:.2f}-'.format(key_dict['$lte']) elif '$in' in key_dict: if not isinstance(key_dict['$in'], list): raise ValueError("$in expects a list of strings") key_dict['$in'] = [evaluate_datetime(v) for v in key_dict['$in']] if not all(isinstance(v, string_types) for v in key_dict['$in']): raise ValueError("$in expects a list of strings") options = ','.join(key_dict['$in']) string = '{}'.format(options) elif '$nin' in key_dict: if not isinstance(key_dict['$nin'], list): raise ValueError("$nin expects a list of strings") key_dict['$nin'] = [evaluate_datetime(v) for v in key_dict['$nin']] if not all(isinstance(v, string_types) for v in key_dict['$nin']): raise ValueError("$nin expects a list of strings") options = ','.join(key_dict['$nin']) string = '~{}'.format(options) elif '$contains' in key_dict: if not isinstance(key_dict['$contains'], string_types): raise ValueError("$contains expects a string.") string = '*{}*'.format(key_dict['$contains']) elif '$begins' in key_dict: if not isinstance(key_dict['$begins'], string_types): raise ValueError("$begins expects a string.") string = '{}*'.format(key_dict['$begins']) elif '$ends' in key_dict: if not isinstance(key_dict['$ends'], string_types): raise ValueError("$ends expects a string.") string = '*{}'.format(key_dict['$ends']) elif '$neq' in key_dict: string = '~{}'.format(key_dict['$neq']) else: # Provided too many or too few operators raise ValueError("Please supply $gte and $lte for getting values between numbers or 1 of {}".format( allowed_operators)) return string dmql_search_filters = [] for filt, value in filter_dict.items(): dmql_string = '({}='.format(filt) if isinstance(value, dict): # Applying an operator. This will need to be recursive because of the or possibility dmql_string += evaluate_operators(key_dict=value) else: # Simle equals statement dmql_string += '{}'.format(evaluate_datetime(value)) dmql_string += ')' dmql_search_filters.append(dmql_string) search_string = ','.join(dmql_search_filters) # Converts the filter dictionary to dmqp string logger.debug("Filter returned the following DMQL: {}".format(search_string)) return search_string
Converts the filter dictionary into DMQL
entailment
def get_attributes(input_dict): """ Get attributes of xml tags in input_dict and creates a dictionary with the attribute name as the key and the attribute value as the value :param input_dict: The xml tag with the attributes and values :return: dict """ return {k.lstrip("@"): v for k, v in input_dict.items() if k[0] == "@"}
Get attributes of xml tags in input_dict and creates a dictionary with the attribute name as the key and the attribute value as the value :param input_dict: The xml tag with the attributes and values :return: dict
entailment
def data_columns_to_dict(columns_string, dict_string, delimiter=None): """ Turns column names in a single string into a dictionary with the key being the column name and the value being the value in that column for each row :param columns_string: A string of column names :param dict_string: A string of values :param delimiter: The delimiter to use to split the column and values :return: dict """ if delimiter: return {k: v for k, v in zip(columns_string.split(delimiter), dict_string.split(delimiter))} else: return {k: v for k, v in zip(columns_string.split(), dict_string.split())}
Turns column names in a single string into a dictionary with the key being the column name and the value being the value in that column for each row :param columns_string: A string of column names :param dict_string: A string of values :param delimiter: The delimiter to use to split the column and values :return: dict
entailment
def analyze_reply_code(self, xml_response_dict): """ Checks the RETS Response Code and handles non-zero answers. :param xml_response_dict: :return: None """ if 'RETS-STATUS' in xml_response_dict: attributes = self.get_attributes(xml_response_dict['RETS-STATUS']) reply_code = attributes['ReplyCode'] reply_text = attributes.get('ReplyText', 'RETS did not supply a Reply Text.') logger.debug("Received ReplyCode of {0!s} from the RETS Server: {0!s}".format(reply_code, reply_text)) if reply_code != '0': raise RETSException(reply_text, reply_code) elif 'RETS' not in xml_response_dict: # pragma: no cover raise RETSException("The <RETS> tag was expected in the response XML but it was not found.") attributes = self.get_attributes(input_dict=xml_response_dict['RETS']) if 'ReplyCode' not in attributes: # pragma: no cover # The RETS server did not return a response code. return True reply_code = attributes['ReplyCode'] reply_text = attributes.get('ReplyText', 'RETS did not supply a Reply Text.') logger.debug("Received ReplyCode of {0!s} from the RETS Server: {0!s}".format(reply_code, reply_text)) if reply_code != '0': raise RETSException(reply_text, reply_code)
Checks the RETS Response Code and handles non-zero answers. :param xml_response_dict: :return: None
entailment
def ids(self, content_ids, object_ids): """Appends the content and object ids how RETS expects them""" result = [] content_ids = self.split(content_ids, False) object_ids = self.split(object_ids) for cid in content_ids: result.append('{}:{}'.format(cid, ':'.join(object_ids))) return result
Appends the content and object ids how RETS expects them
entailment
def split(value, dash_ranges=True): """Splits """ if isinstance(value, list): value = [str(v) for v in value] else: str_value = str(value) dash_matches = re.match(pattern='(\d+)\-(\d+)', string=str_value) if ':' in str_value or ',' in str_value: value = [v.strip() for v in str_value.replace(',', ':').split(':')] elif dash_ranges and dash_matches: start_range = int(dash_matches.group(1)) end_range = int(dash_matches.group(2)) + 1 rng = range(start_range, end_range) value = [str(r) for r in rng] else: value = [str_value] return value
Splits
entailment
def set_value(self, key, value): """ Set key value to the file. The fuction will be make the key and value to dictinary formate. If its exist then it will update the current new key value to the file. Arg: key : cache key value : cache value """ file_cache = self.read_file() if file_cache: file_cache[key] = value else: file_cache = {} file_cache[key] = value self.update_file(file_cache)
Set key value to the file. The fuction will be make the key and value to dictinary formate. If its exist then it will update the current new key value to the file. Arg: key : cache key value : cache value
entailment
def delete_value(self, key): """ Delete the key if the token is expired. Arg: key : cache key """ response = {} response['status'] = False response['msg'] = "key does not exist" file_cache = self.read_file() if key in file_cache: del file_cache[key] self.update_file(file_cache) response['status'] = True response['msg'] = "success" return response
Delete the key if the token is expired. Arg: key : cache key
entailment
def read_file(self): """ Open the file and assiging the permission to read/write and return the content in json formate. Return : json data """ file_obj = open(self.file, 'r') content = file_obj.read() file_obj.close() if content: content = json.loads(content) return content else: return {}
Open the file and assiging the permission to read/write and return the content in json formate. Return : json data
entailment
def update_file(self, content): """ It will convert json content to json string and update into file. Return: Boolean True/False """ updated_content = json.dumps(content) file_obj = open(self.file, 'r+') file_obj.write(str(updated_content)) file_obj.close() return True
It will convert json content to json string and update into file. Return: Boolean True/False
entailment
def parse(self, response, metadata_type): """ Parses RETS metadata using the COMPACT-DECODED format :param response: :param metadata_type: :return: """ xml = xmltodict.parse(response.text) self.analyze_reply_code(xml_response_dict=xml) base = xml.get('RETS', {}).get(metadata_type, {}) attributes = self.get_attributes(base) parsed = [] if base.get('System') or base.get('SYSTEM'): system_obj = {} if base.get('SYSTEM', {}).get('@SystemDescription'): system_obj['system_id'] = str(base['SYSTEM']['@SystemID']) if base.get('SYSTEM', {}).get('@SystemDescription'): system_obj['system_description'] = str(base['SYSTEM']['@SystemDescription']) if base.get('SYSTEM', {}).get('@TimeZoneOffset'): system_obj['timezone_offset'] = str(base['SYSTEM']['@TimeZoneOffset']) if base.get('SYSTEM', {}).get('Comments'): system_obj['comments'] = base['SYSTEM']['Comments'] if base.get('@Version'): system_obj['version'] = base['@Version'] parsed.append(system_obj) elif 'DATA' in base: if not isinstance(base['DATA'], list): # xmltodict could take single entry XML lists and turn them into str base['DATA'] = [base['DATA']] for data in base['DATA']: data_dict = self.data_columns_to_dict(columns_string=base.get('COLUMNS', ''), dict_string=data) data_dict.update(attributes) parsed.append(data_dict) return parsed
Parses RETS metadata using the COMPACT-DECODED format :param response: :param metadata_type: :return:
entailment
def parse(self, response, metadata_type): """ Parses RETS metadata using the STANDARD-XML format :param response: requests Response object :param metadata_type: string :return parsed: list """ xml = xmltodict.parse(response.text) self.analyze_reply_code(xml_response_dict=xml) base = xml.get('RETS', {}).get('METADATA', {}).get(metadata_type, {}) if metadata_type == 'METADATA-SYSTEM': syst = base.get('System', base.get('SYSTEM')) if not syst: raise ParseError("Could not get the System key from a METADATA-SYSTEM request.") system_obj = {} if syst.get('SystemID'): system_obj['system_id'] = str(syst['SystemID']) if syst.get('SystemDescription'): system_obj['system_description'] = str(syst['SystemDescription']) if syst.get('Comments'): system_obj['comments'] = syst['Comments'] if base.get('@Version'): system_obj['version'] = base['@Version'] return [system_obj] elif metadata_type == 'METADATA-CLASS': key = 'class' elif metadata_type == 'METADATA-RESOURCE': key = 'resource' elif metadata_type == 'METADATA-LOOKUP_TYPE': key = 'lookuptype' elif metadata_type == 'METADATA-OBJECT': key = 'object' elif metadata_type == 'METADATA-TABLE': key = 'field' else: msg = "Got an unknown metadata type of {0!s}".format(metadata_type) raise ParseError(msg) # Get the version with the right capitalization from the dictionary key_cap = None for k in base.keys(): if k.lower() == key: key_cap = k if not key_cap: msg = 'Could not find {0!s} in the response XML'.format(key) raise ParseError(msg) if isinstance(base[key_cap], list): return base[key_cap] else: return [base[key_cap]]
Parses RETS metadata using the STANDARD-XML format :param response: requests Response object :param metadata_type: string :return parsed: list
entailment
def _get_multiparts(response): """ From this 'multipart/parallel; boundary="874e43d27ec6d83f30f37841bdaf90c7"; charset=utf-8' get this --874e43d27ec6d83f30f37841bdaf90c7 """ boundary = None for part in response.headers.get('Content-Type', '').split(';'): if 'boundary=' in part: boundary = '--{}'.format(part.split('=', 1)[1].strip('\"')) break if not boundary: raise ParseError("Was not able to find the boundary between objects in a multipart response") if response.content is None: return [] response_string = response.content if six.PY3: # Python3 returns bytes, decode for string operations response_string = response_string.decode('latin-1') # help bad responses be more multipart compliant whole_body = response_string.strip('\r\n') no_front_boundary = whole_body.strip(boundary) # The boundary comes with some characters multi_parts = [] for part in no_front_boundary.split(boundary): multi_parts.append(part.strip('\r\n')) return multi_parts
From this 'multipart/parallel; boundary="874e43d27ec6d83f30f37841bdaf90c7"; charset=utf-8' get this --874e43d27ec6d83f30f37841bdaf90c7
entailment
def parse_image_response(self, response): """ Parse multiple objects from the RETS feed. A lot of string methods are used to handle the response before encoding it back into bytes for the object. :param response: The response from the feed :return: list of SingleObjectParser """ if 'xml' in response.headers.get('Content-Type'): # Got an XML response, likely an error code. xml = xmltodict.parse(response.text) self.analyze_reply_code(xml_response_dict=xml) multi_parts = self._get_multiparts(response) parsed = [] # go through each part of the multipart message for part in multi_parts: clean_part = part.strip('\r\n\r\n') if '\r\n\r\n' in clean_part: header, body = clean_part.split('\r\n\r\n', 1) else: header = clean_part body = None part_header_dict = {k.strip(): v.strip() for k, v in (h.split(':', 1) for h in header.split('\r\n'))} # Some multipart requests respond with a text/XML part stating an error if 'xml' in part_header_dict.get('Content-Type'): # Got an XML response, likely an error code. # Some rets servers give characters after the closing brace. body = body[:body.index('/>') + 2] if '/>' in body else body xml = xmltodict.parse(body) try: self.analyze_reply_code(xml_response_dict=xml) except RETSException as e: if e.reply_code == '20403': # The requested object_id was not found. continue raise e if body: obj = self._response_object_from_header( obj_head_dict=part_header_dict, content=body.encode('latin-1') if six.PY3 else body) else: obj = self._response_object_from_header(obj_head_dict=part_header_dict) parsed.append(obj) return parsed
Parse multiple objects from the RETS feed. A lot of string methods are used to handle the response before encoding it back into bytes for the object. :param response: The response from the feed :return: list of SingleObjectParser
entailment
def parse_image_response(self, response): """ Parse a single object from the RETS feed :param response: The response from the RETS server :return: Object """ if 'xml' in response.headers.get('Content-Type'): # Got an XML response, likely an error code. xml = xmltodict.parse(response.text) self.analyze_reply_code(xml_response_dict=xml) obj = self._response_object_from_header( obj_head_dict=response.headers, content=response.content) return obj
Parse a single object from the RETS feed :param response: The response from the RETS server :return: Object
entailment
def auth(self): """ Auth is used to call the AUTH API of CricketAPI. Access token required for every request call to CricketAPI. Auth functional will post user Cricket API app details to server and return the access token. Return: Access token """ if not self.store_handler.has_value('access_token'): params = {} params["access_key"] = self.access_key params["secret_key"] = self.secret_key params["app_id"] = self.app_id params["device_id"] = self.device_id auth_url = self.api_path + "auth/" response = self.get_response(auth_url, params, "post") if 'auth' in response: self.store_handler.set_value("access_token", response['auth']['access_token']) self.store_handler.set_value("expires", response['auth']['expires']) logger.info('Getting new access token') else: msg = "Error getting access_token, " + \ "please verify your access_key, secret_key and app_id" logger.error(msg) raise Exception("Auth Failed, please check your access details")
Auth is used to call the AUTH API of CricketAPI. Access token required for every request call to CricketAPI. Auth functional will post user Cricket API app details to server and return the access token. Return: Access token
entailment
def get_response(self, url, params={}, method="get"): """ It will return json response based on given url, params and methods. Arg: params: 'dictionary' url: 'url' format method: default 'get', support method 'post' Return: json data """ if method == "post": response_data = json.loads(requests.post(url, params=params).text) else: params["access_token"] = self.get_active_token() response_data = json.loads(requests.get(url, params=params).text) if not response_data['status_code'] == 200: if "status_msg" in response_data: logger.error("Bad response: " + response_data['status_msg']) else: logger.error("Some thing went wrong, please check your " + \ "request params Example: card_type and date") return response_data
It will return json response based on given url, params and methods. Arg: params: 'dictionary' url: 'url' format method: default 'get', support method 'post' Return: json data
entailment
def get_active_token(self): """ Getting the valid access token. Access token expires every 24 hours, It will expires then it will generate a new token. Return: active access token """ expire_time = self.store_handler.has_value("expires") access_token = self.store_handler.has_value("access_token") if expire_time and access_token: expire_time = self.store_handler.get_value("expires") if not datetime.now() < datetime.fromtimestamp(float(expire_time)): self.store_handler.delete_value("access_token") self.store_handler.delete_value("expires") logger.info('Access token expired, going to get new token') self.auth() else: logger.info('Access token noy expired yet') else: self.auth() return self.store_handler.get_value("access_token")
Getting the valid access token. Access token expires every 24 hours, It will expires then it will generate a new token. Return: active access token
entailment
def get_match(self, match_key, card_type="full_card"): """ Calling the Match API. Arg: match_key: key of the match card_type: optional, default to full_card. Accepted values are micro_card, summary_card & full_card. Return: json data """ match_url = self.api_path + "match/" + match_key + "/" params = {} params["card_type"] = card_type response = self.get_response(match_url, params) return response
Calling the Match API. Arg: match_key: key of the match card_type: optional, default to full_card. Accepted values are micro_card, summary_card & full_card. Return: json data
entailment
def get_recent_matches(self, card_type="micro_card"): """ Calling the Recent Matches API. Arg: card_type: optional, default to micro_card. Accepted values are micro_card & summary_card. Return: json data """ recent_matches_url = self.api_path + "recent_matches/" params = {} params["card_type"] = card_type response = self.get_response(recent_matches_url, params) return response
Calling the Recent Matches API. Arg: card_type: optional, default to micro_card. Accepted values are micro_card & summary_card. Return: json data
entailment
def get_player_stats(self, player_key, board_key): """ Calling the Player Stats API Args: player_key: Key of the player board_key: key of the board Return: json data """ player_stats_url = self.api_path + 'player/' + player_key + '/league/' + board_key + '/stats/' response = self.get_response(player_stats_url) return response
Calling the Player Stats API Args: player_key: Key of the player board_key: key of the board Return: json data
entailment
def get_ball_by_ball(self, match_key, over_key=None): """ match_key: key of the match over_key : key of the over Return: json data: """ if over_key: ball_by_ball_url = "{base_path}match/{match_key}/balls/{over_key}/".format(base_path=self.api_path, match_key=match_key, over_key=over_key) else: ball_by_ball_url = "{base_path}match/{match_key}/balls/".format(base_path=self.api_path, match_key=match_key) response = self.get_response(ball_by_ball_url) return response
match_key: key of the match over_key : key of the over Return: json data:
entailment
def get_recent_season_matches(self, season_key): """ Calling specific season recent matches. Arg: season_key: key of the season. Return: json date """ season_recent_matches_url = self.api_path + "season/" + season_key + "/recent_matches/" response = self.get_response(season_recent_matches_url) return response
Calling specific season recent matches. Arg: season_key: key of the season. Return: json date
entailment
def get_recent_seasons(self): """ Calling the Recent Season API. Return: json data """ recent_seasons_url = self.api_path + "recent_seasons/" response = self.get_response(recent_seasons_url) return response
Calling the Recent Season API. Return: json data
entailment
def get_schedule(self, date=None): """ Calling the Schedule API. Return: json data """ schedule_url = self.api_path + "schedule/" params = {} if date: params['date'] = date response = self.get_response(schedule_url, params) return response
Calling the Schedule API. Return: json data
entailment
def get_season_schedule(self, season_key): """ Calling specific season schedule Arg: season_key: key of the season Return: json data """ schedule_url = self.api_path + "season/" + season_key + "/schedule/" response = self.get_response(schedule_url) return response
Calling specific season schedule Arg: season_key: key of the season Return: json data
entailment
def get_season(self, season_key, card_type="micro_card"): """ Calling Season API. Arg: season_key: key of the season card_type: optional, default to micro_card. Accepted values are micro_card & summary_card Return: json data """ season_url = self.api_path + "season/" + season_key + "/" params = {} params["card_type"] = card_type response = self.get_response(season_url, params) return response
Calling Season API. Arg: season_key: key of the season card_type: optional, default to micro_card. Accepted values are micro_card & summary_card Return: json data
entailment
def get_season_stats(self, season_key): """ Calling Season Stats API. Arg: season_key: key of the season Return: json data """ season_stats_url = self.api_path + "season/" + season_key + "/stats/" response = self.get_response(season_stats_url) return response
Calling Season Stats API. Arg: season_key: key of the season Return: json data
entailment
def get_season_team(self, season_key, season_team_key,stats_type=None): """ Calling Season teams API Arg: season_key: key of the season Return: json data """ params = {"stats_type": stats_type} season_team_url = self.api_path + 'season/' + season_key + '/team/' + season_team_key + '/' response = self.get_response(season_team_url, params=params) return response
Calling Season teams API Arg: season_key: key of the season Return: json data
entailment
def get_season_points(self, season_key): """ Calling Season Points API. Arg: season_key: key of the season Return: json data """ season_points_url = self.api_path + "season/" + season_key + "/points/" response = self.get_response(season_points_url) return response
Calling Season Points API. Arg: season_key: key of the season Return: json data
entailment
def get_season_player_stats(self, season_key, player_key): """ Calling Season Player Stats API. Arg: season_key: key of the season player_key: key of the player Return: json data """ season_player_stats_url = self.api_path + "season/" + season_key + "/player/" + player_key + "/stats/" response = self.get_response(season_player_stats_url) return response
Calling Season Player Stats API. Arg: season_key: key of the season player_key: key of the player Return: json data
entailment
def get_overs_summary(self, match_key): """ Calling Overs Summary API Arg: match_key: key of the match Return: json data """ overs_summary_url = self.api_path + "match/" + match_key + "/overs_summary/" response = self.get_response(overs_summary_url) return response
Calling Overs Summary API Arg: match_key: key of the match Return: json data
entailment
def get_news_aggregation(self): """ Calling News Aggregation API Return: json data """ news_aggregation_url = self.api_path + "news_aggregation" + "/" response = self.get_response(news_aggregation_url) return response
Calling News Aggregation API Return: json data
entailment
def get_fantasy_credits(self, match_key): """ Calling Fantasy Credit API Arg: match_key: key of the match Return: json data """ fantasy_credit_url = self.api_path_v3 + "fantasy-match-credits/" + match_key + "/" response = self.get_response(fantasy_credit_url) return response
Calling Fantasy Credit API Arg: match_key: key of the match Return: json data
entailment
def get_fantasy_points(self, match_key): """ Calling Fantasy Points API Arg: match_key: key of the match Return: json data """ fantasy_points_url = self.api_path_v3 + "fantasy-match-points/" + match_key + "/" response = self.get_response(fantasy_points_url) return response
Calling Fantasy Points API Arg: match_key: key of the match Return: json data
entailment
def main(self): """ Generates an output string by replacing the keywords in the format string with the corresponding values from a submission dictionary. """ self.manage_submissions() out_string = self.options['format'] # Pop until we get something which len(title) <= max-chars length = float('inf') while length > self.options['max_chars']: self.selected_submission = self.submissions.pop() length = len(self.selected_submission['title']) for k, v in self.selected_submission.items(): out_string = out_string.replace(k, self.h.unescape(str(v))) return self.output(out_string, out_string)
Generates an output string by replacing the keywords in the format string with the corresponding values from a submission dictionary.
entailment
def login(self): """ Logs into Reddit in order to display a personalised front page. """ data = {'user': self.options['username'], 'passwd': self.options['password'], 'api_type': 'json'} response = self.client.post('http://www.reddit.com/api/login', data=data) self.client.modhash = response.json()['json']['data']['modhash']
Logs into Reddit in order to display a personalised front page.
entailment
def manage_submissions(self): """ If there are no or only one submissions left, get new submissions. This function manages URL creation and the specifics for front page or subreddit mode. """ if not hasattr(self, 'submissions') or len(self.submissions) == 1: self.submissions = [] if self.options['mode'] == 'front': # If there are no login details, the standard front # page will be displayed. if self.options['password'] and self.options['username']: self.login() url = 'http://reddit.com/.json?sort={0}'.format(self.options['sort']) self.submissions = self.get_submissions(url) elif self.options['mode'] == 'subreddit': for subreddit in self.options['subreddits']: url = 'http://reddit.com/r/{0}/.json?sort={1}'.format( subreddit, self.options['limit']) self.submissions += self.get_submissions(url) else: return
If there are no or only one submissions left, get new submissions. This function manages URL creation and the specifics for front page or subreddit mode.
entailment
def get_submissions(self, url): """ Connects to Reddit and gets a JSON representation of submissions. This JSON data is then processed and returned. url: A url that requests for submissions should be sent to. """ response = self.client.get(url, params={'limit': self.options['limit']}) submissions = [x['data'] for x in response.json()['data']['children']] return submissions
Connects to Reddit and gets a JSON representation of submissions. This JSON data is then processed and returned. url: A url that requests for submissions should be sent to.
entailment
def main(self): """ A compulsary function that gets the output of the cmus-remote -Q command and converts it to unicode in order for it to be processed and finally output. """ try: # Setting stderr to subprocess.STDOUT seems to stop the error # message returned by the process from being output to STDOUT. cmus_output = subprocess.check_output(['cmus-remote', '-Q'], stderr=subprocess.STDOUT).decode('utf-8') except subprocess.CalledProcessError: return self.output(None, None) if 'duration' in cmus_output: status = self.convert_cmus_output(cmus_output) out_string = self.options['format'] for k, v in status.items(): out_string = out_string.replace(k, v) else: out_string = None return self.output(out_string, out_string)
A compulsary function that gets the output of the cmus-remote -Q command and converts it to unicode in order for it to be processed and finally output.
entailment
def convert_cmus_output(self, cmus_output): """ Change the newline separated string of output data into a dictionary which can then be used to replace the strings in the config format. cmus_output: A string with information about cmus that is newline seperated. Running cmus-remote -Q in a terminal will show you what you're dealing with. """ cmus_output = cmus_output.split('\n') cmus_output = [x.replace('tag ', '') for x in cmus_output if not x in ''] cmus_output = [x.replace('set ', '') for x in cmus_output] status = {} partitioned = (item.partition(' ') for item in cmus_output) status = {item[0]: item[2] for item in partitioned} status['duration'] = self.convert_time(status['duration']) status['position'] = self.convert_time(status['position']) return status
Change the newline separated string of output data into a dictionary which can then be used to replace the strings in the config format. cmus_output: A string with information about cmus that is newline seperated. Running cmus-remote -Q in a terminal will show you what you're dealing with.
entailment
def convert_time(self, time): """ A helper function to convert seconds into hh:mm:ss for better readability. time: A string representing time in seconds. """ time_string = str(datetime.timedelta(seconds=int(time))) if time_string.split(':')[0] == '0': time_string = time_string.partition(':')[2] return time_string
A helper function to convert seconds into hh:mm:ss for better readability. time: A string representing time in seconds.
entailment
def output(self, full_text, short_text): """ Output all of the options and data for a segment. full_text: A string representing the data that should be output to i3bar. short_text: A more concise version of full_text, in case there is minimal room on the i3bar. """ full_text = full_text.replace('\n', '') short_text = short_text.replace('\n', '') self.output_options.update({'full_text': full_text, 'short_text': short_text}) self.output_options = {k: v for k, v in self.output_options.items() if v} return self.output_options
Output all of the options and data for a segment. full_text: A string representing the data that should be output to i3bar. short_text: A more concise version of full_text, in case there is minimal room on the i3bar.
entailment
def on_click(self, event): """ A function that should be overwritten by a plugin that wishes to react to events, if it wants to perform any action other than running the supplied command related to a button. event: A dictionary passed from i3bar (after being decoded from JSON) that has the folowing format: event = {'name': 'my_plugin', 'x': 231, 'y': 423} Note: It is also possible to have an instance key, but i3situation doesn't set it. """ if event['button'] == 1 and 'button1' in self.options: subprocess.call(self.options['button1'].split()) elif event['button'] == 2 and 'button2' in self.options: subprocess.call(self.options['button2'].split()) elif event['button'] == 3 and 'button3' in self.options: subprocess.call(self.options['button3'].split())
A function that should be overwritten by a plugin that wishes to react to events, if it wants to perform any action other than running the supplied command related to a button. event: A dictionary passed from i3bar (after being decoded from JSON) that has the folowing format: event = {'name': 'my_plugin', 'x': 231, 'y': 423} Note: It is also possible to have an instance key, but i3situation doesn't set it.
entailment
def url_reverse(request): """ Reverse the requested URL (passed via GET / POST as `url_name` parameter) :param request: Request object :return: The reversed path """ if request.method in ('GET', 'POST'): data = getattr(request, request.method) url_name = data.get('url_name') try: path = urls.reverse(url_name, args=data.getlist('args')) (view_func, args, kwargs) = urls.resolve(path) return http.HttpResponse(path, content_type='text/plain') except urls.NoReverseMatch: return http.HttpResponse('Error', content_type='text/plain') return http.HttpResponseNotAllowed(('GET', 'POST'))
Reverse the requested URL (passed via GET / POST as `url_name` parameter) :param request: Request object :return: The reversed path
entailment
def url_image(request, image_id, thumb_options=None, width=None, height=None): """ Converts a filer image ID in a complete path :param request: Request object :param image_id: Filer image ID :param thumb_options: ThumbnailOption ID :param width: user-provided width :param height: user-provided height :return: JSON serialized URL components ('url', 'width', 'height') """ image = File.objects.get(pk=image_id) if getattr(image, 'canonical_url', None): url = image.canonical_url else: url = image.url thumb = _return_thumbnail(image, thumb_options, width, height) if thumb: image = thumb url = image.url data = { 'url': url, 'width': image.width, 'height': image.height, } return http.HttpResponse(json.dumps(data), content_type='application/json')
Converts a filer image ID in a complete path :param request: Request object :param image_id: Filer image ID :param thumb_options: ThumbnailOption ID :param width: user-provided width :param height: user-provided height :return: JSON serialized URL components ('url', 'width', 'height')
entailment
def thumbnail_options(request): """ Returns the requested ThumbnailOption as JSON :param request: Request object :return: JSON serialized ThumbnailOption """ response_data = [{'id': opt.pk, 'name': opt.name} for opt in ThumbnailOption.objects.all()] return http.HttpResponse(json.dumps(response_data), content_type="application/json")
Returns the requested ThumbnailOption as JSON :param request: Request object :return: JSON serialized ThumbnailOption
entailment
def serve_image(request, image_id, thumb_options=None, width=None, height=None): """ returns the content of an image sized according to the parameters :param request: Request object :param image_id: Filer image ID :param thumb_options: ThumbnailOption ID :param width: user-provided width :param height: user-provided height :return: JSON serialized URL components ('url', 'width', 'height') """ image = File.objects.get(pk=image_id) if getattr(image, 'canonical_url', None): url = image.canonical_url else: url = image.url thumb = _return_thumbnail(image, thumb_options, width, height) if thumb: return server.serve(request, file_obj=thumb, save_as=False) else: return HttpResponseRedirect(url)
returns the content of an image sized according to the parameters :param request: Request object :param image_id: Filer image ID :param thumb_options: ThumbnailOption ID :param width: user-provided width :param height: user-provided height :return: JSON serialized URL components ('url', 'width', 'height')
entailment
def _touch_dir(self, path): """ A helper function to create a directory if it doesn't exist. path: A string containing a full path to the directory to be created. """ try: os.makedirs(path) except OSError as e: if e.errno != errno.EEXIST: raise
A helper function to create a directory if it doesn't exist. path: A string containing a full path to the directory to be created.
entailment
def reload(self): """ Reload the configuration from the file. This is in its own function so that it can be called at any time by another class. """ self._conf = configparser.ConfigParser() # Preserve the case of sections and keys. self._conf.optionxform = str self._conf.read(self.config_file_path) if 'general' not in self._conf.keys(): raise IncompleteConfigurationFile('Missing the general section') general = self._replace_data_types(dict(self._conf.items('general'))) self._conf.remove_section('general') plugin = [] for section in self._conf.sections(): plugin.append(dict(self._conf.items(section))) plugin[-1].update({'name': section}) plugin[-1] = self._replace_data_types(plugin[-1]) return (plugin, general)
Reload the configuration from the file. This is in its own function so that it can be called at any time by another class.
entailment
def _replace_data_types(dictionary): """ Replaces strings with appropriate data types (int, boolean). Also replaces the human readable logging levels with the integer form. dictionary: A dictionary returned from the config file. """ logging_levels = {'NONE': 0, 'NULL': 0, 'DEBUG': 10, 'INFO': 20, 'WARNING': 30, 'ERROR': 40, 'CRITICAL': 50} for k, v in dictionary.items(): if v in ['true', 'True', 'on']: dictionary[k] = True elif v in ['false', 'False', 'off']: dictionary[k] = False elif k == 'log_file' and '~' in v: dictionary[k] = v.replace('~', os.path.expanduser('~')) elif v in logging_levels: dictionary[k] = logging_levels[v] elif isinstance(v, str) and v.isnumeric(): dictionary[k] = int(v) elif ',' in v: dictionary[k] = [x.strip() for x in v.split(',')] return dictionary
Replaces strings with appropriate data types (int, boolean). Also replaces the human readable logging levels with the integer form. dictionary: A dictionary returned from the config file.
entailment
def main(): """Command line interface for the ``rotate-backups`` program.""" coloredlogs.install(syslog=True) # Command line option defaults. rotation_scheme = {} kw = dict(include_list=[], exclude_list=[]) parallel = False use_sudo = False # Internal state. selected_locations = [] # Parse the command line arguments. try: options, arguments = getopt.getopt(sys.argv[1:], 'M:H:d:w:m:y:I:x:jpri:c:r:uC:nvqh', [ 'minutely=', 'hourly=', 'daily=', 'weekly=', 'monthly=', 'yearly=', 'include=', 'exclude=', 'parallel', 'prefer-recent', 'relaxed', 'ionice=', 'config=', 'use-sudo', 'dry-run', 'removal-command=', 'verbose', 'quiet', 'help', ]) for option, value in options: if option in ('-M', '--minutely'): rotation_scheme['minutely'] = coerce_retention_period(value) elif option in ('-H', '--hourly'): rotation_scheme['hourly'] = coerce_retention_period(value) elif option in ('-d', '--daily'): rotation_scheme['daily'] = coerce_retention_period(value) elif option in ('-w', '--weekly'): rotation_scheme['weekly'] = coerce_retention_period(value) elif option in ('-m', '--monthly'): rotation_scheme['monthly'] = coerce_retention_period(value) elif option in ('-y', '--yearly'): rotation_scheme['yearly'] = coerce_retention_period(value) elif option in ('-I', '--include'): kw['include_list'].append(value) elif option in ('-x', '--exclude'): kw['exclude_list'].append(value) elif option in ('-j', '--parallel'): parallel = True elif option in ('-p', '--prefer-recent'): kw['prefer_recent'] = True elif option in ('-r', '--relaxed'): kw['strict'] = False elif option in ('-i', '--ionice'): value = validate_ionice_class(value.lower().strip()) kw['io_scheduling_class'] = value elif option in ('-c', '--config'): kw['config_file'] = parse_path(value) elif option in ('-u', '--use-sudo'): use_sudo = True elif option in ('-n', '--dry-run'): logger.info("Performing a dry run (because of %s option) ..", option) kw['dry_run'] = True elif option in ('-C', '--removal-command'): removal_command = shlex.split(value) logger.info("Using custom removal command: %s", removal_command) kw['removal_command'] = removal_command elif option in ('-v', '--verbose'): coloredlogs.increase_verbosity() elif option in ('-q', '--quiet'): coloredlogs.decrease_verbosity() elif option in ('-h', '--help'): usage(__doc__) return else: assert False, "Unhandled option! (programming error)" if rotation_scheme: logger.verbose("Rotation scheme defined on command line: %s", rotation_scheme) if arguments: # Rotation of the locations given on the command line. location_source = 'command line arguments' selected_locations.extend(coerce_location(value, sudo=use_sudo) for value in arguments) else: # Rotation of all configured locations. location_source = 'configuration file' selected_locations.extend( location for location, rotation_scheme, options in load_config_file(configuration_file=kw.get('config_file'), expand=True) ) # Inform the user which location(s) will be rotated. if selected_locations: logger.verbose("Selected %s based on %s:", pluralize(len(selected_locations), "location"), location_source) for number, location in enumerate(selected_locations, start=1): logger.verbose(" %i. %s", number, location) else: # Show the usage message when no directories are given nor configured. logger.verbose("No location(s) to rotate selected.") usage(__doc__) return except Exception as e: logger.error("%s", e) sys.exit(1) # Rotate the backups in the selected directories. program = RotateBackups(rotation_scheme, **kw) if parallel: program.rotate_concurrent(*selected_locations) else: for location in selected_locations: program.rotate_backups(location)
Command line interface for the ``rotate-backups`` program.
entailment
def get_battery_state(self, prop): """ Return the first line from the file located at battery_path/prop as a string. """ with open(os.path.join(self.options['battery_path'], prop), 'r') as f: return f.readline().strip()
Return the first line from the file located at battery_path/prop as a string.
entailment
def coerce_location(value, **options): """ Coerce a string to a :class:`Location` object. :param value: The value to coerce (a string or :class:`Location` object). :param options: Any keyword arguments are passed on to :func:`~executor.contexts.create_context()`. :returns: A :class:`Location` object. """ # Location objects pass through untouched. if not isinstance(value, Location): # Other values are expected to be strings. if not isinstance(value, string_types): msg = "Expected Location object or string, got %s instead!" raise ValueError(msg % type(value)) # Try to parse a remote location. ssh_alias, _, directory = value.partition(':') if ssh_alias and directory and '/' not in ssh_alias: options['ssh_alias'] = ssh_alias else: directory = value # Create the location object. value = Location( context=create_context(**options), directory=parse_path(directory), ) return value
Coerce a string to a :class:`Location` object. :param value: The value to coerce (a string or :class:`Location` object). :param options: Any keyword arguments are passed on to :func:`~executor.contexts.create_context()`. :returns: A :class:`Location` object.
entailment
def coerce_retention_period(value): """ Coerce a retention period to a Python value. :param value: A string containing the text 'always', a number or an expression that can be evaluated to a number. :returns: A number or the string 'always'. :raises: :exc:`~exceptions.ValueError` when the string can't be coerced. """ # Numbers pass through untouched. if not isinstance(value, numbers.Number): # Other values are expected to be strings. if not isinstance(value, string_types): msg = "Expected string, got %s instead!" raise ValueError(msg % type(value)) # Check for the literal string `always'. value = value.strip() if value.lower() == 'always': value = 'always' else: # Evaluate other strings as expressions. value = simple_eval(value) if not isinstance(value, numbers.Number): msg = "Expected numeric result, got %s instead!" raise ValueError(msg % type(value)) return value
Coerce a retention period to a Python value. :param value: A string containing the text 'always', a number or an expression that can be evaluated to a number. :returns: A number or the string 'always'. :raises: :exc:`~exceptions.ValueError` when the string can't be coerced.
entailment
def load_config_file(configuration_file=None, expand=True): """ Load a configuration file with backup directories and rotation schemes. :param configuration_file: Override the pathname of the configuration file to load (a string or :data:`None`). :param expand: :data:`True` to expand filename patterns to their matches, :data:`False` otherwise. :returns: A generator of tuples with four values each: 1. An execution context created using :mod:`executor.contexts`. 2. The pathname of a directory with backups (a string). 3. A dictionary with the rotation scheme. 4. A dictionary with additional options. :raises: :exc:`~exceptions.ValueError` when `configuration_file` is given but doesn't exist or can't be loaded. This function is used by :class:`RotateBackups` to discover user defined rotation schemes and by :mod:`rotate_backups.cli` to discover directories for which backup rotation is configured. When `configuration_file` isn't given :class:`~update_dotdee.ConfigLoader` is used to search for configuration files in the following locations: - ``/etc/rotate-backups.ini`` and ``/etc/rotate-backups.d/*.ini`` - ``~/.rotate-backups.ini`` and ``~/.rotate-backups.d/*.ini`` - ``~/.config/rotate-backups.ini`` and ``~/.config/rotate-backups.d/*.ini`` All of the available configuration files are loaded in the order given above, so that sections in user-specific configuration files override sections by the same name in system-wide configuration files. """ expand_notice_given = False if configuration_file: loader = ConfigLoader(available_files=[configuration_file], strict=True) else: loader = ConfigLoader(program_name='rotate-backups', strict=False) for section in loader.section_names: items = dict(loader.get_options(section)) context_options = {} if coerce_boolean(items.get('use-sudo')): context_options['sudo'] = True if items.get('ssh-user'): context_options['ssh_user'] = items['ssh-user'] location = coerce_location(section, **context_options) rotation_scheme = dict((name, coerce_retention_period(items[name])) for name in SUPPORTED_FREQUENCIES if name in items) options = dict(include_list=split(items.get('include-list', '')), exclude_list=split(items.get('exclude-list', '')), io_scheduling_class=items.get('ionice'), strict=coerce_boolean(items.get('strict', 'yes')), prefer_recent=coerce_boolean(items.get('prefer-recent', 'no'))) # Don't override the value of the 'removal_command' property unless the # 'removal-command' configuration file option has a value set. if items.get('removal-command'): options['removal_command'] = shlex.split(items['removal-command']) # Expand filename patterns? if expand and location.have_wildcards: logger.verbose("Expanding filename pattern %s on %s ..", location.directory, location.context) if location.is_remote and not expand_notice_given: logger.notice("Expanding remote filename patterns (may be slow) ..") expand_notice_given = True for match in sorted(location.context.glob(location.directory)): if location.context.is_directory(match): logger.verbose("Matched directory: %s", match) expanded = Location(context=location.context, directory=match) yield expanded, rotation_scheme, options else: logger.verbose("Ignoring match (not a directory): %s", match) else: yield location, rotation_scheme, options
Load a configuration file with backup directories and rotation schemes. :param configuration_file: Override the pathname of the configuration file to load (a string or :data:`None`). :param expand: :data:`True` to expand filename patterns to their matches, :data:`False` otherwise. :returns: A generator of tuples with four values each: 1. An execution context created using :mod:`executor.contexts`. 2. The pathname of a directory with backups (a string). 3. A dictionary with the rotation scheme. 4. A dictionary with additional options. :raises: :exc:`~exceptions.ValueError` when `configuration_file` is given but doesn't exist or can't be loaded. This function is used by :class:`RotateBackups` to discover user defined rotation schemes and by :mod:`rotate_backups.cli` to discover directories for which backup rotation is configured. When `configuration_file` isn't given :class:`~update_dotdee.ConfigLoader` is used to search for configuration files in the following locations: - ``/etc/rotate-backups.ini`` and ``/etc/rotate-backups.d/*.ini`` - ``~/.rotate-backups.ini`` and ``~/.rotate-backups.d/*.ini`` - ``~/.config/rotate-backups.ini`` and ``~/.config/rotate-backups.d/*.ini`` All of the available configuration files are loaded in the order given above, so that sections in user-specific configuration files override sections by the same name in system-wide configuration files.
entailment
def rotate_backups(directory, rotation_scheme, **options): """ Rotate the backups in a directory according to a flexible rotation scheme. .. note:: This function exists to preserve backwards compatibility with older versions of the `rotate-backups` package where all of the logic was exposed as a single function. Please refer to the documentation of the :class:`RotateBackups` initializer and the :func:`~RotateBackups.rotate_backups()` method for an explanation of this function's parameters. """ program = RotateBackups(rotation_scheme=rotation_scheme, **options) program.rotate_backups(directory)
Rotate the backups in a directory according to a flexible rotation scheme. .. note:: This function exists to preserve backwards compatibility with older versions of the `rotate-backups` package where all of the logic was exposed as a single function. Please refer to the documentation of the :class:`RotateBackups` initializer and the :func:`~RotateBackups.rotate_backups()` method for an explanation of this function's parameters.
entailment
def rotate_concurrent(self, *locations, **kw): """ Rotate the backups in the given locations concurrently. :param locations: One or more values accepted by :func:`coerce_location()`. :param kw: Any keyword arguments are passed on to :func:`rotate_backups()`. This function uses :func:`rotate_backups()` to prepare rotation commands for the given locations and then it removes backups in parallel, one backup per mount point at a time. The idea behind this approach is that parallel rotation is most useful when the files to be removed are on different disks and so multiple devices can be utilized at the same time. Because mount points are per system :func:`rotate_concurrent()` will also parallelize over backups located on multiple remote systems. """ timer = Timer() pool = CommandPool(concurrency=10) logger.info("Scanning %s ..", pluralize(len(locations), "backup location")) for location in locations: for cmd in self.rotate_backups(location, prepare=True, **kw): pool.add(cmd) if pool.num_commands > 0: backups = pluralize(pool.num_commands, "backup") logger.info("Preparing to rotate %s (in parallel) ..", backups) pool.run() logger.info("Successfully rotated %s in %s.", backups, timer)
Rotate the backups in the given locations concurrently. :param locations: One or more values accepted by :func:`coerce_location()`. :param kw: Any keyword arguments are passed on to :func:`rotate_backups()`. This function uses :func:`rotate_backups()` to prepare rotation commands for the given locations and then it removes backups in parallel, one backup per mount point at a time. The idea behind this approach is that parallel rotation is most useful when the files to be removed are on different disks and so multiple devices can be utilized at the same time. Because mount points are per system :func:`rotate_concurrent()` will also parallelize over backups located on multiple remote systems.
entailment
def rotate_backups(self, location, load_config=True, prepare=False): """ Rotate the backups in a directory according to a flexible rotation scheme. :param location: Any value accepted by :func:`coerce_location()`. :param load_config: If :data:`True` (so by default) the rotation scheme and other options can be customized by the user in a configuration file. In this case the caller's arguments are only used when the configuration file doesn't define a configuration for the location. :param prepare: If this is :data:`True` (not the default) then :func:`rotate_backups()` will prepare the required rotation commands without running them. :returns: A list with the rotation commands (:class:`~executor.ExternalCommand` objects). :raises: :exc:`~exceptions.ValueError` when the given location doesn't exist, isn't readable or isn't writable. The third check is only performed when dry run isn't enabled. This function binds the main methods of the :class:`RotateBackups` class together to implement backup rotation with an easy to use Python API. If you're using `rotate-backups` as a Python API and the default behavior is not satisfactory, consider writing your own :func:`rotate_backups()` function based on the underlying :func:`collect_backups()`, :func:`group_backups()`, :func:`apply_rotation_scheme()` and :func:`find_preservation_criteria()` methods. """ rotation_commands = [] location = coerce_location(location) # Load configuration overrides by user? if load_config: location = self.load_config_file(location) # Collect the backups in the given directory. sorted_backups = self.collect_backups(location) if not sorted_backups: logger.info("No backups found in %s.", location) return # Make sure the directory is writable. if not self.dry_run: location.ensure_writable() most_recent_backup = sorted_backups[-1] # Group the backups by the rotation frequencies. backups_by_frequency = self.group_backups(sorted_backups) # Apply the user defined rotation scheme. self.apply_rotation_scheme(backups_by_frequency, most_recent_backup.timestamp) # Find which backups to preserve and why. backups_to_preserve = self.find_preservation_criteria(backups_by_frequency) # Apply the calculated rotation scheme. for backup in sorted_backups: friendly_name = backup.pathname if not location.is_remote: # Use human friendly pathname formatting for local backups. friendly_name = format_path(backup.pathname) if backup in backups_to_preserve: matching_periods = backups_to_preserve[backup] logger.info("Preserving %s (matches %s retention %s) ..", friendly_name, concatenate(map(repr, matching_periods)), "period" if len(matching_periods) == 1 else "periods") else: logger.info("Deleting %s ..", friendly_name) if not self.dry_run: # Copy the list with the (possibly user defined) removal command. removal_command = list(self.removal_command) # Add the pathname of the backup as the final argument. removal_command.append(backup.pathname) # Construct the command object. command = location.context.prepare( command=removal_command, group_by=(location.ssh_alias, location.mount_point), ionice=self.io_scheduling_class, ) rotation_commands.append(command) if not prepare: timer = Timer() command.wait() logger.verbose("Deleted %s in %s.", friendly_name, timer) if len(backups_to_preserve) == len(sorted_backups): logger.info("Nothing to do! (all backups preserved)") return rotation_commands
Rotate the backups in a directory according to a flexible rotation scheme. :param location: Any value accepted by :func:`coerce_location()`. :param load_config: If :data:`True` (so by default) the rotation scheme and other options can be customized by the user in a configuration file. In this case the caller's arguments are only used when the configuration file doesn't define a configuration for the location. :param prepare: If this is :data:`True` (not the default) then :func:`rotate_backups()` will prepare the required rotation commands without running them. :returns: A list with the rotation commands (:class:`~executor.ExternalCommand` objects). :raises: :exc:`~exceptions.ValueError` when the given location doesn't exist, isn't readable or isn't writable. The third check is only performed when dry run isn't enabled. This function binds the main methods of the :class:`RotateBackups` class together to implement backup rotation with an easy to use Python API. If you're using `rotate-backups` as a Python API and the default behavior is not satisfactory, consider writing your own :func:`rotate_backups()` function based on the underlying :func:`collect_backups()`, :func:`group_backups()`, :func:`apply_rotation_scheme()` and :func:`find_preservation_criteria()` methods.
entailment
def load_config_file(self, location): """ Load a rotation scheme and other options from a configuration file. :param location: Any value accepted by :func:`coerce_location()`. :returns: The configured or given :class:`Location` object. """ location = coerce_location(location) for configured_location, rotation_scheme, options in load_config_file(self.config_file, expand=False): if configured_location.match(location): logger.verbose("Loading configuration for %s ..", location) if rotation_scheme: self.rotation_scheme = rotation_scheme for name, value in options.items(): if value: setattr(self, name, value) # Create a new Location object based on the directory of the # given location and the execution context of the configured # location, because: # # 1. The directory of the configured location may be a filename # pattern whereas we are interested in the expanded name. # # 2. The execution context of the given location may lack some # details of the configured location. return Location( context=configured_location.context, directory=location.directory, ) logger.verbose("No configuration found for %s.", location) return location
Load a rotation scheme and other options from a configuration file. :param location: Any value accepted by :func:`coerce_location()`. :returns: The configured or given :class:`Location` object.
entailment
def collect_backups(self, location): """ Collect the backups at the given location. :param location: Any value accepted by :func:`coerce_location()`. :returns: A sorted :class:`list` of :class:`Backup` objects (the backups are sorted by their date). :raises: :exc:`~exceptions.ValueError` when the given directory doesn't exist or isn't readable. """ backups = [] location = coerce_location(location) logger.info("Scanning %s for backups ..", location) location.ensure_readable() for entry in natsort(location.context.list_entries(location.directory)): match = TIMESTAMP_PATTERN.search(entry) if match: if self.exclude_list and any(fnmatch.fnmatch(entry, p) for p in self.exclude_list): logger.verbose("Excluded %s (it matched the exclude list).", entry) elif self.include_list and not any(fnmatch.fnmatch(entry, p) for p in self.include_list): logger.verbose("Excluded %s (it didn't match the include list).", entry) else: try: backups.append(Backup( pathname=os.path.join(location.directory, entry), timestamp=datetime.datetime(*(int(group, 10) for group in match.groups('0'))), )) except ValueError as e: logger.notice("Ignoring %s due to invalid date (%s).", entry, e) else: logger.debug("Failed to match time stamp in filename: %s", entry) if backups: logger.info("Found %i timestamped backups in %s.", len(backups), location) return sorted(backups)
Collect the backups at the given location. :param location: Any value accepted by :func:`coerce_location()`. :returns: A sorted :class:`list` of :class:`Backup` objects (the backups are sorted by their date). :raises: :exc:`~exceptions.ValueError` when the given directory doesn't exist or isn't readable.
entailment
def group_backups(self, backups): """ Group backups collected by :func:`collect_backups()` by rotation frequencies. :param backups: A :class:`set` of :class:`Backup` objects. :returns: A :class:`dict` whose keys are the names of rotation frequencies ('hourly', 'daily', etc.) and whose values are dictionaries. Each nested dictionary contains lists of :class:`Backup` objects that are grouped together because they belong into the same time unit for the corresponding rotation frequency. """ backups_by_frequency = dict((frequency, collections.defaultdict(list)) for frequency in SUPPORTED_FREQUENCIES) for b in backups: backups_by_frequency['minutely'][(b.year, b.month, b.day, b.hour, b.minute)].append(b) backups_by_frequency['hourly'][(b.year, b.month, b.day, b.hour)].append(b) backups_by_frequency['daily'][(b.year, b.month, b.day)].append(b) backups_by_frequency['weekly'][(b.year, b.week)].append(b) backups_by_frequency['monthly'][(b.year, b.month)].append(b) backups_by_frequency['yearly'][b.year].append(b) return backups_by_frequency
Group backups collected by :func:`collect_backups()` by rotation frequencies. :param backups: A :class:`set` of :class:`Backup` objects. :returns: A :class:`dict` whose keys are the names of rotation frequencies ('hourly', 'daily', etc.) and whose values are dictionaries. Each nested dictionary contains lists of :class:`Backup` objects that are grouped together because they belong into the same time unit for the corresponding rotation frequency.
entailment
def apply_rotation_scheme(self, backups_by_frequency, most_recent_backup): """ Apply the user defined rotation scheme to the result of :func:`group_backups()`. :param backups_by_frequency: A :class:`dict` in the format generated by :func:`group_backups()`. :param most_recent_backup: The :class:`~datetime.datetime` of the most recent backup. :raises: :exc:`~exceptions.ValueError` when the rotation scheme dictionary is empty (this would cause all backups to be deleted). .. note:: This method mutates the given data structure by removing all backups that should be removed to apply the user defined rotation scheme. """ if not self.rotation_scheme: raise ValueError("Refusing to use empty rotation scheme! (all backups would be deleted)") for frequency, backups in backups_by_frequency.items(): # Ignore frequencies not specified by the user. if frequency not in self.rotation_scheme: backups.clear() else: # Reduce the number of backups in each time slot of this # rotation frequency to a single backup (the oldest one or the # newest one). for period, backups_in_period in backups.items(): index = -1 if self.prefer_recent else 0 selected_backup = sorted(backups_in_period)[index] backups[period] = [selected_backup] # Check if we need to rotate away backups in old periods. retention_period = self.rotation_scheme[frequency] if retention_period != 'always': # Remove backups created before the minimum date of this # rotation frequency? (relative to the most recent backup) if self.strict: minimum_date = most_recent_backup - SUPPORTED_FREQUENCIES[frequency] * retention_period for period, backups_in_period in list(backups.items()): for backup in backups_in_period: if backup.timestamp < minimum_date: backups_in_period.remove(backup) if not backups_in_period: backups.pop(period) # If there are more periods remaining than the user # requested to be preserved we delete the oldest one(s). items_to_preserve = sorted(backups.items())[-retention_period:] backups_by_frequency[frequency] = dict(items_to_preserve)
Apply the user defined rotation scheme to the result of :func:`group_backups()`. :param backups_by_frequency: A :class:`dict` in the format generated by :func:`group_backups()`. :param most_recent_backup: The :class:`~datetime.datetime` of the most recent backup. :raises: :exc:`~exceptions.ValueError` when the rotation scheme dictionary is empty (this would cause all backups to be deleted). .. note:: This method mutates the given data structure by removing all backups that should be removed to apply the user defined rotation scheme.
entailment
def find_preservation_criteria(self, backups_by_frequency): """ Collect the criteria used to decide which backups to preserve. :param backups_by_frequency: A :class:`dict` in the format generated by :func:`group_backups()` which has been processed by :func:`apply_rotation_scheme()`. :returns: A :class:`dict` with :class:`Backup` objects as keys and :class:`list` objects containing strings (rotation frequencies) as values. """ backups_to_preserve = collections.defaultdict(list) for frequency, delta in ORDERED_FREQUENCIES: for period in backups_by_frequency[frequency].values(): for backup in period: backups_to_preserve[backup].append(frequency) return backups_to_preserve
Collect the criteria used to decide which backups to preserve. :param backups_by_frequency: A :class:`dict` in the format generated by :func:`group_backups()` which has been processed by :func:`apply_rotation_scheme()`. :returns: A :class:`dict` with :class:`Backup` objects as keys and :class:`list` objects containing strings (rotation frequencies) as values.
entailment
def mount_point(self): """ The pathname of the mount point of :attr:`directory` (a string or :data:`None`). If the ``stat --format=%m ...`` command that is used to determine the mount point fails, the value of this property defaults to :data:`None`. This enables graceful degradation on e.g. Mac OS X whose ``stat`` implementation is rather bare bones compared to GNU/Linux. """ try: return self.context.capture('stat', '--format=%m', self.directory, silent=True) except ExternalCommandFailed: return None
The pathname of the mount point of :attr:`directory` (a string or :data:`None`). If the ``stat --format=%m ...`` command that is used to determine the mount point fails, the value of this property defaults to :data:`None`. This enables graceful degradation on e.g. Mac OS X whose ``stat`` implementation is rather bare bones compared to GNU/Linux.
entailment
def ensure_exists(self): """Make sure the location exists.""" if not self.context.is_directory(self.directory): # This can also happen when we don't have permission to one of the # parent directories so we'll point that out in the error message # when it seems applicable (so as not to confuse users). if self.context.have_superuser_privileges: msg = "The directory %s doesn't exist!" raise ValueError(msg % self) else: raise ValueError(compact(""" The directory {location} isn't accessible, most likely because it doesn't exist or because of permissions. If you're sure the directory exists you can use the --use-sudo option. """, location=self))
Make sure the location exists.
entailment
def ensure_readable(self): """Make sure the location exists and is readable.""" self.ensure_exists() if not self.context.is_readable(self.directory): if self.context.have_superuser_privileges: msg = "The directory %s isn't readable!" raise ValueError(msg % self) else: raise ValueError(compact(""" The directory {location} isn't readable, most likely because of permissions. Consider using the --use-sudo option. """, location=self))
Make sure the location exists and is readable.
entailment
def ensure_writable(self): """Make sure the directory exists and is writable.""" self.ensure_exists() if not self.context.is_writable(self.directory): if self.context.have_superuser_privileges: msg = "The directory %s isn't writable!" raise ValueError(msg % self) else: raise ValueError(compact(""" The directory {location} isn't writable, most likely due to permissions. Consider using the --use-sudo option. """, location=self))
Make sure the directory exists and is writable.
entailment
def match(self, location): """ Check if the given location "matches". :param location: The :class:`Location` object to try to match. :returns: :data:`True` if the two locations are on the same system and the :attr:`directory` can be matched as a filename pattern or a literal match on the normalized pathname. """ if self.ssh_alias != location.ssh_alias: # Never match locations on other systems. return False elif self.have_wildcards: # Match filename patterns using fnmatch(). return fnmatch.fnmatch(location.directory, self.directory) else: # Compare normalized directory pathnames. self = os.path.normpath(self.directory) other = os.path.normpath(location.directory) return self == other
Check if the given location "matches". :param location: The :class:`Location` object to try to match. :returns: :data:`True` if the two locations are on the same system and the :attr:`directory` can be matched as a filename pattern or a literal match on the normalized pathname.
entailment
def setup_file_logger(filename, formatting, log_level): """ A helper function for creating a file logger. Accepts arguments, as it is used in Status and LoggingWriter. """ logger = logging.getLogger() # If a stream handler has been attached, remove it. if logger.handlers: logger.removeHandler(logger.handlers[0]) handler = logging.FileHandler(filename) logger.addHandler(handler) formatter = logging.Formatter(*formatting) handler.setFormatter(formatter) logger.setLevel(log_level) handler.setLevel(log_level) return logger
A helper function for creating a file logger. Accepts arguments, as it is used in Status and LoggingWriter.
entailment
def output_to_bar(self, message, comma=True): """ Outputs data to stdout, without buffering. message: A string containing the data to be output. comma: Whether or not a comma should be placed at the end of the output. """ if comma: message += ',' sys.stdout.write(message + '\n') sys.stdout.flush()
Outputs data to stdout, without buffering. message: A string containing the data to be output. comma: Whether or not a comma should be placed at the end of the output.
entailment
def reload(self): """ Reload the installed plugins and the configuration file. This is called when either the plugins or config get updated. """ logging.debug('Reloading config file as files have been modified.') self.config.plugin, self.config.general = self.config.reload() logging.debug('Reloading plugins as files have been modified.') self.loader = plugin_manager.PluginLoader( self._plugin_path, self.config.plugin) self._plugin_mod_time = os.path.getmtime(self._plugin_path) self._config_mod_time = os.path.getmtime(self._config_file_path)
Reload the installed plugins and the configuration file. This is called when either the plugins or config get updated.
entailment
def run_plugins(self): """ Creates a thread for each plugin and lets the thread_manager handle it. """ for obj in self.loader.objects: # Reserve a slot in the output_dict in order to ensure that the # items are in the correct order. self.output_dict[obj.output_options['name']] = None self.thread_manager.add_thread(obj.main, obj.options['interval'])
Creates a thread for each plugin and lets the thread_manager handle it.
entailment
def run(self): """ Monitors if the config file or plugins are updated. Also outputs the JSON data generated by the plugins, without needing to poll the threads. """ self.run_plugins() while True: # Reload plugins and config if either the config file or plugin # directory are modified. if self._config_mod_time != os.path.getmtime(self._config_file_path) or \ self._plugin_mod_time != os.path.getmtime(self._plugin_path): self.thread_manager.kill_all_threads() self.output_dict.clear() self.reload() self.run_plugins() self.output_to_bar(json.dumps(self._remove_empty_output())) time.sleep(self.config.general['interval'])
Monitors if the config file or plugins are updated. Also outputs the JSON data generated by the plugins, without needing to poll the threads.
entailment
def _remove_empty_output(self): """ If plugins haven't been initialised and therefore not sending output or their output is None, there is no reason to take up extra room on the bar. """ output = [] for key in self.output_dict: if self.output_dict[key] is not None and 'full_text' in self.output_dict[key]: output.append(self.output_dict[key]) return output
If plugins haven't been initialised and therefore not sending output or their output is None, there is no reason to take up extra room on the bar.
entailment