idx int64 0 63k | question stringlengths 53 5.28k | target stringlengths 5 805 |
|---|---|---|
18,200 | def find_biclique_embedding ( a , b , m , n = None , t = None , target_edges = None ) : _ , anodes = a _ , bnodes = b m , n , t , target_edges = _chimera_input ( m , n , t , target_edges ) embedding = processor ( target_edges , M = m , N = n , L = t ) . tightestNativeBiClique ( len ( anodes ) , len ( bnodes ) ) if not embedding : raise ValueError ( "cannot find a K{},{} embedding for given Chimera lattice" . format ( a , b ) ) left , right = embedding return dict ( zip ( anodes , left ) ) , dict ( zip ( bnodes , right ) ) | Find an embedding for a biclique in a Chimera graph . |
18,201 | def find_grid_embedding ( dim , m , n = None , t = 4 ) : m , n , t , target_edges = _chimera_input ( m , n , t , None ) indexer = dnx . generators . chimera . chimera_coordinates ( m , n , t ) dim = list ( dim ) num_dim = len ( dim ) if num_dim == 1 : def _key ( row , col , aisle ) : return row dim . extend ( [ 1 , 1 ] ) elif num_dim == 2 : def _key ( row , col , aisle ) : return row , col dim . append ( 1 ) elif num_dim == 3 : def _key ( row , col , aisle ) : return row , col , aisle else : raise ValueError ( "find_grid_embedding supports between one and three dimensions" ) rows , cols , aisles = dim if rows > m or cols > n or aisles > t : msg = ( "the largest grid that find_grid_embedding can fit in a ({}, {}, {}) Chimera-lattice " "is {}x{}x{}; given grid is {}x{}x{}" ) . format ( m , n , t , m , n , t , rows , cols , aisles ) raise ValueError ( msg ) return { _key ( row , col , aisle ) : [ indexer . int ( ( row , col , 0 , aisle ) ) , indexer . int ( ( row , col , 1 , aisle ) ) ] for row in range ( dim [ 0 ] ) for col in range ( dim [ 1 ] ) for aisle in range ( dim [ 2 ] ) } | Find an embedding for a grid in a Chimera graph . |
18,202 | def sample ( self , bqm , ** parameters ) : child = self . child cutoff = self . _cutoff cutoff_vartype = self . _cutoff_vartype comp = self . _comparison if cutoff_vartype is dimod . SPIN : original = bqm . spin else : original = bqm . binary new = type ( bqm ) ( original . linear , ( ( u , v , bias ) for ( u , v ) , bias in original . quadratic . items ( ) if not comp ( abs ( bias ) , cutoff ) ) , original . offset , original . vartype ) isolated = [ v for v in new if not new . adj [ v ] ] new . remove_variables_from ( isolated ) if isolated and len ( new ) == 0 : v = isolated . pop ( ) new . linear [ v ] = original . linear [ v ] sampleset = child . sample ( new , ** parameters ) . change_vartype ( bqm . vartype , inplace = True ) if isolated : samples , variables = _restore_isolated ( sampleset , bqm , isolated ) else : samples = sampleset . record . sample variables = sampleset . variables vectors = sampleset . data_vectors vectors . pop ( 'energy' ) return dimod . SampleSet . from_samples_bqm ( ( samples , variables ) , bqm , ** vectors ) | Cutoff and sample from the provided binary quadratic model . |
18,203 | def sample_poly ( self , poly , ** kwargs ) : child = self . child cutoff = self . _cutoff cutoff_vartype = self . _cutoff_vartype comp = self . _comparison if cutoff_vartype is dimod . SPIN : original = poly . to_spin ( copy = False ) else : original = poly . to_binary ( copy = False ) new = type ( poly ) ( ( ( term , bias ) for term , bias in original . items ( ) if len ( term ) > 1 and not comp ( abs ( bias ) , cutoff ) ) , cutoff_vartype ) for v in new . variables : term = v , if term in original : new [ term ] = original [ term ] isolated = list ( original . variables . difference ( new . variables ) ) if isolated and len ( new ) == 0 : term = isolated . pop ( ) , new [ term ] = original [ term ] sampleset = child . sample_poly ( new , ** kwargs ) . change_vartype ( poly . vartype , inplace = True ) if isolated : samples , variables = _restore_isolated_higherorder ( sampleset , poly , isolated ) else : samples = sampleset . record . sample variables = sampleset . variables vectors = sampleset . data_vectors vectors . pop ( 'energy' ) return dimod . SampleSet . from_samples_bqm ( ( samples , variables ) , poly , ** vectors ) | Cutoff and sample from the provided binary polynomial . |
18,204 | def diagnose_embedding ( emb , source , target ) : if not hasattr ( source , 'edges' ) : source = nx . Graph ( source ) if not hasattr ( target , 'edges' ) : target = nx . Graph ( target ) label = { } embedded = set ( ) for x in source : try : embx = emb [ x ] missing_chain = len ( embx ) == 0 except KeyError : missing_chain = True if missing_chain : yield MissingChainError , x continue all_present = True for q in embx : if label . get ( q , x ) != x : yield ChainOverlapError , q , x , label [ q ] elif q not in target : all_present = False yield InvalidNodeError , x , q else : label [ q ] = x if all_present : embedded . add ( x ) if not nx . is_connected ( target . subgraph ( embx ) ) : yield DisconnectedChainError , x yielded = nx . Graph ( ) for p , q in target . subgraph ( label ) . edges ( ) : yielded . add_edge ( label [ p ] , label [ q ] ) for x , y in source . edges ( ) : if x == y : continue if x in embedded and y in embedded and not yielded . has_edge ( x , y ) : yield MissingEdgeError , x , y | A detailed diagnostic for minor embeddings . |
18,205 | def model ( self , name = None , model = None , mask = None , ** kwargs ) : if isinstance ( model , ( flask_marshmallow . Schema , flask_marshmallow . base_fields . FieldABC ) ) : if not name : name = model . __class__ . __name__ api_model = Model ( name , model , mask = mask ) api_model . __apidoc__ = kwargs return self . add_model ( name , api_model ) return super ( Namespace , self ) . model ( name = name , model = model , ** kwargs ) | Model registration decorator . |
18,206 | def parameters ( self , parameters , locations = None ) : def decorator ( func ) : if locations is None and parameters . many : _locations = ( 'json' , ) else : _locations = locations if _locations is not None : parameters . context [ 'in' ] = _locations return self . doc ( params = parameters ) ( self . response ( code = HTTPStatus . UNPROCESSABLE_ENTITY ) ( self . WEBARGS_PARSER . use_args ( parameters , locations = _locations ) ( func ) ) ) return decorator | Endpoint parameters registration decorator . |
18,207 | def response ( self , model = None , code = HTTPStatus . OK , description = None , ** kwargs ) : code = HTTPStatus ( code ) if code is HTTPStatus . NO_CONTENT : assert model is None if model is None and code not in { HTTPStatus . ACCEPTED , HTTPStatus . NO_CONTENT } : if code . value not in http_exceptions . default_exceptions : raise ValueError ( "`model` parameter is required for code %d" % code ) model = self . model ( name = 'HTTPError%d' % code , model = DefaultHTTPErrorSchema ( http_code = code ) ) if description is None : description = code . description def response_serializer_decorator ( func ) : def dump_wrapper ( * args , ** kwargs ) : response = func ( * args , ** kwargs ) extra_headers = None if response is None : if model is not None : raise ValueError ( "Response cannot not be None with HTTP status %d" % code ) return flask . Response ( status = code ) elif isinstance ( response , flask . Response ) or model is None : return response elif isinstance ( response , tuple ) : response , _code , extra_headers = unpack ( response ) else : _code = code if HTTPStatus ( _code ) is code : response = model . dump ( response ) . data return response , _code , extra_headers return dump_wrapper def decorator ( func_or_class ) : if code . value in http_exceptions . default_exceptions : decorated_func_or_class = func_or_class elif isinstance ( func_or_class , type ) : func_or_class . _apply_decorator_to_methods ( response_serializer_decorator ) decorated_func_or_class = func_or_class else : decorated_func_or_class = wraps ( func_or_class ) ( response_serializer_decorator ( func_or_class ) ) if model is None : api_model = None else : if isinstance ( model , Model ) : api_model = model else : api_model = self . model ( model = model ) if getattr ( model , 'many' , False ) : api_model = [ api_model ] doc_decorator = self . doc ( responses = { code . value : ( description , api_model ) } ) return doc_decorator ( decorated_func_or_class ) return decorator | Endpoint response OpenAPI documentation decorator . |
18,208 | def _apply_decorator_to_methods ( cls , decorator ) : for method in cls . methods : method_name = method . lower ( ) decorated_method_func = decorator ( getattr ( cls , method_name ) ) setattr ( cls , method_name , decorated_method_func ) | This helper can apply a given decorator to all methods on the current Resource . |
18,209 | def options ( self , * args , ** kwargs ) : method_funcs = [ getattr ( self , m . lower ( ) ) for m in self . methods ] allowed_methods = [ ] request_oauth_backup = getattr ( flask . request , 'oauth' , None ) for method_func in method_funcs : if getattr ( method_func , '_access_restriction_decorators' , None ) : if not hasattr ( method_func , '_cached_fake_method_func' ) : fake_method_func = lambda * args , ** kwargs : True fake_method_func . __name__ = 'options' for decorator in method_func . _access_restriction_decorators : fake_method_func = decorator ( fake_method_func ) method_func . __dict__ [ '_cached_fake_method_func' ] = fake_method_func else : fake_method_func = method_func . _cached_fake_method_func flask . request . oauth = None try : fake_method_func ( self , * args , ** kwargs ) except HTTPException : continue allowed_methods . append ( method_func . __name__ . upper ( ) ) flask . request . oauth = request_oauth_backup return flask . Response ( status = HTTPStatus . NO_CONTENT , headers = { 'Allow' : ", " . join ( allowed_methods ) } ) | Check which methods are allowed . |
18,210 | def validate_patch_structure ( self , data ) : if data [ 'op' ] not in self . NO_VALUE_OPERATIONS and 'value' not in data : raise ValidationError ( 'value is required' ) if 'path' not in data : raise ValidationError ( 'Path is required and must always begin with /' ) else : data [ 'field_name' ] = data [ 'path' ] [ 1 : ] | Common validation of PATCH structure |
18,211 | def perform_patch ( cls , operations , obj , state = None ) : if state is None : state = { } for operation in operations : if not cls . _process_patch_operation ( operation , obj = obj , state = state ) : log . info ( "%s patching has been stopped because of unknown operation %s" , obj . __class__ . __name__ , operation ) raise ValidationError ( "Failed to update %s details. Operation %s could not succeed." % ( obj . __class__ . __name__ , operation ) ) return True | Performs all necessary operations by calling class methods with corresponding names . |
18,212 | def replace ( cls , obj , field , value , state ) : if not hasattr ( obj , field ) : raise ValidationError ( "Field '%s' does not exist, so it cannot be patched" % field ) setattr ( obj , field , value ) return True | This is method for replace operation . It is separated to provide a possibility to easily override it in your Parameters . |
18,213 | def __related_categories ( self , category_id ) : related = [ ] for cat in self . categories_tree : if category_id in self . categories_tree [ cat ] : related . append ( self . categories [ cat ] ) return related | Get all related categories to a given one |
18,214 | def _create_projects_file ( project_name , data_source , items ) : repositories = [ ] for item in items : if item [ 'origin' ] not in repositories : repositories . append ( item [ 'origin' ] ) projects = { project_name : { data_source : repositories } } projects_file , projects_file_path = tempfile . mkstemp ( prefix = 'track_items_' ) with open ( projects_file_path , "w" ) as pfile : json . dump ( projects , pfile , indent = True ) return projects_file_path | Create a projects file from the items origin data |
18,215 | def enrich_items ( self , ocean_backend , events = False ) : max_items = self . elastic . max_items_bulk current = 0 total = 0 bulk_json = "" items = ocean_backend . fetch ( ) images_items = { } url = self . elastic . index_url + '/items/_bulk' logger . debug ( "Adding items to %s (in %i packs)" , self . elastic . anonymize_url ( url ) , max_items ) for item in items : if current >= max_items : total += self . elastic . safe_put_bulk ( url , bulk_json ) json_size = sys . getsizeof ( bulk_json ) / ( 1024 * 1024 ) logger . debug ( "Added %i items to %s (%0.2f MB)" , total , self . elastic . anonymize_url ( url ) , json_size ) bulk_json = "" current = 0 rich_item = self . get_rich_item ( item ) data_json = json . dumps ( rich_item ) bulk_json += '{"index" : {"_id" : "%s" } }\n' % ( item [ self . get_field_unique_id ( ) ] ) bulk_json += data_json + "\n" current += 1 if rich_item [ 'id' ] not in images_items : rich_item [ 'is_docker_image' ] = 1 rich_item [ 'is_event' ] = 0 images_items [ rich_item [ 'id' ] ] = rich_item else : image_date = images_items [ rich_item [ 'id' ] ] [ 'last_updated' ] if image_date <= rich_item [ 'last_updated' ] : rich_item [ 'is_docker_image' ] = 1 rich_item [ 'is_event' ] = 0 images_items [ rich_item [ 'id' ] ] = rich_item if current > 0 : total += self . elastic . safe_put_bulk ( url , bulk_json ) if total == 0 : return total for image in images_items : data = images_items [ image ] data_json = json . dumps ( data ) bulk_json += '{"index" : {"_id" : "%s" } }\n' % ( data [ 'id' ] + "_image" ) bulk_json += data_json + "\n" total += self . elastic . safe_put_bulk ( url , bulk_json ) return total | A custom enrich items is needed because apart from the enriched events from raw items a image item with the last data for an image must be created |
18,216 | def get_owner_repos_url ( owner , token ) : url_org = GITHUB_API_URL + "/orgs/" + owner + "/repos" url_user = GITHUB_API_URL + "/users/" + owner + "/repos" url_owner = url_org try : r = requests . get ( url_org , params = get_payload ( ) , headers = get_headers ( token ) ) r . raise_for_status ( ) except requests . exceptions . HTTPError as e : if r . status_code == 403 : rate_limit_reset_ts = datetime . fromtimestamp ( int ( r . headers [ 'X-RateLimit-Reset' ] ) ) seconds_to_reset = ( rate_limit_reset_ts - datetime . utcnow ( ) ) . seconds + 1 logging . info ( "GitHub rate limit exhausted. Waiting %i secs for rate limit reset." % ( seconds_to_reset ) ) sleep ( seconds_to_reset ) else : url_owner = url_user return url_owner | The owner could be a org or a user . It waits if need to have rate limit . Also it fixes a djando issue changing - with _ |
18,217 | def get_repositores ( owner_url , token , nrepos ) : all_repos = [ ] url = owner_url while True : logging . debug ( "Getting repos from: %s" % ( url ) ) try : r = requests . get ( url , params = get_payload ( ) , headers = get_headers ( token ) ) r . raise_for_status ( ) all_repos += r . json ( ) logging . debug ( "Rate limit: %s" % ( r . headers [ 'X-RateLimit-Remaining' ] ) ) if 'next' not in r . links : break url = r . links [ 'next' ] [ 'url' ] except requests . exceptions . ConnectionError : logging . error ( "Can not connect to GitHub" ) break nrepos_recent = [ repo for repo in all_repos if not repo [ 'fork' ] ] nrepos_sorted = sorted ( nrepos_recent , key = lambda repo : parser . parse ( repo [ 'updated_at' ] ) , reverse = True ) nrepos_sorted = nrepos_sorted [ 0 : nrepos ] nrepos_sorted = sorted ( nrepos_sorted , key = lambda repo : repo [ 'size' ] ) for repo in nrepos_sorted : logging . debug ( "%s %i %s" % ( repo [ 'updated_at' ] , repo [ 'size' ] , repo [ 'name' ] ) ) return nrepos_sorted | owner could be an org or and user |
18,218 | def publish_twitter ( twitter_contact , owner ) : dashboard_url = CAULDRON_DASH_URL + "/%s" % ( owner ) tweet = "@%s your http://cauldron.io dashboard for #%s at GitHub is ready: %s. Check it out! #oscon" % ( twitter_contact , owner , dashboard_url ) status = quote_plus ( tweet ) oauth = get_oauth ( ) r = requests . post ( url = "https://api.twitter.com/1.1/statuses/update.json?status=" + status , auth = oauth ) | Publish in twitter the dashboard |
18,219 | def get_perceval_params_from_url ( cls , urls ) : params = [ ] dparam = cls . get_arthur_params_from_url ( urls ) params . append ( dparam [ "url" ] ) return params | Get the perceval params given the URLs for the data source |
18,220 | def add_identity ( cls , db , identity , backend ) : uuid = None try : uuid = api . add_identity ( db , backend , identity [ 'email' ] , identity [ 'name' ] , identity [ 'username' ] ) logger . debug ( "New sortinghat identity %s %s,%s,%s " , uuid , identity [ 'username' ] , identity [ 'name' ] , identity [ 'email' ] ) profile = { "name" : identity [ 'name' ] if identity [ 'name' ] else identity [ 'username' ] , "email" : identity [ 'email' ] } api . edit_profile ( db , uuid , ** profile ) except AlreadyExistsError as ex : uuid = ex . eid except InvalidValueError as ex : logger . warning ( "Trying to add a None identity. Ignoring it." ) except UnicodeEncodeError as ex : logger . warning ( "UnicodeEncodeError. Ignoring it. %s %s %s" , identity [ 'email' ] , identity [ 'name' ] , identity [ 'username' ] ) except Exception as ex : logger . warning ( "Unknown exception adding identity. Ignoring it. %s %s %s" , identity [ 'email' ] , identity [ 'name' ] , identity [ 'username' ] , exc_info = True ) if 'company' in identity and identity [ 'company' ] is not None : try : api . add_organization ( db , identity [ 'company' ] ) api . add_enrollment ( db , uuid , identity [ 'company' ] , datetime ( 1900 , 1 , 1 ) , datetime ( 2100 , 1 , 1 ) ) except AlreadyExistsError : pass return uuid | Load and identity list from backend in Sorting Hat |
18,221 | def add_identities ( cls , db , identities , backend ) : logger . info ( "Adding the identities to SortingHat" ) total = 0 for identity in identities : try : cls . add_identity ( db , identity , backend ) total += 1 except Exception as e : logger . error ( "Unexcepted error when adding identities: %s" % e ) continue logger . info ( "Total identities added to SH: %i" , total ) | Load identities list from backend in Sorting Hat |
18,222 | def remove_identity ( cls , sh_db , ident_id ) : success = False try : api . delete_identity ( sh_db , ident_id ) logger . debug ( "Identity %s deleted" , ident_id ) success = True except Exception as e : logger . debug ( "Identity not deleted due to %s" , str ( e ) ) return success | Delete an identity from SortingHat . |
18,223 | def remove_unique_identity ( cls , sh_db , uuid ) : success = False try : api . delete_unique_identity ( sh_db , uuid ) logger . debug ( "Unique identity %s deleted" , uuid ) success = True except Exception as e : logger . debug ( "Unique identity not deleted due to %s" , str ( e ) ) return success | Delete a unique identity from SortingHat . |
18,224 | def unique_identities ( cls , sh_db ) : try : for unique_identity in api . unique_identities ( sh_db ) : yield unique_identity except Exception as e : logger . debug ( "Unique identities not returned from SortingHat due to %s" , str ( e ) ) | List the unique identities available in SortingHat . |
18,225 | def get_rich_events ( self , item ) : module = item [ 'data' ] if not item [ 'data' ] [ 'releases' ] : return [ ] for release in item [ 'data' ] [ 'releases' ] : event = self . get_rich_item ( item ) event [ "uuid" ] += "_" + release [ 'slug' ] event [ "author_url" ] = 'https://forge.puppet.com/' + release [ 'module' ] [ 'owner' ] [ 'username' ] event [ "gravatar_id" ] = release [ 'module' ] [ 'owner' ] [ 'gravatar_id' ] event [ "downloads" ] = release [ 'downloads' ] event [ "slug" ] = release [ 'slug' ] event [ "version" ] = release [ 'version' ] event [ "uri" ] = release [ 'uri' ] event [ "validation_score" ] = release [ 'validation_score' ] event [ "homepage_url" ] = None if 'project_page' in release [ 'metadata' ] : event [ "homepage_url" ] = release [ 'metadata' ] [ 'project_page' ] event [ "issues_url" ] = None if "issues_url" in release [ 'metadata' ] : event [ "issues_url" ] = release [ 'metadata' ] [ 'issues_url' ] event [ "tags" ] = release [ 'tags' ] event [ "license" ] = release [ 'metadata' ] [ 'license' ] event [ "source_url" ] = release [ 'metadata' ] [ 'source' ] event [ "summary" ] = release [ 'metadata' ] [ 'summary' ] event [ "metadata__updated_on" ] = parser . parse ( release [ 'updated_at' ] ) . isoformat ( ) if self . sortinghat : release [ "metadata__updated_on" ] = event [ "metadata__updated_on" ] event . update ( self . get_item_sh ( release ) ) if self . prjs_map : event . update ( self . get_item_project ( event ) ) event . update ( self . get_grimoire_fields ( release [ "created_at" ] , "release" ) ) yield event | Get the enriched events related to a module |
18,226 | def _connect ( self ) : try : db = pymysql . connect ( user = self . user , passwd = self . passwd , host = self . host , port = self . port , db = self . shdb , use_unicode = True ) return db , db . cursor ( ) except Exception : logger . error ( "Database connection error" ) raise | Connect to the MySQL database . |
18,227 | def refresh_identities ( enrich_backend , author_field = None , author_values = None ) : def update_items ( new_filter_author ) : for eitem in enrich_backend . fetch ( new_filter_author ) : roles = None try : roles = enrich_backend . roles except AttributeError : pass new_identities = enrich_backend . get_item_sh_from_id ( eitem , roles ) eitem . update ( new_identities ) yield eitem logger . debug ( "Refreshing identities fields from %s" , enrich_backend . elastic . anonymize_url ( enrich_backend . elastic . index_url ) ) total = 0 max_ids = enrich_backend . elastic . max_items_clause logger . debug ( 'Refreshing identities' ) if author_field is None : for item in update_items ( None ) : yield item total += 1 else : to_refresh = [ ] for author_value in author_values : to_refresh . append ( author_value ) if len ( to_refresh ) > max_ids : filter_author = { "name" : author_field , "value" : to_refresh } for item in update_items ( filter_author ) : yield item total += 1 to_refresh = [ ] if len ( to_refresh ) > 0 : filter_author = { "name" : author_field , "value" : to_refresh } for item in update_items ( filter_author ) : yield item total += 1 logger . info ( "Total eitems refreshed for identities fields %i" , total ) | Refresh identities in enriched index . |
18,228 | def get_ocean_backend ( backend_cmd , enrich_backend , no_incremental , filter_raw = None , filter_raw_should = None ) : if no_incremental : last_enrich = None else : last_enrich = get_last_enrich ( backend_cmd , enrich_backend , filter_raw = filter_raw ) logger . debug ( "Last enrichment: %s" , last_enrich ) backend = None connector = get_connectors ( ) [ enrich_backend . get_connector_name ( ) ] if backend_cmd : backend_cmd = init_backend ( backend_cmd ) backend = backend_cmd . backend signature = inspect . signature ( backend . fetch ) if 'from_date' in signature . parameters : ocean_backend = connector [ 1 ] ( backend , from_date = last_enrich ) elif 'offset' in signature . parameters : ocean_backend = connector [ 1 ] ( backend , offset = last_enrich ) else : if last_enrich : ocean_backend = connector [ 1 ] ( backend , from_date = last_enrich ) else : ocean_backend = connector [ 1 ] ( backend ) else : params = enrich_backend . backend_params if params : try : date_pos = params . index ( '--from-date' ) last_enrich = parser . parse ( params [ date_pos + 1 ] ) except ValueError : pass if last_enrich : ocean_backend = connector [ 1 ] ( backend , from_date = last_enrich ) else : ocean_backend = connector [ 1 ] ( backend ) if filter_raw : ocean_backend . set_filter_raw ( filter_raw ) if filter_raw_should : ocean_backend . set_filter_raw_should ( filter_raw_should ) return ocean_backend | Get the ocean backend configured to start from the last enriched date |
18,229 | def do_studies ( ocean_backend , enrich_backend , studies_args , retention_time = None ) : for study in enrich_backend . studies : selected_studies = [ ( s [ 'name' ] , s [ 'params' ] ) for s in studies_args if s [ 'type' ] == study . __name__ ] for ( name , params ) in selected_studies : logger . info ( "Starting study: %s, params %s" , name , str ( params ) ) try : study ( ocean_backend , enrich_backend , ** params ) except Exception as e : logger . error ( "Problem executing study %s, %s" , name , str ( e ) ) raise e if name . startswith ( 'enrich_onion' ) : continue index_params = [ p for p in params if 'out_index' in p ] for ip in index_params : index_name = params [ ip ] elastic = get_elastic ( enrich_backend . elastic_url , index_name ) elastic . delete_items ( retention_time ) | Execute studies related to a given enrich backend . If retention_time is not None the study data is deleted based on the number of minutes declared in retention_time . |
18,230 | def delete_orphan_unique_identities ( es , sortinghat_db , current_data_source , active_data_sources ) : def get_uuids_in_index ( target_uuids ) : page = es . search ( index = IDENTITIES_INDEX , scroll = "360m" , size = SIZE_SCROLL_IDENTITIES_INDEX , body = { "query" : { "bool" : { "filter" : [ { "terms" : { "sh_uuid" : target_uuids } } ] } } } ) hits = [ ] if page [ 'hits' ] [ 'total' ] != 0 : hits = page [ 'hits' ] [ 'hits' ] return hits def delete_unique_identities ( target_uuids ) : count = 0 for uuid in target_uuids : success = SortingHat . remove_unique_identity ( sortinghat_db , uuid ) count = count + 1 if success else count return count def delete_identities ( unique_ident , data_sources ) : count = 0 for ident in unique_ident . identities : if ident . source not in data_sources : success = SortingHat . remove_identity ( sortinghat_db , ident . id ) count = count + 1 if success else count return count def has_identities_in_data_sources ( unique_ident , data_sources ) : in_active = False for ident in unique_ident . identities : if ident . source in data_sources : in_active = True break return in_active deleted_unique_identities = 0 deleted_identities = 0 uuids_to_process = [ ] for unique_identity in SortingHat . unique_identities ( sortinghat_db ) : if not has_identities_in_data_sources ( unique_identity , active_data_sources ) : deleted_unique_identities += delete_unique_identities ( [ unique_identity . uuid ] ) continue deleted_identities += delete_identities ( unique_identity , active_data_sources ) if not has_identities_in_data_sources ( unique_identity , [ current_data_source ] ) : continue uuids_to_process . append ( unique_identity . uuid ) if len ( uuids_to_process ) != SIZE_SCROLL_IDENTITIES_INDEX : continue results = get_uuids_in_index ( uuids_to_process ) uuids_found = [ item [ '_source' ] [ 'sh_uuid' ] for item in results ] orphan_uuids = set ( uuids_to_process ) - set ( uuids_found ) deleted_unique_identities += delete_unique_identities ( orphan_uuids ) uuids_to_process = [ ] if uuids_to_process : results = get_uuids_in_index ( uuids_to_process ) uuids_found = [ item [ '_source' ] [ 'sh_uuid' ] for item in results ] orphan_uuids = set ( uuids_to_process ) - set ( uuids_found ) deleted_unique_identities += delete_unique_identities ( orphan_uuids ) logger . debug ( "[identities retention] Total orphan unique identities deleted from SH: %i" , deleted_unique_identities ) logger . debug ( "[identities retention] Total identities in non-active data sources deleted from SH: %i" , deleted_identities ) | Delete all unique identities which appear in SortingHat but not in the IDENTITIES_INDEX . |
18,231 | def delete_inactive_unique_identities ( es , sortinghat_db , before_date ) : page = es . search ( index = IDENTITIES_INDEX , scroll = "360m" , size = SIZE_SCROLL_IDENTITIES_INDEX , body = { "query" : { "range" : { "last_seen" : { "lte" : before_date } } } } ) sid = page [ '_scroll_id' ] scroll_size = page [ 'hits' ] [ 'total' ] if scroll_size == 0 : logging . warning ( "[identities retention] No inactive identities found in %s after %s!" , IDENTITIES_INDEX , before_date ) return count = 0 while scroll_size > 0 : for item in page [ 'hits' ] [ 'hits' ] : to_delete = item [ '_source' ] [ 'sh_uuid' ] success = SortingHat . remove_unique_identity ( sortinghat_db , to_delete ) count = count + 1 if success else count page = es . scroll ( scroll_id = sid , scroll = '60m' ) sid = page [ '_scroll_id' ] scroll_size = len ( page [ 'hits' ] [ 'hits' ] ) logger . debug ( "[identities retention] Total inactive identities deleted from SH: %i" , count ) | Select the unique identities not seen before before_date and delete them from SortingHat . |
18,232 | def retain_identities ( retention_time , es_enrichment_url , sortinghat_db , data_source , active_data_sources ) : before_date = get_diff_current_date ( minutes = retention_time ) before_date_str = before_date . isoformat ( ) es = Elasticsearch ( [ es_enrichment_url ] , timeout = 120 , max_retries = 20 , retry_on_timeout = True , verify_certs = False ) delete_inactive_unique_identities ( es , sortinghat_db , before_date_str ) delete_orphan_unique_identities ( es , sortinghat_db , data_source , active_data_sources ) | Select the unique identities not seen before retention_time and delete them from SortingHat . Furthermore it deletes also the orphan unique identities those ones stored in SortingHat but not in IDENTITIES_INDEX . |
18,233 | def init_backend ( backend_cmd ) : try : backend_cmd . backend except AttributeError : parsed_args = vars ( backend_cmd . parsed_args ) init_args = find_signature_parameters ( backend_cmd . BACKEND , parsed_args ) backend_cmd . backend = backend_cmd . BACKEND ( ** init_args ) return backend_cmd | Init backend within the backend_cmd |
18,234 | def safe_index ( cls , unique_id ) : index = unique_id if unique_id : index = unique_id . replace ( "/" , "_" ) . lower ( ) return index | Return a valid elastic index generated from unique_id |
18,235 | def _check_instance ( url , insecure ) : res = grimoire_con ( insecure ) . get ( url ) if res . status_code != 200 : logger . error ( "Didn't get 200 OK from url %s" , url ) raise ElasticConnectException else : try : version_str = res . json ( ) [ 'version' ] [ 'number' ] version_major = version_str . split ( '.' ) [ 0 ] return version_major except Exception : logger . error ( "Could not read proper welcome message from url %s" , ElasticSearch . anonymize_url ( url ) ) logger . error ( "Message read: %s" , res . text ) raise ElasticConnectException | Checks if there is an instance of Elasticsearch in url . |
18,236 | def safe_put_bulk ( self , url , bulk_json ) : headers = { "Content-Type" : "application/x-ndjson" } try : res = self . requests . put ( url + '?refresh=true' , data = bulk_json , headers = headers ) res . raise_for_status ( ) except UnicodeEncodeError : logger . error ( "Encondig error ... converting bulk to iso-8859-1" ) bulk_json = bulk_json . encode ( 'iso-8859-1' , 'ignore' ) res = self . requests . put ( url , data = bulk_json , headers = headers ) res . raise_for_status ( ) result = res . json ( ) failed_items = [ ] if result [ 'errors' ] : failed_items = [ item [ 'index' ] for item in result [ 'items' ] if 'error' in item [ 'index' ] ] error = str ( failed_items [ 0 ] [ 'error' ] ) logger . error ( "Failed to insert data to ES: %s, %s" , error , self . anonymize_url ( url ) ) inserted_items = len ( result [ 'items' ] ) - len ( failed_items ) try : if failed_items : raise ELKError ( cause = error ) except ELKError : pass logger . debug ( "%i items uploaded to ES (%s)" , inserted_items , self . anonymize_url ( url ) ) return inserted_items | Bulk PUT controlling unicode issues |
18,237 | def all_es_aliases ( self ) : r = self . requests . get ( self . url + "/_aliases" , headers = HEADER_JSON , verify = False ) try : r . raise_for_status ( ) except requests . exceptions . HTTPError as ex : logger . warning ( "Something went wrong when retrieving aliases on %s." , self . anonymize_url ( self . index_url ) ) logger . warning ( ex ) return aliases = [ ] for index in r . json ( ) . keys ( ) : aliases . extend ( list ( r . json ( ) [ index ] [ 'aliases' ] . keys ( ) ) ) aliases = list ( set ( aliases ) ) return aliases | List all aliases used in ES |
18,238 | def list_aliases ( self ) : r = self . requests . get ( self . index_url + "/_alias" , headers = HEADER_JSON , verify = False ) try : r . raise_for_status ( ) except requests . exceptions . HTTPError as ex : logger . warning ( "Something went wrong when retrieving aliases on %s." , self . anonymize_url ( self . index_url ) ) logger . warning ( ex ) return aliases = r . json ( ) [ self . index ] [ 'aliases' ] return aliases | List aliases linked to the index |
18,239 | def bulk_upload ( self , items , field_id ) : current = 0 new_items = 0 bulk_json = "" if not items : return new_items url = self . index_url + '/items/_bulk' logger . debug ( "Adding items to %s (in %i packs)" , self . anonymize_url ( url ) , self . max_items_bulk ) task_init = time ( ) for item in items : if current >= self . max_items_bulk : task_init = time ( ) new_items += self . safe_put_bulk ( url , bulk_json ) current = 0 json_size = sys . getsizeof ( bulk_json ) / ( 1024 * 1024 ) logger . debug ( "bulk packet sent (%.2f sec, %i total, %.2f MB)" % ( time ( ) - task_init , new_items , json_size ) ) bulk_json = "" data_json = json . dumps ( item ) bulk_json += '{"index" : {"_id" : "%s" } }\n' % ( item [ field_id ] ) bulk_json += data_json + "\n" current += 1 if current > 0 : new_items += self . safe_put_bulk ( url , bulk_json ) json_size = sys . getsizeof ( bulk_json ) / ( 1024 * 1024 ) logger . debug ( "bulk packet sent (%.2f sec prev, %i total, %.2f MB)" % ( time ( ) - task_init , new_items , json_size ) ) return new_items | Upload in controlled packs items to ES using bulk API |
18,240 | def all_properties ( self ) : properties = { } r = self . requests . get ( self . index_url + "/_mapping" , headers = HEADER_JSON , verify = False ) try : r . raise_for_status ( ) r_json = r . json ( ) if 'items' not in r_json [ self . index ] [ 'mappings' ] : return properties if 'properties' not in r_json [ self . index ] [ 'mappings' ] [ 'items' ] : return properties properties = r_json [ self . index ] [ 'mappings' ] [ 'items' ] [ 'properties' ] except requests . exceptions . HTTPError as ex : logger . error ( "Error all attributes for %s." , self . anonymize_url ( self . index_url ) ) logger . error ( ex ) return return properties | Get all properties of a given index |
18,241 | def get_kibiter_version ( url ) : config_url = '.kibana/config/_search' if url [ - 1 ] != '/' : url += "/" url += config_url r = requests . get ( url ) r . raise_for_status ( ) if len ( r . json ( ) [ 'hits' ] [ 'hits' ] ) == 0 : logger . error ( "Can not get the Kibiter version" ) return None version = r . json ( ) [ 'hits' ] [ 'hits' ] [ 0 ] [ '_id' ] major_version = version . split ( "." , 1 ) [ 0 ] return major_version | Return kibiter major number version |
18,242 | def get_params ( ) : parser = get_params_parser ( ) args = parser . parse_args ( ) if not args . enrich_only and not args . only_identities and not args . only_studies : if not args . index : print ( "[error] --index <name> param is required when collecting items from raw" ) sys . exit ( 1 ) return args | Get params definition from ElasticOcean and from all the backends |
18,243 | def get_time_diff_days ( start_txt , end_txt ) : if start_txt is None or end_txt is None : return None start = parser . parse ( start_txt ) end = parser . parse ( end_txt ) seconds_day = float ( 60 * 60 * 24 ) diff_days = ( end - start ) . total_seconds ( ) / seconds_day diff_days = float ( '%.2f' % diff_days ) return diff_days | Number of days between two days |
18,244 | def enrich_fields ( cls , fields , eitem ) : for field in fields : if field . startswith ( 'customfield_' ) : if type ( fields [ field ] ) is dict : if 'name' in fields [ field ] : if fields [ field ] [ 'name' ] == "Story Points" : eitem [ 'story_points' ] = fields [ field ] [ 'value' ] elif fields [ field ] [ 'name' ] == "Sprint" : value = fields [ field ] [ 'value' ] if value : sprint = value [ 0 ] . partition ( ",name=" ) [ 2 ] . split ( ',' ) [ 0 ] sprint_start = value [ 0 ] . partition ( ",startDate=" ) [ 2 ] . split ( ',' ) [ 0 ] sprint_end = value [ 0 ] . partition ( ",endDate=" ) [ 2 ] . split ( ',' ) [ 0 ] sprint_complete = value [ 0 ] . partition ( ",completeDate=" ) [ 2 ] . split ( ',' ) [ 0 ] eitem [ 'sprint' ] = sprint eitem [ 'sprint_start' ] = cls . fix_value_null ( sprint_start ) eitem [ 'sprint_end' ] = cls . fix_value_null ( sprint_end ) eitem [ 'sprint_complete' ] = cls . fix_value_null ( sprint_complete ) | Enrich the fields property of an issue . |
18,245 | def get_review_sh ( self , revision , item ) : identity = self . get_sh_identity ( revision ) update = parser . parse ( item [ self . get_field_date ( ) ] ) erevision = self . get_item_sh_fields ( identity , update ) return erevision | Add sorting hat enrichment fields for the author of the revision |
18,246 | def get_github_cache ( self , kind , key_ ) : cache = { } res_size = 100 from_ = 0 index_github = "github/" + kind url = self . elastic . url + "/" + index_github url += "/_search" + "?" + "size=%i" % res_size r = self . requests . get ( url ) type_items = r . json ( ) if 'hits' not in type_items : logger . info ( "No github %s data in ES" % ( kind ) ) else : while len ( type_items [ 'hits' ] [ 'hits' ] ) > 0 : for hit in type_items [ 'hits' ] [ 'hits' ] : item = hit [ '_source' ] cache [ item [ key_ ] ] = item from_ += res_size r = self . requests . get ( url + "&from=%i" % from_ ) type_items = r . json ( ) if 'hits' not in type_items : break return cache | Get cache data for items of _type using key_ as the cache dict key |
18,247 | def get_time_to_first_attention ( self , item ) : comment_dates = [ str_to_datetime ( comment [ 'created_at' ] ) for comment in item [ 'comments_data' ] if item [ 'user' ] [ 'login' ] != comment [ 'user' ] [ 'login' ] ] reaction_dates = [ str_to_datetime ( reaction [ 'created_at' ] ) for reaction in item [ 'reactions_data' ] if item [ 'user' ] [ 'login' ] != reaction [ 'user' ] [ 'login' ] ] reaction_dates . extend ( comment_dates ) if reaction_dates : return min ( reaction_dates ) return None | Get the first date at which a comment or reaction was made to the issue by someone other than the user who created the issue |
18,248 | def get_time_to_merge_request_response ( self , item ) : review_dates = [ str_to_datetime ( review [ 'created_at' ] ) for review in item [ 'review_comments_data' ] if item [ 'user' ] [ 'login' ] != review [ 'user' ] [ 'login' ] ] if review_dates : return min ( review_dates ) return None | Get the first date at which a review was made on the PR by someone other than the user who created the PR |
18,249 | def get_rich_events ( self , item ) : if "version_downloads_data" not in item [ 'data' ] : return [ ] eitem = self . get_rich_item ( item ) for sample in item [ 'data' ] [ "version_downloads_data" ] [ "version_downloads" ] : event = deepcopy ( eitem ) event [ 'download_sample_id' ] = sample [ 'id' ] event [ 'sample_date' ] = sample [ 'date' ] sample_date = parser . parse ( event [ 'sample_date' ] ) event [ 'sample_version' ] = sample [ 'version' ] event [ 'sample_downloads' ] = sample [ 'downloads' ] event . update ( self . get_grimoire_fields ( sample_date . isoformat ( ) , "downloads_event" ) ) yield event | In the events there are some common fields with the crate . The name of the field must be the same in the create and in the downloads event so we can filer using it in crate and event at the same time . |
18,250 | def get_item_project ( self , eitem ) : project = None eitem_project = { } ds_name = self . get_connector_name ( ) if ds_name not in self . prjs_map : return eitem_project for tag in eitem [ 'hashtags_analyzed' ] : tags2project = CaseInsensitiveDict ( self . prjs_map [ ds_name ] ) if tag in tags2project : project = tags2project [ tag ] break if project is None : project = DEFAULT_PROJECT eitem_project = { "project" : project } eitem_project . update ( self . add_project_levels ( project ) ) return eitem_project | Get project mapping enrichment field . |
18,251 | def get_fields_from_job_name ( self , job_name ) : extra_fields = { 'category' : None , 'installer' : None , 'scenario' : None , 'testproject' : None , 'pod' : None , 'loop' : None , 'branch' : None } try : components = job_name . split ( '-' ) if len ( components ) < 2 : return extra_fields kind = components [ 1 ] if kind == 'os' : extra_fields [ 'category' ] = 'parent/main' extra_fields [ 'installer' ] = components [ 0 ] extra_fields [ 'scenario' ] = '-' . join ( components [ 2 : - 3 ] ) elif kind == 'deploy' : extra_fields [ 'category' ] = 'deploy' extra_fields [ 'installer' ] = components [ 0 ] else : extra_fields [ 'category' ] = 'test' extra_fields [ 'testproject' ] = components [ 0 ] extra_fields [ 'installer' ] = components [ 1 ] extra_fields [ 'pod' ] = components [ - 3 ] extra_fields [ 'loop' ] = components [ - 2 ] extra_fields [ 'branch' ] = components [ - 1 ] except IndexError as ex : logger . debug ( 'Problems parsing job name %s' , job_name ) logger . debug ( ex ) return extra_fields | Analyze a Jenkins job name producing a dictionary |
18,252 | def extract_builton ( self , built_on , regex ) : pattern = re . compile ( regex , re . M | re . I ) match = pattern . search ( built_on ) if match and len ( match . groups ( ) ) >= 1 : node_name = match . group ( 1 ) else : msg = "Node name not extracted, using builtOn as it is: " + regex + ":" + built_on logger . warning ( msg ) node_name = built_on return node_name | Extracts node name using a regular expression . Node name is expected to be group 1 . |
18,253 | def onion_study ( in_conn , out_conn , data_source ) : onion = OnionStudy ( in_connector = in_conn , out_connector = out_conn , data_source = data_source ) ndocs = onion . analyze ( ) return ndocs | Build and index for onion from a given Git index . |
18,254 | def read_block ( self , size = None , from_date = None ) : quarters = self . __quarters ( ) for quarter in quarters : logger . info ( self . __log_prefix + " Quarter: " + str ( quarter ) ) date_range = { self . _timeframe_field : { 'gte' : quarter . start_time , 'lte' : quarter . end_time } } orgs = self . __list_uniques ( date_range , self . AUTHOR_ORG ) projects = self . __list_uniques ( date_range , self . PROJECT ) s = self . __build_search ( date_range ) response = s . execute ( ) for timing in response . aggregations [ self . TIMEFRAME ] . buckets : yield self . __build_dataframe ( timing ) . copy ( ) for org_name in orgs : logger . info ( self . __log_prefix + " Quarter: " + str ( quarter ) + " Org: " + org_name ) s = self . __build_search ( date_range , org_name = org_name ) response = s . execute ( ) for timing in response . aggregations [ self . TIMEFRAME ] . buckets : yield self . __build_dataframe ( timing , org_name = org_name ) . copy ( ) for project in projects : logger . info ( self . __log_prefix + " Quarter: " + str ( quarter ) + " Project: " + project ) s = self . __build_search ( date_range , project_name = project ) response = s . execute ( ) for timing in response . aggregations [ self . TIMEFRAME ] . buckets : yield self . __build_dataframe ( timing , project_name = project ) . copy ( ) for org_name in orgs : logger . info ( self . __log_prefix + " Quarter: " + str ( quarter ) + " Project: " + project + " Org: " + org_name ) s = self . __build_search ( date_range , project_name = project , org_name = org_name ) response = s . execute ( ) for timing in response . aggregations [ self . TIMEFRAME ] . buckets : yield self . __build_dataframe ( timing , project_name = project , org_name = org_name ) . copy ( ) | Read author commits by Quarter Org and Project . |
18,255 | def __quarters ( self , from_date = None ) : s = Search ( using = self . _es_conn , index = self . _es_index ) if from_date : q = Q ( 'range' ) q . __setattr__ ( self . _sort_on_field , { 'gte' : from_date } ) s = s . filter ( q ) s = s [ 0 : 0 ] s . aggs . bucket ( self . TIMEFRAME , 'date_histogram' , field = self . _timeframe_field , interval = 'quarter' , min_doc_count = 1 ) response = s . execute ( ) quarters = [ ] for quarter in response . aggregations [ self . TIMEFRAME ] . buckets : period = pandas . Period ( quarter . key_as_string , 'Q' ) quarters . append ( period ) return quarters | Get a set of quarters with available items from a given index date . |
18,256 | def __list_uniques ( self , date_range , field_name ) : s = Search ( using = self . _es_conn , index = self . _es_index ) s = s . filter ( 'range' , ** date_range ) s = s [ 0 : 0 ] s . aggs . bucket ( 'uniques' , 'terms' , field = field_name , size = 1000 ) response = s . execute ( ) uniques_list = [ ] for item in response . aggregations . uniques . buckets : uniques_list . append ( item . key ) return uniques_list | Retrieve a list of unique values in a given field within a date range . |
18,257 | def __build_dataframe ( self , timing , project_name = None , org_name = None ) : date_list = [ ] uuid_list = [ ] name_list = [ ] contribs_list = [ ] latest_ts_list = [ ] logger . debug ( self . __log_prefix + " timing: " + timing . key_as_string ) for author in timing [ self . AUTHOR_UUID ] . buckets : latest_ts_list . append ( timing [ self . LATEST_TS ] . value_as_string ) date_list . append ( timing . key_as_string ) uuid_list . append ( author . key ) if author [ self . AUTHOR_NAME ] and author [ self . AUTHOR_NAME ] . buckets and len ( author [ self . AUTHOR_NAME ] . buckets ) > 0 : name_list . append ( author [ self . AUTHOR_NAME ] . buckets [ 0 ] . key ) else : name_list . append ( "Unknown" ) contribs_list . append ( author [ self . CONTRIBUTIONS ] . value ) df = pandas . DataFrame ( ) df [ self . TIMEFRAME ] = date_list df [ self . AUTHOR_UUID ] = uuid_list df [ self . AUTHOR_NAME ] = name_list df [ self . CONTRIBUTIONS ] = contribs_list df [ self . TIMESTAMP ] = latest_ts_list if not project_name : project_name = "_Global_" df [ self . PROJECT ] = project_name if not org_name : org_name = "_Global_" df [ self . AUTHOR_ORG ] = org_name return df | Build a DataFrame from a time bucket . |
18,258 | def process ( self , items_block ) : logger . info ( self . __log_prefix + " Authors to process: " + str ( len ( items_block ) ) ) onion_enrich = Onion ( items_block ) df_onion = onion_enrich . enrich ( member_column = ESOnionConnector . AUTHOR_UUID , events_column = ESOnionConnector . CONTRIBUTIONS ) df_onion [ 'quarter' ] = df_onion [ ESOnionConnector . TIMEFRAME ] . map ( lambda x : str ( pandas . Period ( x , 'Q' ) ) ) df_onion [ 'metadata__enriched_on' ] = datetime . utcnow ( ) . isoformat ( ) df_onion [ 'data_source' ] = self . data_source df_onion [ 'grimoire_creation_date' ] = df_onion [ ESOnionConnector . TIMEFRAME ] logger . info ( self . __log_prefix + " Final new events: " + str ( len ( df_onion ) ) ) return self . ProcessResults ( processed = len ( df_onion ) , out_items = df_onion ) | Process a DataFrame to compute Onion . |
18,259 | def get_projects ( self ) : repos_list = [ ] gerrit_projects_db = self . projects_db db = Database ( user = "root" , passwd = "" , host = "localhost" , port = 3306 , scrdb = None , shdb = gerrit_projects_db , prjdb = None ) sql = repos_list_raw = db . execute ( sql ) for repo in repos_list_raw : repo_name = repo [ 0 ] . replace ( self . repository + "_" , "" ) repos_list . append ( repo_name ) return repos_list | Get the projects list from database |
18,260 | def metadata ( func ) : @ functools . wraps ( func ) def decorator ( self , * args , ** kwargs ) : eitem = func ( self , * args , ** kwargs ) metadata = { 'metadata__gelk_version' : self . gelk_version , 'metadata__gelk_backend_name' : self . __class__ . __name__ , 'metadata__enriched_on' : datetime_utcnow ( ) . isoformat ( ) } eitem . update ( metadata ) return eitem return decorator | Add metadata to an item . |
18,261 | def get_grimoire_fields ( self , creation_date , item_name ) : grimoire_date = None try : grimoire_date = str_to_datetime ( creation_date ) . isoformat ( ) except Exception as ex : pass name = "is_" + self . get_connector_name ( ) + "_" + item_name return { "grimoire_creation_date" : grimoire_date , name : 1 } | Return common grimoire fields for all data sources |
18,262 | def add_project_levels ( cls , project ) : eitem_path = '' eitem_project_levels = { } if project is not None : subprojects = project . split ( '.' ) for i in range ( 0 , len ( subprojects ) ) : if i > 0 : eitem_path += "." eitem_path += subprojects [ i ] eitem_project_levels [ 'project_' + str ( i + 1 ) ] = eitem_path return eitem_project_levels | Add project sub levels extra items |
18,263 | def get_item_metadata ( self , eitem ) : eitem_metadata = { } project = self . find_item_project ( eitem ) if project and 'meta' in self . json_projects [ project ] : meta_fields = self . json_projects [ project ] [ 'meta' ] if isinstance ( meta_fields , dict ) : eitem_metadata = { CUSTOM_META_PREFIX + "_" + field : value for field , value in meta_fields . items ( ) } return eitem_metadata | In the projects . json file inside each project there is a field called meta which has a dictionary with fields to be added to the enriched items for this project . |
18,264 | def get_domain ( self , identity ) : domain = None if identity [ 'email' ] : try : domain = identity [ 'email' ] . split ( "@" ) [ 1 ] except IndexError : pass return domain | Get the domain from a SH identity |
18,265 | def get_enrollment ( self , uuid , item_date ) : if item_date and item_date . tzinfo : item_date = ( item_date - item_date . utcoffset ( ) ) . replace ( tzinfo = None ) enrollments = self . get_enrollments ( uuid ) enroll = self . unaffiliated_group if enrollments : for enrollment in enrollments : if not item_date : enroll = enrollment . organization . name break elif item_date >= enrollment . start and item_date <= enrollment . end : enroll = enrollment . organization . name break return enroll | Get the enrollment for the uuid when the item was done |
18,266 | def __get_item_sh_fields_empty ( self , rol , undefined = False ) : empty_field = '' if not undefined else '-- UNDEFINED --' return { rol + "_id" : empty_field , rol + "_uuid" : empty_field , rol + "_name" : empty_field , rol + "_user_name" : empty_field , rol + "_domain" : empty_field , rol + "_gender" : empty_field , rol + "_gender_acc" : None , rol + "_org_name" : empty_field , rol + "_bot" : False } | Return a SH identity with all fields to empty_field |
18,267 | def get_item_sh_fields ( self , identity = None , item_date = None , sh_id = None , rol = 'author' ) : eitem_sh = self . __get_item_sh_fields_empty ( rol ) if identity : sh_ids = self . get_sh_ids ( identity , self . get_connector_name ( ) ) eitem_sh [ rol + "_id" ] = sh_ids . get ( 'id' , '' ) eitem_sh [ rol + "_uuid" ] = sh_ids . get ( 'uuid' , '' ) eitem_sh [ rol + "_name" ] = identity . get ( 'name' , '' ) eitem_sh [ rol + "_user_name" ] = identity . get ( 'username' , '' ) eitem_sh [ rol + "_domain" ] = self . get_identity_domain ( identity ) elif sh_id : eitem_sh [ rol + "_id" ] = sh_id eitem_sh [ rol + "_uuid" ] = self . get_uuid_from_id ( sh_id ) else : return eitem_sh if rol + "_uuid" not in eitem_sh or not eitem_sh [ rol + "_uuid" ] : return self . __get_item_sh_fields_empty ( rol , undefined = True ) profile = self . get_profile_sh ( eitem_sh [ rol + "_uuid" ] ) if profile : eitem_sh [ rol + "_name" ] = profile . get ( 'name' , eitem_sh [ rol + "_name" ] ) email = profile . get ( 'email' , None ) if email : eitem_sh [ rol + "_domain" ] = self . get_email_domain ( email ) eitem_sh [ rol + "_gender" ] = profile . get ( 'gender' , self . unknown_gender ) eitem_sh [ rol + "_gender_acc" ] = profile . get ( 'gender_acc' , 0 ) elif not profile and sh_id : logger . warning ( "Can't find SH identity profile: %s" , sh_id ) if not eitem_sh . get ( rol + "_gender" ) : eitem_sh [ rol + "_gender" ] = self . unknown_gender eitem_sh [ rol + "_gender_acc" ] = 0 eitem_sh [ rol + "_org_name" ] = self . get_enrollment ( eitem_sh [ rol + "_uuid" ] , item_date ) eitem_sh [ rol + "_bot" ] = self . is_bot ( eitem_sh [ rol + '_uuid' ] ) return eitem_sh | Get standard SH fields from a SH identity |
18,268 | def get_item_sh ( self , item , roles = None , date_field = None ) : eitem_sh = { } author_field = self . get_field_author ( ) if not roles : roles = [ author_field ] if not date_field : item_date = str_to_datetime ( item [ self . get_field_date ( ) ] ) else : item_date = str_to_datetime ( item [ date_field ] ) users_data = self . get_users_data ( item ) for rol in roles : if rol in users_data : identity = self . get_sh_identity ( item , rol ) eitem_sh . update ( self . get_item_sh_fields ( identity , item_date , rol = rol ) ) if not eitem_sh [ rol + '_org_name' ] : eitem_sh [ rol + '_org_name' ] = SH_UNKNOWN_VALUE if not eitem_sh [ rol + '_name' ] : eitem_sh [ rol + '_name' ] = SH_UNKNOWN_VALUE if not eitem_sh [ rol + '_user_name' ] : eitem_sh [ rol + '_user_name' ] = SH_UNKNOWN_VALUE rol_author = 'author' if author_field in users_data and author_field != rol_author : identity = self . get_sh_identity ( item , author_field ) eitem_sh . update ( self . get_item_sh_fields ( identity , item_date , rol = rol_author ) ) if not eitem_sh [ 'author_org_name' ] : eitem_sh [ 'author_org_name' ] = SH_UNKNOWN_VALUE if not eitem_sh [ 'author_name' ] : eitem_sh [ 'author_name' ] = SH_UNKNOWN_VALUE if not eitem_sh [ 'author_user_name' ] : eitem_sh [ 'author_user_name' ] = SH_UNKNOWN_VALUE return eitem_sh | Add sorting hat enrichment fields for different roles |
18,269 | def get_sh_ids ( self , identity , backend_name ) : identity_tuple = tuple ( identity . items ( ) ) sh_ids = self . __get_sh_ids_cache ( identity_tuple , backend_name ) return sh_ids | Return the Sorting Hat id and uuid for an identity |
18,270 | def get_repository_filter_raw ( self , term = False ) : perceval_backend_name = self . get_connector_name ( ) filter_ = get_repository_filter ( self . perceval_backend , perceval_backend_name , term ) return filter_ | Returns the filter to be used in queries in a repository items |
18,271 | def set_filter_raw ( self , filter_raw ) : self . filter_raw = filter_raw self . filter_raw_dict = [ ] splitted = re . compile ( FILTER_SEPARATOR ) . split ( filter_raw ) for fltr_raw in splitted : fltr = self . __process_filter ( fltr_raw ) self . filter_raw_dict . append ( fltr ) | Filter to be used when getting items from Ocean index |
18,272 | def set_filter_raw_should ( self , filter_raw_should ) : self . filter_raw_should = filter_raw_should self . filter_raw_should_dict = [ ] splitted = re . compile ( FILTER_SEPARATOR ) . split ( filter_raw_should ) for fltr_raw in splitted : fltr = self . __process_filter ( fltr_raw ) self . filter_raw_should_dict . append ( fltr ) | Bool filter should to be used when getting items from Ocean index |
18,273 | def fetch ( self , _filter = None , ignore_incremental = False ) : logger . debug ( "Creating a elastic items generator." ) scroll_id = None page = self . get_elastic_items ( scroll_id , _filter = _filter , ignore_incremental = ignore_incremental ) if not page : return [ ] scroll_id = page [ "_scroll_id" ] scroll_size = page [ 'hits' ] [ 'total' ] if scroll_size == 0 : logger . warning ( "No results found from %s" , self . elastic . anonymize_url ( self . elastic . index_url ) ) return while scroll_size > 0 : logger . debug ( "Fetching from %s: %d received" , self . elastic . anonymize_url ( self . elastic . index_url ) , len ( page [ 'hits' ] [ 'hits' ] ) ) for item in page [ 'hits' ] [ 'hits' ] : eitem = item [ '_source' ] yield eitem page = self . get_elastic_items ( scroll_id , _filter = _filter , ignore_incremental = ignore_incremental ) if not page : break scroll_size = len ( page [ 'hits' ] [ 'hits' ] ) logger . debug ( "Fetching from %s: done receiving" , self . elastic . anonymize_url ( self . elastic . index_url ) ) | Fetch the items from raw or enriched index . An optional _filter could be provided to filter the data collected |
18,274 | def find_uuid ( es_url , index ) : uid_field = None res = requests . get ( '%s/%s/_search?size=1' % ( es_url , index ) ) first_item = res . json ( ) [ 'hits' ] [ 'hits' ] [ 0 ] [ '_source' ] fields = first_item . keys ( ) if 'uuid' in fields : uid_field = 'uuid' else : uuid_value = res . json ( ) [ 'hits' ] [ 'hits' ] [ 0 ] [ '_id' ] logging . debug ( "Finding unique id for %s with value %s" , index , uuid_value ) for field in fields : if first_item [ field ] == uuid_value : logging . debug ( "Found unique id for %s: %s" , index , field ) uid_field = field break if not uid_field : logging . error ( "Can not find uid field for %s. Can not copy the index." , index ) logging . error ( "Try to copy it directly with elasticdump or similar." ) sys . exit ( 1 ) return uid_field | Find the unique identifier field for a given index |
18,275 | def find_mapping ( es_url , index ) : mapping = None backend = find_perceval_backend ( es_url , index ) if backend : mapping = backend . get_elastic_mappings ( ) if mapping : logging . debug ( "MAPPING FOUND:\n%s" , json . dumps ( json . loads ( mapping [ 'items' ] ) , indent = True ) ) return mapping | Find the mapping given an index |
18,276 | def get_elastic_items ( elastic , elastic_scroll_id = None , limit = None ) : scroll_size = limit if not limit : scroll_size = DEFAULT_LIMIT if not elastic : return None url = elastic . index_url max_process_items_pack_time = "5m" url += "/_search?scroll=%s&size=%i" % ( max_process_items_pack_time , scroll_size ) if elastic_scroll_id : url = elastic . url url += "/_search/scroll" scroll_data = { "scroll" : max_process_items_pack_time , "scroll_id" : elastic_scroll_id } res = requests . post ( url , data = json . dumps ( scroll_data ) ) else : query = logging . debug ( "%s\n%s" , url , json . dumps ( json . loads ( query ) , indent = 4 ) ) res = requests . post ( url , data = query ) rjson = None try : rjson = res . json ( ) except Exception : logging . error ( "No JSON found in %s" , res . text ) logging . error ( "No results found from %s" , url ) return rjson | Get the items from the index |
18,277 | def fetch ( elastic , backend , limit = None , search_after_value = None , scroll = True ) : logging . debug ( "Creating a elastic items generator." ) elastic_scroll_id = None search_after = search_after_value while True : if scroll : rjson = get_elastic_items ( elastic , elastic_scroll_id , limit ) else : rjson = get_elastic_items_search ( elastic , search_after , limit ) if rjson and "_scroll_id" in rjson : elastic_scroll_id = rjson [ "_scroll_id" ] if rjson and "hits" in rjson : if not rjson [ "hits" ] [ "hits" ] : break for hit in rjson [ "hits" ] [ "hits" ] : item = hit [ '_source' ] if 'sort' in hit : search_after = hit [ 'sort' ] try : backend . _fix_item ( item ) except Exception : pass yield item else : logging . error ( "No results found from %s" , elastic . index_url ) break return | Fetch the items from raw or enriched index |
18,278 | def export_items ( elastic_url , in_index , out_index , elastic_url_out = None , search_after = False , search_after_value = None , limit = None , copy = False ) : if not limit : limit = DEFAULT_LIMIT if search_after_value : search_after_value_timestamp = int ( search_after_value [ 0 ] ) search_after_value_uuid = search_after_value [ 1 ] search_after_value = [ search_after_value_timestamp , search_after_value_uuid ] logging . info ( "Exporting items from %s/%s to %s" , elastic_url , in_index , out_index ) count_res = requests . get ( '%s/%s/_count' % ( elastic_url , in_index ) ) try : count_res . raise_for_status ( ) except requests . exceptions . HTTPError : if count_res . status_code == 404 : logging . error ( "The index does not exists: %s" , in_index ) else : logging . error ( count_res . text ) sys . exit ( 1 ) logging . info ( "Total items to copy: %i" , count_res . json ( ) [ 'count' ] ) elastic_in = ElasticSearch ( elastic_url , in_index ) if not copy : ds_mapping = find_mapping ( elastic_url , in_index ) else : logging . debug ( 'Using the input index mapping' ) ds_mapping = extract_mapping ( elastic_url , in_index ) if not elastic_url_out : elastic_out = ElasticSearch ( elastic_url , out_index , mappings = ds_mapping ) else : elastic_out = ElasticSearch ( elastic_url_out , out_index , mappings = ds_mapping ) uid_field = find_uuid ( elastic_url , in_index ) backend = find_perceval_backend ( elastic_url , in_index ) if search_after : total = elastic_out . bulk_upload ( fetch ( elastic_in , backend , limit , search_after_value , scroll = False ) , uid_field ) else : total = elastic_out . bulk_upload ( fetch ( elastic_in , backend , limit ) , uid_field ) logging . info ( "Total items copied: %i" , total ) | Export items from in_index to out_index using the correct mapping |
18,279 | def _fix_review_dates ( self , item ) : for date_field in [ 'timestamp' , 'createdOn' , 'lastUpdated' ] : if date_field in item . keys ( ) : date_ts = item [ date_field ] item [ date_field ] = unixtime_to_datetime ( date_ts ) . isoformat ( ) if 'patchSets' in item . keys ( ) : for patch in item [ 'patchSets' ] : pdate_ts = patch [ 'createdOn' ] patch [ 'createdOn' ] = unixtime_to_datetime ( pdate_ts ) . isoformat ( ) if 'approvals' in patch : for approval in patch [ 'approvals' ] : adate_ts = approval [ 'grantedOn' ] approval [ 'grantedOn' ] = unixtime_to_datetime ( adate_ts ) . isoformat ( ) if 'comments' in item . keys ( ) : for comment in item [ 'comments' ] : cdate_ts = comment [ 'timestamp' ] comment [ 'timestamp' ] = unixtime_to_datetime ( cdate_ts ) . isoformat ( ) | Convert dates so ES detect them |
18,280 | def get_sh_identity ( self , item , identity_field = None ) : def fill_list_identity ( identity , user_list_data ) : identity [ 'username' ] = user_list_data [ 0 ] [ '__text__' ] if '@' in identity [ 'username' ] : identity [ 'email' ] = identity [ 'username' ] if 'name' in user_list_data [ 0 ] : identity [ 'name' ] = user_list_data [ 0 ] [ 'name' ] return identity identity = { } for field in [ 'name' , 'email' , 'username' ] : identity [ field ] = None user = item if 'data' in item and type ( item ) == dict : user = item [ 'data' ] [ identity_field ] identity = fill_list_identity ( identity , user ) return identity | Return a Sorting Hat identity using bugzilla user data |
18,281 | def analyze ( self ) : from_date = self . _out . latest_date ( ) if from_date : logger . info ( "Reading items since " + from_date ) else : logger . info ( "Reading items since the beginning of times" ) cont = 0 total_processed = 0 total_written = 0 for item_block in self . _in . read_block ( size = self . _block_size , from_date = from_date ) : cont = cont + len ( item_block ) process_results = self . process ( item_block ) total_processed += process_results . processed if len ( process_results . out_items ) > 0 : self . _out . write ( process_results . out_items ) total_written += len ( process_results . out_items ) else : logger . info ( "No new items to be written this time." ) logger . info ( "Items read/to be written/total read/total processed/total written: " "{0}/{1}/{2}/{3}/{4}" . format ( str ( len ( item_block ) ) , str ( len ( process_results . out_items ) ) , str ( cont ) , str ( total_processed ) , str ( total_written ) ) ) logger . info ( "SUMMARY: Items total read/total processed/total written: " "{0}/{1}/{2}" . format ( str ( cont ) , str ( total_processed ) , str ( total_written ) ) ) logger . info ( "This is the end." ) return total_written | Populate an enriched index by processing input items in blocks . |
18,282 | def read_item ( self , from_date = None ) : search_query = self . _build_search_query ( from_date ) for hit in helpers . scan ( self . _es_conn , search_query , scroll = '300m' , index = self . _es_index , preserve_order = True ) : yield hit | Read items and return them one by one . |
18,283 | def read_block ( self , size , from_date = None ) : search_query = self . _build_search_query ( from_date ) hits_block = [ ] for hit in helpers . scan ( self . _es_conn , search_query , scroll = '300m' , index = self . _es_index , preserve_order = True ) : hits_block . append ( hit ) if len ( hits_block ) % size == 0 : yield hits_block hits_block = [ ] if len ( hits_block ) > 0 : yield hits_block | Read items and return them in blocks . |
18,284 | def write ( self , items ) : if self . _read_only : raise IOError ( "Cannot write, Connector created as Read Only" ) docs = [ ] for item in items : doc = { "_index" : self . _es_index , "_type" : "item" , "_id" : item [ "_id" ] , "_source" : item [ "_source" ] } docs . append ( doc ) helpers . bulk ( self . _es_conn , docs ) logger . info ( self . __log_prefix + " Written: " + str ( len ( docs ) ) ) | Upload items to ElasticSearch . |
18,285 | def create_alias ( self , alias_name ) : return self . _es_conn . indices . put_alias ( index = self . _es_index , name = alias_name ) | Creates an alias pointing to the index configured in this connection |
18,286 | def exists_alias ( self , alias_name , index_name = None ) : return self . _es_conn . indices . exists_alias ( index = index_name , name = alias_name ) | Check whether or not the given alias exists |
18,287 | def _build_search_query ( self , from_date ) : sort = [ { self . _sort_on_field : { "order" : "asc" } } ] filters = [ ] if self . _repo : filters . append ( { "term" : { "origin" : self . _repo } } ) if from_date : filters . append ( { "range" : { self . _sort_on_field : { "gte" : from_date } } } ) if filters : query = { "bool" : { "filter" : filters } } else : query = { "match_all" : { } } search_query = { "query" : query , "sort" : sort } return search_query | Build an ElasticSearch search query to retrieve items for read methods . |
18,288 | def add_params ( cls , cmdline_parser ) : parser = cmdline_parser parser . add_argument ( "-e" , "--elastic_url" , default = "http://127.0.0.1:9200" , help = "Host with elastic search (default: http://127.0.0.1:9200)" ) parser . add_argument ( "--elastic_url-enrich" , help = "Host with elastic search and enriched indexes" ) | Shared params in all backends |
18,289 | def get_p2o_params_from_url ( cls , url ) : if PRJ_JSON_FILTER_SEPARATOR not in url : return { "url" : url } params = { 'url' : url . split ( ' ' , 1 ) [ 0 ] } tokens = url . split ( PRJ_JSON_FILTER_SEPARATOR ) [ 1 : ] if len ( tokens ) > 1 : cause = "Too many filters defined for %s, only the first one is considered" % url logger . warning ( cause ) token = tokens [ 0 ] filter_tokens = token . split ( PRJ_JSON_FILTER_OP_ASSIGNMENT ) if len ( filter_tokens ) != 2 : cause = "Too many tokens after splitting for %s in %s" % ( token , url ) logger . error ( cause ) raise ELKError ( cause = cause ) fltr_name = filter_tokens [ 0 ] . strip ( ) fltr_value = filter_tokens [ 1 ] . strip ( ) params [ 'filter-' + fltr_name ] = fltr_value return params | Get the p2o params given a URL for the data source |
18,290 | def feed ( self , from_date = None , from_offset = None , category = None , latest_items = None , arthur_items = None , filter_classified = None ) : if self . fetch_archive : items = self . perceval_backend . fetch_from_archive ( ) self . feed_items ( items ) return elif arthur_items : items = arthur_items self . feed_items ( items ) return if from_date and from_offset : raise RuntimeError ( "Can't not feed using from_date and from_offset." ) filters_ = [ get_repository_filter ( self . perceval_backend , self . get_connector_name ( ) ) ] signature = inspect . signature ( self . perceval_backend . fetch ) last_update = None if 'from_date' in signature . parameters : if from_date : last_update = from_date else : self . last_update = self . get_last_update_from_es ( filters_ = filters_ ) last_update = self . last_update logger . info ( "Incremental from: %s" , last_update ) offset = None if 'offset' in signature . parameters : if from_offset : offset = from_offset else : offset = self . elastic . get_last_offset ( "offset" , filters_ = filters_ ) if offset is not None : logger . info ( "Incremental from: %i offset" , offset ) else : logger . info ( "Not incremental" ) params = { } if category is not None : params [ 'category' ] = category if filter_classified is not None : params [ 'filter_classified' ] = filter_classified if latest_items : params [ 'latest_items' ] = latest_items items = self . perceval_backend . fetch ( ** params ) elif last_update : last_update = last_update . replace ( tzinfo = None ) params [ 'from_date' ] = last_update items = self . perceval_backend . fetch ( ** params ) elif offset is not None : params [ 'offset' ] = offset items = self . perceval_backend . fetch ( ** params ) else : items = self . perceval_backend . fetch ( ** params ) self . feed_items ( items ) self . update_items ( ) | Feed data in Elastic from Perceval or Arthur |
18,291 | def get_identities ( self , item ) : def add_sh_github_identity ( user , user_field , rol ) : github_repo = None if GITHUB in item [ 'origin' ] : github_repo = item [ 'origin' ] . replace ( GITHUB , '' ) github_repo = re . sub ( '.git$' , '' , github_repo ) if not github_repo : return user_data = item [ 'data' ] [ user_field ] sh_identity = SortingHat . get_github_commit_username ( self . sh_db , user , SH_GIT_COMMIT ) if not sh_identity : gh_username = self . get_github_login ( user_data , rol , commit_hash , github_repo ) logger . debug ( "Adding new identity %s to SH %s: %s" , gh_username , SH_GIT_COMMIT , user ) user = self . get_sh_identity ( user_data ) user [ 'username' ] = gh_username SortingHat . add_identity ( self . sh_db , user , SH_GIT_COMMIT ) else : if user_data not in self . github_logins : self . github_logins [ user_data ] = sh_identity [ 'username' ] logger . debug ( "GitHub-commit exists. username:%s user:%s" , sh_identity [ 'username' ] , user_data ) commit_hash = item [ 'data' ] [ 'commit' ] if item [ 'data' ] [ 'Author' ] : m = self . AUTHOR_P2P_REGEX . match ( item [ 'data' ] [ "Author" ] ) n = self . AUTHOR_P2P_NEW_REGEX . match ( item [ 'data' ] [ "Author" ] ) if ( m or n ) and self . pair_programming : authors = self . __get_authors ( item [ 'data' ] [ "Author" ] ) for author in authors : user = self . get_sh_identity ( author ) yield user else : user = self . get_sh_identity ( item [ 'data' ] [ "Author" ] ) yield user if self . github_token : add_sh_github_identity ( user , 'Author' , 'author' ) if item [ 'data' ] [ 'Commit' ] : m = self . AUTHOR_P2P_REGEX . match ( item [ 'data' ] [ "Commit" ] ) n = self . AUTHOR_P2P_NEW_REGEX . match ( item [ 'data' ] [ "Author" ] ) if ( m or n ) and self . pair_programming : committers = self . __get_authors ( item [ 'data' ] [ 'Commit' ] ) for committer in committers : user = self . get_sh_identity ( committer ) yield user else : user = self . get_sh_identity ( item [ 'data' ] [ 'Commit' ] ) yield user if self . github_token : add_sh_github_identity ( user , 'Commit' , 'committer' ) if 'Signed-off-by' in item [ 'data' ] and self . pair_programming : signers = item [ 'data' ] [ "Signed-off-by" ] for signer in signers : user = self . get_sh_identity ( signer ) yield user | Return the identities from an item . If the repo is in GitHub get the usernames from GitHub . |
18,292 | def __fix_field_date ( self , item , attribute ) : field_date = str_to_datetime ( item [ attribute ] ) try : _ = int ( field_date . strftime ( "%z" ) [ 0 : 3 ] ) except ValueError : logger . warning ( "%s in commit %s has a wrong format" , attribute , item [ 'commit' ] ) item [ attribute ] = field_date . replace ( tzinfo = None ) . isoformat ( ) | Fix possible errors in the field date |
18,293 | def update_items ( self , ocean_backend , enrich_backend ) : fltr = { 'name' : 'origin' , 'value' : [ self . perceval_backend . origin ] } logger . debug ( "[update-items] Checking commits for %s." , self . perceval_backend . origin ) git_repo = GitRepository ( self . perceval_backend . uri , self . perceval_backend . gitpath ) try : current_hashes = set ( [ commit for commit in git_repo . rev_list ( ) ] ) except Exception as e : logger . error ( "Skip updating branch info for repo %s, git rev-list command failed: %s" , git_repo . uri , e ) return raw_hashes = set ( [ item [ 'data' ] [ 'commit' ] for item in ocean_backend . fetch ( ignore_incremental = True , _filter = fltr ) ] ) hashes_to_delete = list ( raw_hashes . difference ( current_hashes ) ) to_process = [ ] for _hash in hashes_to_delete : to_process . append ( _hash ) if len ( to_process ) != MAX_BULK_UPDATE_SIZE : continue self . remove_commits ( to_process , ocean_backend . elastic . index_url , 'data.commit' , self . perceval_backend . origin ) self . remove_commits ( to_process , enrich_backend . elastic . index_url , 'hash' , self . perceval_backend . origin ) to_process = [ ] if to_process : self . remove_commits ( to_process , ocean_backend . elastic . index_url , 'data.commit' , self . perceval_backend . origin ) self . remove_commits ( to_process , enrich_backend . elastic . index_url , 'hash' , self . perceval_backend . origin ) logger . debug ( "[update-items] %s commits deleted from %s with origin %s." , len ( hashes_to_delete ) , ocean_backend . elastic . anonymize_url ( ocean_backend . elastic . index_url ) , self . perceval_backend . origin ) logger . debug ( "[update-items] %s commits deleted from %s with origin %s." , len ( hashes_to_delete ) , enrich_backend . elastic . anonymize_url ( enrich_backend . elastic . index_url ) , self . perceval_backend . origin ) self . delete_commit_branches ( enrich_backend ) self . add_commit_branches ( git_repo , enrich_backend ) | Retrieve the commits not present in the original repository and delete the corresponding documents from the raw and enriched indexes |
18,294 | def add_commit_branches ( self , git_repo , enrich_backend ) : to_process = [ ] for hash , refname in git_repo . _discover_refs ( remote = True ) : if not refname . startswith ( 'refs/heads/' ) : continue commit_count = 0 branch_name = refname . replace ( 'refs/heads/' , '' ) try : commits = git_repo . rev_list ( [ branch_name ] ) for commit in commits : to_process . append ( commit ) commit_count += 1 if commit_count == MAX_BULK_UPDATE_SIZE : self . __process_commits_in_branch ( enrich_backend , branch_name , to_process ) to_process = [ ] commit_count = 0 if commit_count : self . __process_commits_in_branch ( enrich_backend , branch_name , to_process ) except Exception as e : logger . error ( "Skip adding branch info for repo %s due to %s" , git_repo . uri , e ) return | Add the information about branches to the documents representing commits in the enriched index . Branches are obtained using the command git ls - remote then for each branch the list of commits is retrieved via the command git rev - list branch - name and used to update the corresponding items in the enriched index . |
18,295 | def find_ds_mapping ( data_source , es_major_version ) : mappings = { "raw" : None , "enriched" : None } connectors = get_connectors ( ) try : raw_klass = connectors [ data_source ] [ 1 ] enrich_klass = connectors [ data_source ] [ 2 ] except KeyError : print ( "Data source not found" , data_source ) sys . exit ( 1 ) backend = raw_klass ( None ) if backend : mapping = json . loads ( backend . mapping . get_elastic_mappings ( es_major_version ) [ 'items' ] ) mappings [ 'raw' ] = [ mapping , find_general_mappings ( es_major_version ) ] backend = enrich_klass ( None ) if backend : mapping = json . loads ( backend . mapping . get_elastic_mappings ( es_major_version ) [ 'items' ] ) mappings [ 'enriched' ] = [ mapping , find_general_mappings ( es_major_version ) ] return mappings | Find the mapping given a perceval data source |
18,296 | def areas_of_code ( git_enrich , in_conn , out_conn , block_size = 100 ) : aoc = AreasOfCode ( in_connector = in_conn , out_connector = out_conn , block_size = block_size , git_enrich = git_enrich ) ndocs = aoc . analyze ( ) return ndocs | Build and index for areas of code from a given Perceval RAW index . |
18,297 | def process ( self , items_block ) : logger . info ( self . __log_prefix + " New commits: " + str ( len ( items_block ) ) ) git_events = Git ( items_block , self . _git_enrich ) events_df = git_events . eventize ( 2 ) logger . info ( self . __log_prefix + " New events: " + str ( len ( events_df ) ) ) if len ( events_df ) > 0 : data_filtered = FilterRows ( events_df ) events_df = data_filtered . filter_ ( [ "filepath" ] , "-" ) logger . info ( self . __log_prefix + " New events filtered: " + str ( len ( events_df ) ) ) events_df [ 'message' ] = events_df [ 'message' ] . str . slice ( stop = AreasOfCode . MESSAGE_MAX_SIZE ) logger . info ( self . __log_prefix + " Remove message content" ) enriched_filetype = FileType ( events_df ) events_df = enriched_filetype . enrich ( 'filepath' ) logger . info ( self . __log_prefix + " New Filetype events: " + str ( len ( events_df ) ) ) enriched_filepath = FilePath ( events_df ) events_df = enriched_filepath . enrich ( 'filepath' ) logger . info ( self . __log_prefix + " New Filepath events: " + str ( len ( events_df ) ) ) convert = ToUTF8 ( events_df ) events_df = convert . enrich ( [ "owner" ] ) logger . info ( self . __log_prefix + " Final new events: " + str ( len ( events_df ) ) ) return self . ProcessResults ( processed = len ( events_df ) , out_items = events_df ) | Process items to add file related information . |
18,298 | def get_time_diff_days ( start , end ) : if start is None or end is None : return None if type ( start ) is not datetime . datetime : start = parser . parse ( start ) . replace ( tzinfo = None ) if type ( end ) is not datetime . datetime : end = parser . parse ( end ) . replace ( tzinfo = None ) seconds_day = float ( 60 * 60 * 24 ) diff_days = ( end - start ) . total_seconds ( ) / seconds_day diff_days = float ( '%.2f' % diff_days ) return diff_days | Number of days between two dates in UTC format |
18,299 | def __fill_phab_ids ( self , item ) : for p in item [ 'projects' ] : if p and 'name' in p and 'phid' in p : self . phab_ids_names [ p [ 'phid' ] ] = p [ 'name' ] if 'authorData' not in item [ 'fields' ] or not item [ 'fields' ] [ 'authorData' ] : return self . phab_ids_names [ item [ 'fields' ] [ 'authorData' ] [ 'phid' ] ] = item [ 'fields' ] [ 'authorData' ] [ 'userName' ] if 'ownerData' in item [ 'fields' ] and item [ 'fields' ] [ 'ownerData' ] : self . phab_ids_names [ item [ 'fields' ] [ 'ownerData' ] [ 'phid' ] ] = item [ 'fields' ] [ 'ownerData' ] [ 'userName' ] if 'priority' in item [ 'fields' ] : val = item [ 'fields' ] [ 'priority' ] [ 'value' ] self . phab_ids_names [ str ( val ) ] = item [ 'fields' ] [ 'priority' ] [ 'name' ] for t in item [ 'transactions' ] : if 'authorData' in t and t [ 'authorData' ] and 'userName' in t [ 'authorData' ] : self . phab_ids_names [ t [ 'authorData' ] [ 'phid' ] ] = t [ 'authorData' ] [ 'userName' ] elif t [ 'authorData' ] and 'name' in t [ 'authorData' ] : self . phab_ids_names [ t [ 'authorData' ] [ 'phid' ] ] = t [ 'authorData' ] [ 'name' ] | Get mappings between phab ids and names |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.