partition
stringclasses
3 values
func_name
stringlengths
1
134
docstring
stringlengths
1
46.9k
path
stringlengths
4
223
original_string
stringlengths
75
104k
code
stringlengths
75
104k
docstring_tokens
listlengths
1
1.97k
repo
stringlengths
7
55
language
stringclasses
1 value
url
stringlengths
87
315
code_tokens
listlengths
19
28.4k
sha
stringlengths
40
40
valid
quadratic_weighted_kappa
Calculates kappa correlation between rater_a and rater_b. Kappa measures how well 2 quantities vary together. rater_a is a list of rater a scores rater_b is a list of rater b scores min_rating is an optional argument describing the minimum rating possible on the data set max_rating is an optional argument describing the maximum rating possible on the data set Returns a float corresponding to the kappa correlation
ease/util_functions.py
def quadratic_weighted_kappa(rater_a, rater_b, min_rating=None, max_rating=None): """ Calculates kappa correlation between rater_a and rater_b. Kappa measures how well 2 quantities vary together. rater_a is a list of rater a scores rater_b is a list of rater b scores min_rating is an optional argument describing the minimum rating possible on the data set max_rating is an optional argument describing the maximum rating possible on the data set Returns a float corresponding to the kappa correlation """ assert(len(rater_a) == len(rater_b)) rater_a = [int(a) for a in rater_a] rater_b = [int(b) for b in rater_b] if min_rating is None: min_rating = min(rater_a + rater_b) if max_rating is None: max_rating = max(rater_a + rater_b) conf_mat = confusion_matrix(rater_a, rater_b, min_rating, max_rating) num_ratings = len(conf_mat) num_scored_items = float(len(rater_a)) hist_rater_a = histogram(rater_a, min_rating, max_rating) hist_rater_b = histogram(rater_b, min_rating, max_rating) numerator = 0.0 denominator = 0.0 if(num_ratings > 1): for i in range(num_ratings): for j in range(num_ratings): expected_count = (hist_rater_a[i] * hist_rater_b[j] / num_scored_items) d = pow(i - j, 2.0) / pow(num_ratings - 1, 2.0) numerator += d * conf_mat[i][j] / num_scored_items denominator += d * expected_count / num_scored_items return 1.0 - numerator / denominator else: return 1.0
def quadratic_weighted_kappa(rater_a, rater_b, min_rating=None, max_rating=None): """ Calculates kappa correlation between rater_a and rater_b. Kappa measures how well 2 quantities vary together. rater_a is a list of rater a scores rater_b is a list of rater b scores min_rating is an optional argument describing the minimum rating possible on the data set max_rating is an optional argument describing the maximum rating possible on the data set Returns a float corresponding to the kappa correlation """ assert(len(rater_a) == len(rater_b)) rater_a = [int(a) for a in rater_a] rater_b = [int(b) for b in rater_b] if min_rating is None: min_rating = min(rater_a + rater_b) if max_rating is None: max_rating = max(rater_a + rater_b) conf_mat = confusion_matrix(rater_a, rater_b, min_rating, max_rating) num_ratings = len(conf_mat) num_scored_items = float(len(rater_a)) hist_rater_a = histogram(rater_a, min_rating, max_rating) hist_rater_b = histogram(rater_b, min_rating, max_rating) numerator = 0.0 denominator = 0.0 if(num_ratings > 1): for i in range(num_ratings): for j in range(num_ratings): expected_count = (hist_rater_a[i] * hist_rater_b[j] / num_scored_items) d = pow(i - j, 2.0) / pow(num_ratings - 1, 2.0) numerator += d * conf_mat[i][j] / num_scored_items denominator += d * expected_count / num_scored_items return 1.0 - numerator / denominator else: return 1.0
[ "Calculates", "kappa", "correlation", "between", "rater_a", "and", "rater_b", ".", "Kappa", "measures", "how", "well", "2", "quantities", "vary", "together", ".", "rater_a", "is", "a", "list", "of", "rater", "a", "scores", "rater_b", "is", "a", "list", "of",...
edx/ease
python
https://github.com/edx/ease/blob/a7890ed403da94d03726b0639cd8ebda45af6bbb/ease/util_functions.py#L345-L384
[ "def", "quadratic_weighted_kappa", "(", "rater_a", ",", "rater_b", ",", "min_rating", "=", "None", ",", "max_rating", "=", "None", ")", ":", "assert", "(", "len", "(", "rater_a", ")", "==", "len", "(", "rater_b", ")", ")", "rater_a", "=", "[", "int", "...
a7890ed403da94d03726b0639cd8ebda45af6bbb
valid
confusion_matrix
Generates a confusion matrix between rater_a and rater_b A confusion matrix shows how often 2 values agree and disagree See quadratic_weighted_kappa for argument descriptions
ease/util_functions.py
def confusion_matrix(rater_a, rater_b, min_rating=None, max_rating=None): """ Generates a confusion matrix between rater_a and rater_b A confusion matrix shows how often 2 values agree and disagree See quadratic_weighted_kappa for argument descriptions """ assert(len(rater_a) == len(rater_b)) rater_a = [int(a) for a in rater_a] rater_b = [int(b) for b in rater_b] min_rating = int(min_rating) max_rating = int(max_rating) if min_rating is None: min_rating = min(rater_a) if max_rating is None: max_rating = max(rater_a) num_ratings = int(max_rating - min_rating + 1) conf_mat = [[0 for i in range(num_ratings)] for j in range(num_ratings)] for a, b in zip(rater_a, rater_b): conf_mat[int(a - min_rating)][int(b - min_rating)] += 1 return conf_mat
def confusion_matrix(rater_a, rater_b, min_rating=None, max_rating=None): """ Generates a confusion matrix between rater_a and rater_b A confusion matrix shows how often 2 values agree and disagree See quadratic_weighted_kappa for argument descriptions """ assert(len(rater_a) == len(rater_b)) rater_a = [int(a) for a in rater_a] rater_b = [int(b) for b in rater_b] min_rating = int(min_rating) max_rating = int(max_rating) if min_rating is None: min_rating = min(rater_a) if max_rating is None: max_rating = max(rater_a) num_ratings = int(max_rating - min_rating + 1) conf_mat = [[0 for i in range(num_ratings)] for j in range(num_ratings)] for a, b in zip(rater_a, rater_b): conf_mat[int(a - min_rating)][int(b - min_rating)] += 1 return conf_mat
[ "Generates", "a", "confusion", "matrix", "between", "rater_a", "and", "rater_b", "A", "confusion", "matrix", "shows", "how", "often", "2", "values", "agree", "and", "disagree", "See", "quadratic_weighted_kappa", "for", "argument", "descriptions" ]
edx/ease
python
https://github.com/edx/ease/blob/a7890ed403da94d03726b0639cd8ebda45af6bbb/ease/util_functions.py#L387-L407
[ "def", "confusion_matrix", "(", "rater_a", ",", "rater_b", ",", "min_rating", "=", "None", ",", "max_rating", "=", "None", ")", ":", "assert", "(", "len", "(", "rater_a", ")", "==", "len", "(", "rater_b", ")", ")", "rater_a", "=", "[", "int", "(", "a...
a7890ed403da94d03726b0639cd8ebda45af6bbb
valid
histogram
Generates a frequency count of each rating on the scale ratings is a list of scores Returns a list of frequencies
ease/util_functions.py
def histogram(ratings, min_rating=None, max_rating=None): """ Generates a frequency count of each rating on the scale ratings is a list of scores Returns a list of frequencies """ ratings = [int(r) for r in ratings] if min_rating is None: min_rating = min(ratings) if max_rating is None: max_rating = max(ratings) num_ratings = int(max_rating - min_rating + 1) hist_ratings = [0 for x in range(num_ratings)] for r in ratings: hist_ratings[r - min_rating] += 1 return hist_ratings
def histogram(ratings, min_rating=None, max_rating=None): """ Generates a frequency count of each rating on the scale ratings is a list of scores Returns a list of frequencies """ ratings = [int(r) for r in ratings] if min_rating is None: min_rating = min(ratings) if max_rating is None: max_rating = max(ratings) num_ratings = int(max_rating - min_rating + 1) hist_ratings = [0 for x in range(num_ratings)] for r in ratings: hist_ratings[r - min_rating] += 1 return hist_ratings
[ "Generates", "a", "frequency", "count", "of", "each", "rating", "on", "the", "scale", "ratings", "is", "a", "list", "of", "scores", "Returns", "a", "list", "of", "frequencies" ]
edx/ease
python
https://github.com/edx/ease/blob/a7890ed403da94d03726b0639cd8ebda45af6bbb/ease/util_functions.py#L410-L425
[ "def", "histogram", "(", "ratings", ",", "min_rating", "=", "None", ",", "max_rating", "=", "None", ")", ":", "ratings", "=", "[", "int", "(", "r", ")", "for", "r", "in", "ratings", "]", "if", "min_rating", "is", "None", ":", "min_rating", "=", "min"...
a7890ed403da94d03726b0639cd8ebda45af6bbb
valid
get_wordnet_syns
Utilize wordnet (installed with nltk) to get synonyms for words word is the input word returns a list of unique synonyms
ease/util_functions.py
def get_wordnet_syns(word): """ Utilize wordnet (installed with nltk) to get synonyms for words word is the input word returns a list of unique synonyms """ synonyms = [] regex = r"_" pat = re.compile(regex) synset = nltk.wordnet.wordnet.synsets(word) for ss in synset: for swords in ss.lemma_names: synonyms.append(pat.sub(" ", swords.lower())) synonyms = f7(synonyms) return synonyms
def get_wordnet_syns(word): """ Utilize wordnet (installed with nltk) to get synonyms for words word is the input word returns a list of unique synonyms """ synonyms = [] regex = r"_" pat = re.compile(regex) synset = nltk.wordnet.wordnet.synsets(word) for ss in synset: for swords in ss.lemma_names: synonyms.append(pat.sub(" ", swords.lower())) synonyms = f7(synonyms) return synonyms
[ "Utilize", "wordnet", "(", "installed", "with", "nltk", ")", "to", "get", "synonyms", "for", "words", "word", "is", "the", "input", "word", "returns", "a", "list", "of", "unique", "synonyms" ]
edx/ease
python
https://github.com/edx/ease/blob/a7890ed403da94d03726b0639cd8ebda45af6bbb/ease/util_functions.py#L428-L442
[ "def", "get_wordnet_syns", "(", "word", ")", ":", "synonyms", "=", "[", "]", "regex", "=", "r\"_\"", "pat", "=", "re", ".", "compile", "(", "regex", ")", "synset", "=", "nltk", ".", "wordnet", ".", "wordnet", ".", "synsets", "(", "word", ")", "for", ...
a7890ed403da94d03726b0639cd8ebda45af6bbb
valid
get_separator_words
Finds the words that separate a list of tokens from a background corpus Basically this generates a list of informative/interesting words in a set toks1 is a list of words Returns a list of separator words
ease/util_functions.py
def get_separator_words(toks1): """ Finds the words that separate a list of tokens from a background corpus Basically this generates a list of informative/interesting words in a set toks1 is a list of words Returns a list of separator words """ tab_toks1 = nltk.FreqDist(word.lower() for word in toks1) if(os.path.isfile(ESSAY_COR_TOKENS_PATH)): toks2 = pickle.load(open(ESSAY_COR_TOKENS_PATH, 'rb')) else: essay_corpus = open(ESSAY_CORPUS_PATH).read() essay_corpus = sub_chars(essay_corpus) toks2 = nltk.FreqDist(word.lower() for word in nltk.word_tokenize(essay_corpus)) pickle.dump(toks2, open(ESSAY_COR_TOKENS_PATH, 'wb')) sep_words = [] for word in tab_toks1.keys(): tok1_present = tab_toks1[word] if(tok1_present > 2): tok1_total = tab_toks1._N tok2_present = toks2[word] tok2_total = toks2._N fish_val = pvalue(tok1_present, tok2_present, tok1_total, tok2_total).two_tail if(fish_val < .001 and tok1_present / float(tok1_total) > (tok2_present / float(tok2_total)) * 2): sep_words.append(word) sep_words = [w for w in sep_words if not w in nltk.corpus.stopwords.words("english") and len(w) > 5] return sep_words
def get_separator_words(toks1): """ Finds the words that separate a list of tokens from a background corpus Basically this generates a list of informative/interesting words in a set toks1 is a list of words Returns a list of separator words """ tab_toks1 = nltk.FreqDist(word.lower() for word in toks1) if(os.path.isfile(ESSAY_COR_TOKENS_PATH)): toks2 = pickle.load(open(ESSAY_COR_TOKENS_PATH, 'rb')) else: essay_corpus = open(ESSAY_CORPUS_PATH).read() essay_corpus = sub_chars(essay_corpus) toks2 = nltk.FreqDist(word.lower() for word in nltk.word_tokenize(essay_corpus)) pickle.dump(toks2, open(ESSAY_COR_TOKENS_PATH, 'wb')) sep_words = [] for word in tab_toks1.keys(): tok1_present = tab_toks1[word] if(tok1_present > 2): tok1_total = tab_toks1._N tok2_present = toks2[word] tok2_total = toks2._N fish_val = pvalue(tok1_present, tok2_present, tok1_total, tok2_total).two_tail if(fish_val < .001 and tok1_present / float(tok1_total) > (tok2_present / float(tok2_total)) * 2): sep_words.append(word) sep_words = [w for w in sep_words if not w in nltk.corpus.stopwords.words("english") and len(w) > 5] return sep_words
[ "Finds", "the", "words", "that", "separate", "a", "list", "of", "tokens", "from", "a", "background", "corpus", "Basically", "this", "generates", "a", "list", "of", "informative", "/", "interesting", "words", "in", "a", "set", "toks1", "is", "a", "list", "o...
edx/ease
python
https://github.com/edx/ease/blob/a7890ed403da94d03726b0639cd8ebda45af6bbb/ease/util_functions.py#L445-L471
[ "def", "get_separator_words", "(", "toks1", ")", ":", "tab_toks1", "=", "nltk", ".", "FreqDist", "(", "word", ".", "lower", "(", ")", "for", "word", "in", "toks1", ")", "if", "(", "os", ".", "path", ".", "isfile", "(", "ESSAY_COR_TOKENS_PATH", ")", ")"...
a7890ed403da94d03726b0639cd8ebda45af6bbb
valid
encode_plus
Literally encodes the plus sign input is a string returns the string with plus signs encoded
ease/util_functions.py
def encode_plus(s): """ Literally encodes the plus sign input is a string returns the string with plus signs encoded """ regex = r"\+" pat = re.compile(regex) return pat.sub("%2B", s)
def encode_plus(s): """ Literally encodes the plus sign input is a string returns the string with plus signs encoded """ regex = r"\+" pat = re.compile(regex) return pat.sub("%2B", s)
[ "Literally", "encodes", "the", "plus", "sign", "input", "is", "a", "string", "returns", "the", "string", "with", "plus", "signs", "encoded" ]
edx/ease
python
https://github.com/edx/ease/blob/a7890ed403da94d03726b0639cd8ebda45af6bbb/ease/util_functions.py#L474-L482
[ "def", "encode_plus", "(", "s", ")", ":", "regex", "=", "r\"\\+\"", "pat", "=", "re", ".", "compile", "(", "regex", ")", "return", "pat", ".", "sub", "(", "\"%2B\"", ",", "s", ")" ]
a7890ed403da94d03726b0639cd8ebda45af6bbb
valid
getMedian
Gets the median of a list of values Returns a float/int
ease/util_functions.py
def getMedian(numericValues): """ Gets the median of a list of values Returns a float/int """ theValues = sorted(numericValues) if len(theValues) % 2 == 1: return theValues[(len(theValues) + 1) / 2 - 1] else: lower = theValues[len(theValues) / 2 - 1] upper = theValues[len(theValues) / 2] return (float(lower + upper)) / 2
def getMedian(numericValues): """ Gets the median of a list of values Returns a float/int """ theValues = sorted(numericValues) if len(theValues) % 2 == 1: return theValues[(len(theValues) + 1) / 2 - 1] else: lower = theValues[len(theValues) / 2 - 1] upper = theValues[len(theValues) / 2] return (float(lower + upper)) / 2
[ "Gets", "the", "median", "of", "a", "list", "of", "values", "Returns", "a", "float", "/", "int" ]
edx/ease
python
https://github.com/edx/ease/blob/a7890ed403da94d03726b0639cd8ebda45af6bbb/ease/util_functions.py#L485-L498
[ "def", "getMedian", "(", "numericValues", ")", ":", "theValues", "=", "sorted", "(", "numericValues", ")", "if", "len", "(", "theValues", ")", "%", "2", "==", "1", ":", "return", "theValues", "[", "(", "len", "(", "theValues", ")", "+", "1", ")", "/"...
a7890ed403da94d03726b0639cd8ebda45af6bbb
valid
FeatureExtractor.initialize_dictionaries
Initializes dictionaries from an essay set object Dictionaries must be initialized prior to using this to extract features e_set is an input essay set returns a confirmation of initialization
ease/feature_extractor.py
def initialize_dictionaries(self, e_set, max_feats2 = 200): """ Initializes dictionaries from an essay set object Dictionaries must be initialized prior to using this to extract features e_set is an input essay set returns a confirmation of initialization """ if(hasattr(e_set, '_type')): if(e_set._type == "train"): #normal text (unstemmed) useful words/bigrams nvocab = util_functions.get_vocab(e_set._text, e_set._score, max_feats2 = max_feats2) #stemmed and spell corrected vocab useful words/ngrams svocab = util_functions.get_vocab(e_set._clean_stem_text, e_set._score, max_feats2 = max_feats2) #dictionary trained on proper vocab self._normal_dict = CountVectorizer(ngram_range=(1,2), vocabulary=nvocab) #dictionary trained on proper vocab self._stem_dict = CountVectorizer(ngram_range=(1,2), vocabulary=svocab) self.dict_initialized = True #Average spelling errors in set. needed later for spelling detection self._mean_spelling_errors=sum(e_set._spelling_errors)/float(len(e_set._spelling_errors)) self._spell_errors_per_character=sum(e_set._spelling_errors)/float(sum([len(t) for t in e_set._text])) #Gets the number and positions of grammar errors good_pos_tags,bad_pos_positions=self._get_grammar_errors(e_set._pos,e_set._text,e_set._tokens) self._grammar_errors_per_character=(sum(good_pos_tags)/float(sum([len(t) for t in e_set._text]))) #Generate bag of words features bag_feats=self.gen_bag_feats(e_set) #Sum of a row of bag of words features (topical words in an essay) f_row_sum=numpy.sum(bag_feats[:,:]) #Average index of how "topical" essays are self._mean_f_prop=f_row_sum/float(sum([len(t) for t in e_set._text])) ret = "ok" else: raise util_functions.InputError(e_set, "needs to be an essay set of the train type.") else: raise util_functions.InputError(e_set, "wrong input. need an essay set object") return ret
def initialize_dictionaries(self, e_set, max_feats2 = 200): """ Initializes dictionaries from an essay set object Dictionaries must be initialized prior to using this to extract features e_set is an input essay set returns a confirmation of initialization """ if(hasattr(e_set, '_type')): if(e_set._type == "train"): #normal text (unstemmed) useful words/bigrams nvocab = util_functions.get_vocab(e_set._text, e_set._score, max_feats2 = max_feats2) #stemmed and spell corrected vocab useful words/ngrams svocab = util_functions.get_vocab(e_set._clean_stem_text, e_set._score, max_feats2 = max_feats2) #dictionary trained on proper vocab self._normal_dict = CountVectorizer(ngram_range=(1,2), vocabulary=nvocab) #dictionary trained on proper vocab self._stem_dict = CountVectorizer(ngram_range=(1,2), vocabulary=svocab) self.dict_initialized = True #Average spelling errors in set. needed later for spelling detection self._mean_spelling_errors=sum(e_set._spelling_errors)/float(len(e_set._spelling_errors)) self._spell_errors_per_character=sum(e_set._spelling_errors)/float(sum([len(t) for t in e_set._text])) #Gets the number and positions of grammar errors good_pos_tags,bad_pos_positions=self._get_grammar_errors(e_set._pos,e_set._text,e_set._tokens) self._grammar_errors_per_character=(sum(good_pos_tags)/float(sum([len(t) for t in e_set._text]))) #Generate bag of words features bag_feats=self.gen_bag_feats(e_set) #Sum of a row of bag of words features (topical words in an essay) f_row_sum=numpy.sum(bag_feats[:,:]) #Average index of how "topical" essays are self._mean_f_prop=f_row_sum/float(sum([len(t) for t in e_set._text])) ret = "ok" else: raise util_functions.InputError(e_set, "needs to be an essay set of the train type.") else: raise util_functions.InputError(e_set, "wrong input. need an essay set object") return ret
[ "Initializes", "dictionaries", "from", "an", "essay", "set", "object", "Dictionaries", "must", "be", "initialized", "prior", "to", "using", "this", "to", "extract", "features", "e_set", "is", "an", "input", "essay", "set", "returns", "a", "confirmation", "of", ...
edx/ease
python
https://github.com/edx/ease/blob/a7890ed403da94d03726b0639cd8ebda45af6bbb/ease/feature_extractor.py#L38-L73
[ "def", "initialize_dictionaries", "(", "self", ",", "e_set", ",", "max_feats2", "=", "200", ")", ":", "if", "(", "hasattr", "(", "e_set", ",", "'_type'", ")", ")", ":", "if", "(", "e_set", ".", "_type", "==", "\"train\"", ")", ":", "#normal text (unstemm...
a7890ed403da94d03726b0639cd8ebda45af6bbb
valid
FeatureExtractor.get_good_pos_ngrams
Gets a set of gramatically correct part of speech sequences from an input file called essaycorpus.txt Returns the set and caches the file
ease/feature_extractor.py
def get_good_pos_ngrams(self): """ Gets a set of gramatically correct part of speech sequences from an input file called essaycorpus.txt Returns the set and caches the file """ if(os.path.isfile(NGRAM_PATH)): good_pos_ngrams = pickle.load(open(NGRAM_PATH, 'rb')) elif os.path.isfile(ESSAY_CORPUS_PATH): essay_corpus = open(ESSAY_CORPUS_PATH).read() essay_corpus = util_functions.sub_chars(essay_corpus) good_pos_ngrams = util_functions.regenerate_good_tokens(essay_corpus) pickle.dump(good_pos_ngrams, open(NGRAM_PATH, 'wb')) else: #Hard coded list in case the needed files cannot be found good_pos_ngrams=['NN PRP', 'NN PRP .', 'NN PRP . DT', 'PRP .', 'PRP . DT', 'PRP . DT NNP', '. DT', '. DT NNP', '. DT NNP NNP', 'DT NNP', 'DT NNP NNP', 'DT NNP NNP NNP', 'NNP NNP', 'NNP NNP NNP', 'NNP NNP NNP NNP', 'NNP NNP NNP .', 'NNP NNP .', 'NNP NNP . TO', 'NNP .', 'NNP . TO', 'NNP . TO NNP', '. TO', '. TO NNP', '. TO NNP NNP', 'TO NNP', 'TO NNP NNP'] return set(good_pos_ngrams)
def get_good_pos_ngrams(self): """ Gets a set of gramatically correct part of speech sequences from an input file called essaycorpus.txt Returns the set and caches the file """ if(os.path.isfile(NGRAM_PATH)): good_pos_ngrams = pickle.load(open(NGRAM_PATH, 'rb')) elif os.path.isfile(ESSAY_CORPUS_PATH): essay_corpus = open(ESSAY_CORPUS_PATH).read() essay_corpus = util_functions.sub_chars(essay_corpus) good_pos_ngrams = util_functions.regenerate_good_tokens(essay_corpus) pickle.dump(good_pos_ngrams, open(NGRAM_PATH, 'wb')) else: #Hard coded list in case the needed files cannot be found good_pos_ngrams=['NN PRP', 'NN PRP .', 'NN PRP . DT', 'PRP .', 'PRP . DT', 'PRP . DT NNP', '. DT', '. DT NNP', '. DT NNP NNP', 'DT NNP', 'DT NNP NNP', 'DT NNP NNP NNP', 'NNP NNP', 'NNP NNP NNP', 'NNP NNP NNP NNP', 'NNP NNP NNP .', 'NNP NNP .', 'NNP NNP . TO', 'NNP .', 'NNP . TO', 'NNP . TO NNP', '. TO', '. TO NNP', '. TO NNP NNP', 'TO NNP', 'TO NNP NNP'] return set(good_pos_ngrams)
[ "Gets", "a", "set", "of", "gramatically", "correct", "part", "of", "speech", "sequences", "from", "an", "input", "file", "called", "essaycorpus", ".", "txt", "Returns", "the", "set", "and", "caches", "the", "file" ]
edx/ease
python
https://github.com/edx/ease/blob/a7890ed403da94d03726b0639cd8ebda45af6bbb/ease/feature_extractor.py#L75-L95
[ "def", "get_good_pos_ngrams", "(", "self", ")", ":", "if", "(", "os", ".", "path", ".", "isfile", "(", "NGRAM_PATH", ")", ")", ":", "good_pos_ngrams", "=", "pickle", ".", "load", "(", "open", "(", "NGRAM_PATH", ",", "'rb'", ")", ")", "elif", "os", "....
a7890ed403da94d03726b0639cd8ebda45af6bbb
valid
FeatureExtractor._get_grammar_errors
Internal function to get the number of grammar errors in given text pos - part of speech tagged text (list) text - normal text (list) tokens - list of lists of tokenized text
ease/feature_extractor.py
def _get_grammar_errors(self,pos,text,tokens): """ Internal function to get the number of grammar errors in given text pos - part of speech tagged text (list) text - normal text (list) tokens - list of lists of tokenized text """ word_counts = [max(len(t),1) for t in tokens] good_pos_tags = [] min_pos_seq=2 max_pos_seq=4 bad_pos_positions=[] for i in xrange(0, len(text)): pos_seq = [tag[1] for tag in pos[i]] pos_ngrams = util_functions.ngrams(pos_seq, min_pos_seq, max_pos_seq) long_pos_ngrams=[z for z in pos_ngrams if z.count(' ')==(max_pos_seq-1)] bad_pos_tuples=[[z,z+max_pos_seq] for z in xrange(0,len(long_pos_ngrams)) if long_pos_ngrams[z] not in self._good_pos_ngrams] bad_pos_tuples.sort(key=operator.itemgetter(1)) to_delete=[] for m in reversed(xrange(len(bad_pos_tuples)-1)): start, end = bad_pos_tuples[m] for j in xrange(m+1, len(bad_pos_tuples)): lstart, lend = bad_pos_tuples[j] if lstart >= start and lstart <= end: bad_pos_tuples[m][1]=bad_pos_tuples[j][1] to_delete.append(j) fixed_bad_pos_tuples=[bad_pos_tuples[z] for z in xrange(0,len(bad_pos_tuples)) if z not in to_delete] bad_pos_positions.append(fixed_bad_pos_tuples) overlap_ngrams = [z for z in pos_ngrams if z in self._good_pos_ngrams] if (len(pos_ngrams)-len(overlap_ngrams))>0: divisor=len(pos_ngrams)/len(pos_seq) else: divisor=1 if divisor == 0: divisor=1 good_grammar_ratio = (len(pos_ngrams)-len(overlap_ngrams))/divisor good_pos_tags.append(good_grammar_ratio) return good_pos_tags,bad_pos_positions
def _get_grammar_errors(self,pos,text,tokens): """ Internal function to get the number of grammar errors in given text pos - part of speech tagged text (list) text - normal text (list) tokens - list of lists of tokenized text """ word_counts = [max(len(t),1) for t in tokens] good_pos_tags = [] min_pos_seq=2 max_pos_seq=4 bad_pos_positions=[] for i in xrange(0, len(text)): pos_seq = [tag[1] for tag in pos[i]] pos_ngrams = util_functions.ngrams(pos_seq, min_pos_seq, max_pos_seq) long_pos_ngrams=[z for z in pos_ngrams if z.count(' ')==(max_pos_seq-1)] bad_pos_tuples=[[z,z+max_pos_seq] for z in xrange(0,len(long_pos_ngrams)) if long_pos_ngrams[z] not in self._good_pos_ngrams] bad_pos_tuples.sort(key=operator.itemgetter(1)) to_delete=[] for m in reversed(xrange(len(bad_pos_tuples)-1)): start, end = bad_pos_tuples[m] for j in xrange(m+1, len(bad_pos_tuples)): lstart, lend = bad_pos_tuples[j] if lstart >= start and lstart <= end: bad_pos_tuples[m][1]=bad_pos_tuples[j][1] to_delete.append(j) fixed_bad_pos_tuples=[bad_pos_tuples[z] for z in xrange(0,len(bad_pos_tuples)) if z not in to_delete] bad_pos_positions.append(fixed_bad_pos_tuples) overlap_ngrams = [z for z in pos_ngrams if z in self._good_pos_ngrams] if (len(pos_ngrams)-len(overlap_ngrams))>0: divisor=len(pos_ngrams)/len(pos_seq) else: divisor=1 if divisor == 0: divisor=1 good_grammar_ratio = (len(pos_ngrams)-len(overlap_ngrams))/divisor good_pos_tags.append(good_grammar_ratio) return good_pos_tags,bad_pos_positions
[ "Internal", "function", "to", "get", "the", "number", "of", "grammar", "errors", "in", "given", "text", "pos", "-", "part", "of", "speech", "tagged", "text", "(", "list", ")", "text", "-", "normal", "text", "(", "list", ")", "tokens", "-", "list", "of"...
edx/ease
python
https://github.com/edx/ease/blob/a7890ed403da94d03726b0639cd8ebda45af6bbb/ease/feature_extractor.py#L97-L135
[ "def", "_get_grammar_errors", "(", "self", ",", "pos", ",", "text", ",", "tokens", ")", ":", "word_counts", "=", "[", "max", "(", "len", "(", "t", ")", ",", "1", ")", "for", "t", "in", "tokens", "]", "good_pos_tags", "=", "[", "]", "min_pos_seq", "...
a7890ed403da94d03726b0639cd8ebda45af6bbb
valid
FeatureExtractor.gen_length_feats
Generates length based features from an essay set Generally an internal function called by gen_feats Returns an array of length features e_set - EssaySet object
ease/feature_extractor.py
def gen_length_feats(self, e_set): """ Generates length based features from an essay set Generally an internal function called by gen_feats Returns an array of length features e_set - EssaySet object """ text = e_set._text lengths = [len(e) for e in text] word_counts = [max(len(t),1) for t in e_set._tokens] comma_count = [e.count(",") for e in text] ap_count = [e.count("'") for e in text] punc_count = [e.count(".") + e.count("?") + e.count("!") for e in text] chars_per_word = [lengths[m] / float(word_counts[m]) for m in xrange(0, len(text))] good_pos_tags,bad_pos_positions= self._get_grammar_errors(e_set._pos,e_set._text,e_set._tokens) good_pos_tag_prop = [good_pos_tags[m] / float(word_counts[m]) for m in xrange(0, len(text))] length_arr = numpy.array(( lengths, word_counts, comma_count, ap_count, punc_count, chars_per_word, good_pos_tags, good_pos_tag_prop)).transpose() return length_arr.copy()
def gen_length_feats(self, e_set): """ Generates length based features from an essay set Generally an internal function called by gen_feats Returns an array of length features e_set - EssaySet object """ text = e_set._text lengths = [len(e) for e in text] word_counts = [max(len(t),1) for t in e_set._tokens] comma_count = [e.count(",") for e in text] ap_count = [e.count("'") for e in text] punc_count = [e.count(".") + e.count("?") + e.count("!") for e in text] chars_per_word = [lengths[m] / float(word_counts[m]) for m in xrange(0, len(text))] good_pos_tags,bad_pos_positions= self._get_grammar_errors(e_set._pos,e_set._text,e_set._tokens) good_pos_tag_prop = [good_pos_tags[m] / float(word_counts[m]) for m in xrange(0, len(text))] length_arr = numpy.array(( lengths, word_counts, comma_count, ap_count, punc_count, chars_per_word, good_pos_tags, good_pos_tag_prop)).transpose() return length_arr.copy()
[ "Generates", "length", "based", "features", "from", "an", "essay", "set", "Generally", "an", "internal", "function", "called", "by", "gen_feats", "Returns", "an", "array", "of", "length", "features", "e_set", "-", "EssaySet", "object" ]
edx/ease
python
https://github.com/edx/ease/blob/a7890ed403da94d03726b0639cd8ebda45af6bbb/ease/feature_extractor.py#L137-L159
[ "def", "gen_length_feats", "(", "self", ",", "e_set", ")", ":", "text", "=", "e_set", ".", "_text", "lengths", "=", "[", "len", "(", "e", ")", "for", "e", "in", "text", "]", "word_counts", "=", "[", "max", "(", "len", "(", "t", ")", ",", "1", "...
a7890ed403da94d03726b0639cd8ebda45af6bbb
valid
FeatureExtractor.gen_bag_feats
Generates bag of words features from an input essay set and trained FeatureExtractor Generally called by gen_feats Returns an array of features e_set - EssaySet object
ease/feature_extractor.py
def gen_bag_feats(self, e_set): """ Generates bag of words features from an input essay set and trained FeatureExtractor Generally called by gen_feats Returns an array of features e_set - EssaySet object """ if(hasattr(self, '_stem_dict')): sfeats = self._stem_dict.transform(e_set._clean_stem_text) nfeats = self._normal_dict.transform(e_set._text) bag_feats = numpy.concatenate((sfeats.toarray(), nfeats.toarray()), axis=1) else: raise util_functions.InputError(self, "Dictionaries must be initialized prior to generating bag features.") return bag_feats.copy()
def gen_bag_feats(self, e_set): """ Generates bag of words features from an input essay set and trained FeatureExtractor Generally called by gen_feats Returns an array of features e_set - EssaySet object """ if(hasattr(self, '_stem_dict')): sfeats = self._stem_dict.transform(e_set._clean_stem_text) nfeats = self._normal_dict.transform(e_set._text) bag_feats = numpy.concatenate((sfeats.toarray(), nfeats.toarray()), axis=1) else: raise util_functions.InputError(self, "Dictionaries must be initialized prior to generating bag features.") return bag_feats.copy()
[ "Generates", "bag", "of", "words", "features", "from", "an", "input", "essay", "set", "and", "trained", "FeatureExtractor", "Generally", "called", "by", "gen_feats", "Returns", "an", "array", "of", "features", "e_set", "-", "EssaySet", "object" ]
edx/ease
python
https://github.com/edx/ease/blob/a7890ed403da94d03726b0639cd8ebda45af6bbb/ease/feature_extractor.py#L161-L174
[ "def", "gen_bag_feats", "(", "self", ",", "e_set", ")", ":", "if", "(", "hasattr", "(", "self", ",", "'_stem_dict'", ")", ")", ":", "sfeats", "=", "self", ".", "_stem_dict", ".", "transform", "(", "e_set", ".", "_clean_stem_text", ")", "nfeats", "=", "...
a7890ed403da94d03726b0639cd8ebda45af6bbb
valid
FeatureExtractor.gen_feats
Generates bag of words, length, and prompt features from an essay set object returns an array of features e_set - EssaySet object
ease/feature_extractor.py
def gen_feats(self, e_set): """ Generates bag of words, length, and prompt features from an essay set object returns an array of features e_set - EssaySet object """ bag_feats = self.gen_bag_feats(e_set) length_feats = self.gen_length_feats(e_set) prompt_feats = self.gen_prompt_feats(e_set) overall_feats = numpy.concatenate((length_feats, prompt_feats, bag_feats), axis=1) overall_feats = overall_feats.copy() return overall_feats
def gen_feats(self, e_set): """ Generates bag of words, length, and prompt features from an essay set object returns an array of features e_set - EssaySet object """ bag_feats = self.gen_bag_feats(e_set) length_feats = self.gen_length_feats(e_set) prompt_feats = self.gen_prompt_feats(e_set) overall_feats = numpy.concatenate((length_feats, prompt_feats, bag_feats), axis=1) overall_feats = overall_feats.copy() return overall_feats
[ "Generates", "bag", "of", "words", "length", "and", "prompt", "features", "from", "an", "essay", "set", "object", "returns", "an", "array", "of", "features", "e_set", "-", "EssaySet", "object" ]
edx/ease
python
https://github.com/edx/ease/blob/a7890ed403da94d03726b0639cd8ebda45af6bbb/ease/feature_extractor.py#L176-L188
[ "def", "gen_feats", "(", "self", ",", "e_set", ")", ":", "bag_feats", "=", "self", ".", "gen_bag_feats", "(", "e_set", ")", "length_feats", "=", "self", ".", "gen_length_feats", "(", "e_set", ")", "prompt_feats", "=", "self", ".", "gen_prompt_feats", "(", ...
a7890ed403da94d03726b0639cd8ebda45af6bbb
valid
FeatureExtractor.gen_prompt_feats
Generates prompt based features from an essay set object and internal prompt variable. Generally called internally by gen_feats Returns an array of prompt features e_set - EssaySet object
ease/feature_extractor.py
def gen_prompt_feats(self, e_set): """ Generates prompt based features from an essay set object and internal prompt variable. Generally called internally by gen_feats Returns an array of prompt features e_set - EssaySet object """ prompt_toks = nltk.word_tokenize(e_set._prompt) expand_syns = [] for word in prompt_toks: synonyms = util_functions.get_wordnet_syns(word) expand_syns.append(synonyms) expand_syns = list(chain.from_iterable(expand_syns)) prompt_overlap = [] prompt_overlap_prop = [] for j in e_set._tokens: tok_length=len(j) if(tok_length==0): tok_length=1 prompt_overlap.append(len([i for i in j if i in prompt_toks])) prompt_overlap_prop.append(prompt_overlap[len(prompt_overlap) - 1] / float(tok_length)) expand_overlap = [] expand_overlap_prop = [] for j in e_set._tokens: tok_length=len(j) if(tok_length==0): tok_length=1 expand_overlap.append(len([i for i in j if i in expand_syns])) expand_overlap_prop.append(expand_overlap[len(expand_overlap) - 1] / float(tok_length)) prompt_arr = numpy.array((prompt_overlap, prompt_overlap_prop, expand_overlap, expand_overlap_prop)).transpose() return prompt_arr.copy()
def gen_prompt_feats(self, e_set): """ Generates prompt based features from an essay set object and internal prompt variable. Generally called internally by gen_feats Returns an array of prompt features e_set - EssaySet object """ prompt_toks = nltk.word_tokenize(e_set._prompt) expand_syns = [] for word in prompt_toks: synonyms = util_functions.get_wordnet_syns(word) expand_syns.append(synonyms) expand_syns = list(chain.from_iterable(expand_syns)) prompt_overlap = [] prompt_overlap_prop = [] for j in e_set._tokens: tok_length=len(j) if(tok_length==0): tok_length=1 prompt_overlap.append(len([i for i in j if i in prompt_toks])) prompt_overlap_prop.append(prompt_overlap[len(prompt_overlap) - 1] / float(tok_length)) expand_overlap = [] expand_overlap_prop = [] for j in e_set._tokens: tok_length=len(j) if(tok_length==0): tok_length=1 expand_overlap.append(len([i for i in j if i in expand_syns])) expand_overlap_prop.append(expand_overlap[len(expand_overlap) - 1] / float(tok_length)) prompt_arr = numpy.array((prompt_overlap, prompt_overlap_prop, expand_overlap, expand_overlap_prop)).transpose() return prompt_arr.copy()
[ "Generates", "prompt", "based", "features", "from", "an", "essay", "set", "object", "and", "internal", "prompt", "variable", ".", "Generally", "called", "internally", "by", "gen_feats", "Returns", "an", "array", "of", "prompt", "features", "e_set", "-", "EssaySe...
edx/ease
python
https://github.com/edx/ease/blob/a7890ed403da94d03726b0639cd8ebda45af6bbb/ease/feature_extractor.py#L190-L222
[ "def", "gen_prompt_feats", "(", "self", ",", "e_set", ")", ":", "prompt_toks", "=", "nltk", ".", "word_tokenize", "(", "e_set", ".", "_prompt", ")", "expand_syns", "=", "[", "]", "for", "word", "in", "prompt_toks", ":", "synonyms", "=", "util_functions", "...
a7890ed403da94d03726b0639cd8ebda45af6bbb
valid
FeatureExtractor.gen_feedback
Generate feedback for a given set of essays e_set - EssaySet object features - optionally, pass in a matrix of features extracted from e_set using FeatureExtractor in order to get off topic feedback. Returns a list of lists (one list per essay in e_set) e_set - EssaySet object
ease/feature_extractor.py
def gen_feedback(self, e_set, features=None): """ Generate feedback for a given set of essays e_set - EssaySet object features - optionally, pass in a matrix of features extracted from e_set using FeatureExtractor in order to get off topic feedback. Returns a list of lists (one list per essay in e_set) e_set - EssaySet object """ #Set ratio to modify thresholds for grammar/spelling errors modifier_ratio=1.05 #Calc number of grammar and spelling errors per character set_grammar,bad_pos_positions=self._get_grammar_errors(e_set._pos,e_set._text,e_set._tokens) set_grammar_per_character=[set_grammar[m]/float(len(e_set._text[m])+.1) for m in xrange(0,len(e_set._text))] set_spell_errors_per_character=[e_set._spelling_errors[m]/float(len(e_set._text[m])+.1) for m in xrange(0,len(e_set._text))] #Iterate through essays and create a feedback dict for each all_feedback=[] for m in xrange(0,len(e_set._text)): #Be very careful about changing these messages! individual_feedback={'grammar' : "Grammar: Ok.", 'spelling' : "Spelling: Ok.", 'markup_text' : "", 'grammar_per_char' : set_grammar_per_character[m], 'spelling_per_char' : set_spell_errors_per_character[m], 'too_similar_to_prompt' : False, } markup_tokens=e_set._markup_text[m].split(" ") #This loop ensures that sequences of bad grammar get put together into one sequence instead of staying #disjointed bad_pos_starts=[z[0] for z in bad_pos_positions[m]] bad_pos_ends=[z[1]-1 for z in bad_pos_positions[m]] for z in xrange(0,len(markup_tokens)): if z in bad_pos_starts: markup_tokens[z]='<bg>' + markup_tokens[z] elif z in bad_pos_ends: markup_tokens[z]=markup_tokens[z] + "</bg>" if(len(bad_pos_ends)>0 and len(bad_pos_starts)>0 and len(markup_tokens)>1): if max(bad_pos_ends)>(len(markup_tokens)-1) and max(bad_pos_starts)<(len(markup_tokens)-1): markup_tokens[len(markup_tokens)-1]+="</bg>" #Display messages if grammar/spelling errors greater than average in training set if set_grammar_per_character[m]>(self._grammar_errors_per_character*modifier_ratio): individual_feedback['grammar']="Grammar: More grammar errors than average." if set_spell_errors_per_character[m]>(self._spell_errors_per_character*modifier_ratio): individual_feedback['spelling']="Spelling: More spelling errors than average." #Test topicality by calculating # of on topic words per character and comparing to the training set #mean. Requires features to be passed in if features is not None: f_row_sum=numpy.sum(features[m,12:]) f_row_prop=f_row_sum/len(e_set._text[m]) if f_row_prop<(self._mean_f_prop/1.5) or len(e_set._text[m])<20: individual_feedback['topicality']="Topicality: Essay may be off topic." if(features[m,9]>.6): individual_feedback['prompt_overlap']="Prompt Overlap: Too much overlap with prompt." individual_feedback['too_similar_to_prompt']=True log.debug(features[m,9]) #Create string representation of markup text markup_string=" ".join(markup_tokens) individual_feedback['markup_text']=markup_string all_feedback.append(individual_feedback) return all_feedback
def gen_feedback(self, e_set, features=None): """ Generate feedback for a given set of essays e_set - EssaySet object features - optionally, pass in a matrix of features extracted from e_set using FeatureExtractor in order to get off topic feedback. Returns a list of lists (one list per essay in e_set) e_set - EssaySet object """ #Set ratio to modify thresholds for grammar/spelling errors modifier_ratio=1.05 #Calc number of grammar and spelling errors per character set_grammar,bad_pos_positions=self._get_grammar_errors(e_set._pos,e_set._text,e_set._tokens) set_grammar_per_character=[set_grammar[m]/float(len(e_set._text[m])+.1) for m in xrange(0,len(e_set._text))] set_spell_errors_per_character=[e_set._spelling_errors[m]/float(len(e_set._text[m])+.1) for m in xrange(0,len(e_set._text))] #Iterate through essays and create a feedback dict for each all_feedback=[] for m in xrange(0,len(e_set._text)): #Be very careful about changing these messages! individual_feedback={'grammar' : "Grammar: Ok.", 'spelling' : "Spelling: Ok.", 'markup_text' : "", 'grammar_per_char' : set_grammar_per_character[m], 'spelling_per_char' : set_spell_errors_per_character[m], 'too_similar_to_prompt' : False, } markup_tokens=e_set._markup_text[m].split(" ") #This loop ensures that sequences of bad grammar get put together into one sequence instead of staying #disjointed bad_pos_starts=[z[0] for z in bad_pos_positions[m]] bad_pos_ends=[z[1]-1 for z in bad_pos_positions[m]] for z in xrange(0,len(markup_tokens)): if z in bad_pos_starts: markup_tokens[z]='<bg>' + markup_tokens[z] elif z in bad_pos_ends: markup_tokens[z]=markup_tokens[z] + "</bg>" if(len(bad_pos_ends)>0 and len(bad_pos_starts)>0 and len(markup_tokens)>1): if max(bad_pos_ends)>(len(markup_tokens)-1) and max(bad_pos_starts)<(len(markup_tokens)-1): markup_tokens[len(markup_tokens)-1]+="</bg>" #Display messages if grammar/spelling errors greater than average in training set if set_grammar_per_character[m]>(self._grammar_errors_per_character*modifier_ratio): individual_feedback['grammar']="Grammar: More grammar errors than average." if set_spell_errors_per_character[m]>(self._spell_errors_per_character*modifier_ratio): individual_feedback['spelling']="Spelling: More spelling errors than average." #Test topicality by calculating # of on topic words per character and comparing to the training set #mean. Requires features to be passed in if features is not None: f_row_sum=numpy.sum(features[m,12:]) f_row_prop=f_row_sum/len(e_set._text[m]) if f_row_prop<(self._mean_f_prop/1.5) or len(e_set._text[m])<20: individual_feedback['topicality']="Topicality: Essay may be off topic." if(features[m,9]>.6): individual_feedback['prompt_overlap']="Prompt Overlap: Too much overlap with prompt." individual_feedback['too_similar_to_prompt']=True log.debug(features[m,9]) #Create string representation of markup text markup_string=" ".join(markup_tokens) individual_feedback['markup_text']=markup_string all_feedback.append(individual_feedback) return all_feedback
[ "Generate", "feedback", "for", "a", "given", "set", "of", "essays", "e_set", "-", "EssaySet", "object", "features", "-", "optionally", "pass", "in", "a", "matrix", "of", "features", "extracted", "from", "e_set", "using", "FeatureExtractor", "in", "order", "to"...
edx/ease
python
https://github.com/edx/ease/blob/a7890ed403da94d03726b0639cd8ebda45af6bbb/ease/feature_extractor.py#L224-L292
[ "def", "gen_feedback", "(", "self", ",", "e_set", ",", "features", "=", "None", ")", ":", "#Set ratio to modify thresholds for grammar/spelling errors", "modifier_ratio", "=", "1.05", "#Calc number of grammar and spelling errors per character", "set_grammar", ",", "bad_pos_posi...
a7890ed403da94d03726b0639cd8ebda45af6bbb
valid
EssaySet.add_essay
Add new (essay_text,essay_score) pair to the essay set. essay_text must be a string. essay_score must be an int. essay_generated should not be changed by the user. Returns a confirmation that essay was added.
ease/essay_set.py
def add_essay(self, essay_text, essay_score, essay_generated=0): """ Add new (essay_text,essay_score) pair to the essay set. essay_text must be a string. essay_score must be an int. essay_generated should not be changed by the user. Returns a confirmation that essay was added. """ # Get maximum current essay id, or set to 0 if this is the first essay added if(len(self._id) > 0): max_id = max(self._id) else: max_id = 0 # Verify that essay_score is an int, essay_text is a string, and essay_generated equals 0 or 1 try: essay_text = essay_text.encode('ascii', 'ignore') if len(essay_text) < 5: essay_text = "Invalid essay." except: log.exception("Could not parse essay into ascii.") try: # Try conversion of types essay_score = int(essay_score) essay_text = str(essay_text) except: # Nothing needed here, will return error in any case. log.exception("Invalid type for essay score : {0} or essay text : {1}".format(type(essay_score), type(essay_text))) if isinstance(essay_score, int) and isinstance(essay_text, basestring)\ and (essay_generated == 0 or essay_generated == 1): self._id.append(max_id + 1) self._score.append(essay_score) # Clean text by removing non digit/work/punctuation characters try: essay_text = str(essay_text.encode('ascii', 'ignore')) except: essay_text = (essay_text.decode('utf-8', 'replace')).encode('ascii', 'ignore') cleaned_essay = util_functions.sub_chars(essay_text).lower() if(len(cleaned_essay) > MAXIMUM_ESSAY_LENGTH): cleaned_essay = cleaned_essay[0:MAXIMUM_ESSAY_LENGTH] self._text.append(cleaned_essay) # Spell correct text using aspell cleaned_text, spell_errors, markup_text = util_functions.spell_correct(self._text[len(self._text) - 1]) self._clean_text.append(cleaned_text) self._spelling_errors.append(spell_errors) self._markup_text.append(markup_text) # Tokenize text self._tokens.append(nltk.word_tokenize(self._clean_text[len(self._clean_text) - 1])) # Part of speech tag text self._pos.append(nltk.pos_tag(self._clean_text[len(self._clean_text) - 1].split(" "))) self._generated.append(essay_generated) # Stem spell corrected text porter = nltk.PorterStemmer() por_toks = " ".join([porter.stem(w) for w in self._tokens[len(self._tokens) - 1]]) self._clean_stem_text.append(por_toks) ret = "text: " + self._text[len(self._text) - 1] + " score: " + str(essay_score) else: raise util_functions.InputError(essay_text, "arguments need to be in format " "(text,score). text needs to be string," " score needs to be int.")
def add_essay(self, essay_text, essay_score, essay_generated=0): """ Add new (essay_text,essay_score) pair to the essay set. essay_text must be a string. essay_score must be an int. essay_generated should not be changed by the user. Returns a confirmation that essay was added. """ # Get maximum current essay id, or set to 0 if this is the first essay added if(len(self._id) > 0): max_id = max(self._id) else: max_id = 0 # Verify that essay_score is an int, essay_text is a string, and essay_generated equals 0 or 1 try: essay_text = essay_text.encode('ascii', 'ignore') if len(essay_text) < 5: essay_text = "Invalid essay." except: log.exception("Could not parse essay into ascii.") try: # Try conversion of types essay_score = int(essay_score) essay_text = str(essay_text) except: # Nothing needed here, will return error in any case. log.exception("Invalid type for essay score : {0} or essay text : {1}".format(type(essay_score), type(essay_text))) if isinstance(essay_score, int) and isinstance(essay_text, basestring)\ and (essay_generated == 0 or essay_generated == 1): self._id.append(max_id + 1) self._score.append(essay_score) # Clean text by removing non digit/work/punctuation characters try: essay_text = str(essay_text.encode('ascii', 'ignore')) except: essay_text = (essay_text.decode('utf-8', 'replace')).encode('ascii', 'ignore') cleaned_essay = util_functions.sub_chars(essay_text).lower() if(len(cleaned_essay) > MAXIMUM_ESSAY_LENGTH): cleaned_essay = cleaned_essay[0:MAXIMUM_ESSAY_LENGTH] self._text.append(cleaned_essay) # Spell correct text using aspell cleaned_text, spell_errors, markup_text = util_functions.spell_correct(self._text[len(self._text) - 1]) self._clean_text.append(cleaned_text) self._spelling_errors.append(spell_errors) self._markup_text.append(markup_text) # Tokenize text self._tokens.append(nltk.word_tokenize(self._clean_text[len(self._clean_text) - 1])) # Part of speech tag text self._pos.append(nltk.pos_tag(self._clean_text[len(self._clean_text) - 1].split(" "))) self._generated.append(essay_generated) # Stem spell corrected text porter = nltk.PorterStemmer() por_toks = " ".join([porter.stem(w) for w in self._tokens[len(self._tokens) - 1]]) self._clean_stem_text.append(por_toks) ret = "text: " + self._text[len(self._text) - 1] + " score: " + str(essay_score) else: raise util_functions.InputError(essay_text, "arguments need to be in format " "(text,score). text needs to be string," " score needs to be int.")
[ "Add", "new", "(", "essay_text", "essay_score", ")", "pair", "to", "the", "essay", "set", ".", "essay_text", "must", "be", "a", "string", ".", "essay_score", "must", "be", "an", "int", ".", "essay_generated", "should", "not", "be", "changed", "by", "the", ...
edx/ease
python
https://github.com/edx/ease/blob/a7890ed403da94d03726b0639cd8ebda45af6bbb/ease/essay_set.py#L46-L108
[ "def", "add_essay", "(", "self", ",", "essay_text", ",", "essay_score", ",", "essay_generated", "=", "0", ")", ":", "# Get maximum current essay id, or set to 0 if this is the first essay added", "if", "(", "len", "(", "self", ".", "_id", ")", ">", "0", ")", ":", ...
a7890ed403da94d03726b0639cd8ebda45af6bbb
valid
EssaySet.update_prompt
Update the default prompt string, which is "". prompt_text should be a string. Returns the prompt as a confirmation.
ease/essay_set.py
def update_prompt(self, prompt_text): """ Update the default prompt string, which is "". prompt_text should be a string. Returns the prompt as a confirmation. """ if(isinstance(prompt_text, basestring)): self._prompt = util_functions.sub_chars(prompt_text) ret = self._prompt else: raise util_functions.InputError(prompt_text, "Invalid prompt. Need to enter a string value.") return ret
def update_prompt(self, prompt_text): """ Update the default prompt string, which is "". prompt_text should be a string. Returns the prompt as a confirmation. """ if(isinstance(prompt_text, basestring)): self._prompt = util_functions.sub_chars(prompt_text) ret = self._prompt else: raise util_functions.InputError(prompt_text, "Invalid prompt. Need to enter a string value.") return ret
[ "Update", "the", "default", "prompt", "string", "which", "is", ".", "prompt_text", "should", "be", "a", "string", ".", "Returns", "the", "prompt", "as", "a", "confirmation", "." ]
edx/ease
python
https://github.com/edx/ease/blob/a7890ed403da94d03726b0639cd8ebda45af6bbb/ease/essay_set.py#L110-L121
[ "def", "update_prompt", "(", "self", ",", "prompt_text", ")", ":", "if", "(", "isinstance", "(", "prompt_text", ",", "basestring", ")", ")", ":", "self", ".", "_prompt", "=", "util_functions", ".", "sub_chars", "(", "prompt_text", ")", "ret", "=", "self", ...
a7890ed403da94d03726b0639cd8ebda45af6bbb
valid
EssaySet.generate_additional_essays
Substitute synonyms to generate extra essays from existing ones. This is done to increase the amount of training data. Should only be used with lowest scoring essays. e_text is the text of the original essay. e_score is the score of the original essay. dictionary is a fixed dictionary (list) of words to replace. max_syns defines the maximum number of additional essays to generate. Do not set too high.
ease/essay_set.py
def generate_additional_essays(self, e_text, e_score, dictionary=None, max_syns=3): """ Substitute synonyms to generate extra essays from existing ones. This is done to increase the amount of training data. Should only be used with lowest scoring essays. e_text is the text of the original essay. e_score is the score of the original essay. dictionary is a fixed dictionary (list) of words to replace. max_syns defines the maximum number of additional essays to generate. Do not set too high. """ e_toks = nltk.word_tokenize(e_text) all_syns = [] for word in e_toks: synonyms = util_functions.get_wordnet_syns(word) if(len(synonyms) > max_syns): synonyms = random.sample(synonyms, max_syns) all_syns.append(synonyms) new_essays = [] for i in range(0, max_syns): syn_toks = e_toks for z in range(0, len(e_toks)): if len(all_syns[z]) > i and (dictionary == None or e_toks[z] in dictionary): syn_toks[z] = all_syns[z][i] new_essays.append(" ".join(syn_toks)) for z in xrange(0, len(new_essays)): self.add_essay(new_essays[z], e_score, 1)
def generate_additional_essays(self, e_text, e_score, dictionary=None, max_syns=3): """ Substitute synonyms to generate extra essays from existing ones. This is done to increase the amount of training data. Should only be used with lowest scoring essays. e_text is the text of the original essay. e_score is the score of the original essay. dictionary is a fixed dictionary (list) of words to replace. max_syns defines the maximum number of additional essays to generate. Do not set too high. """ e_toks = nltk.word_tokenize(e_text) all_syns = [] for word in e_toks: synonyms = util_functions.get_wordnet_syns(word) if(len(synonyms) > max_syns): synonyms = random.sample(synonyms, max_syns) all_syns.append(synonyms) new_essays = [] for i in range(0, max_syns): syn_toks = e_toks for z in range(0, len(e_toks)): if len(all_syns[z]) > i and (dictionary == None or e_toks[z] in dictionary): syn_toks[z] = all_syns[z][i] new_essays.append(" ".join(syn_toks)) for z in xrange(0, len(new_essays)): self.add_essay(new_essays[z], e_score, 1)
[ "Substitute", "synonyms", "to", "generate", "extra", "essays", "from", "existing", "ones", ".", "This", "is", "done", "to", "increase", "the", "amount", "of", "training", "data", ".", "Should", "only", "be", "used", "with", "lowest", "scoring", "essays", "."...
edx/ease
python
https://github.com/edx/ease/blob/a7890ed403da94d03726b0639cd8ebda45af6bbb/ease/essay_set.py#L123-L148
[ "def", "generate_additional_essays", "(", "self", ",", "e_text", ",", "e_score", ",", "dictionary", "=", "None", ",", "max_syns", "=", "3", ")", ":", "e_toks", "=", "nltk", ".", "word_tokenize", "(", "e_text", ")", "all_syns", "=", "[", "]", "for", "word...
a7890ed403da94d03726b0639cd8ebda45af6bbb
valid
create
Creates a machine learning model from input text, associated scores, a prompt, and a path to the model TODO: Remove model path argument, it is needed for now to support legacy code text - A list of strings containing the text of the essays score - a list of integers containing score values prompt_string - the common prompt for the set of essays
ease/create.py
def create(text,score,prompt_string, dump_data=False): """ Creates a machine learning model from input text, associated scores, a prompt, and a path to the model TODO: Remove model path argument, it is needed for now to support legacy code text - A list of strings containing the text of the essays score - a list of integers containing score values prompt_string - the common prompt for the set of essays """ if dump_data: dump_input_data(text, score) algorithm = select_algorithm(score) #Initialize a results dictionary to return results = {'errors': [],'success' : False, 'cv_kappa' : 0, 'cv_mean_absolute_error': 0, 'feature_ext' : "", 'classifier' : "", 'algorithm' : algorithm, 'score' : score, 'text' : text, 'prompt' : prompt_string} if len(text)!=len(score): msg = "Target and text lists must be same length." results['errors'].append(msg) log.exception(msg) return results try: #Create an essay set object that encapsulates all the essays and alternate representations (tokens, etc) e_set = model_creator.create_essay_set(text, score, prompt_string) except: msg = "essay set creation failed." results['errors'].append(msg) log.exception(msg) try: #Gets features from the essay set and computes error feature_ext, classifier, cv_error_results = model_creator.extract_features_and_generate_model(e_set, algorithm = algorithm) results['cv_kappa']=cv_error_results['kappa'] results['cv_mean_absolute_error']=cv_error_results['mae'] results['feature_ext']=feature_ext results['classifier']=classifier results['algorithm'] = algorithm results['success']=True except: msg = "feature extraction and model creation failed." results['errors'].append(msg) log.exception(msg) return results
def create(text,score,prompt_string, dump_data=False): """ Creates a machine learning model from input text, associated scores, a prompt, and a path to the model TODO: Remove model path argument, it is needed for now to support legacy code text - A list of strings containing the text of the essays score - a list of integers containing score values prompt_string - the common prompt for the set of essays """ if dump_data: dump_input_data(text, score) algorithm = select_algorithm(score) #Initialize a results dictionary to return results = {'errors': [],'success' : False, 'cv_kappa' : 0, 'cv_mean_absolute_error': 0, 'feature_ext' : "", 'classifier' : "", 'algorithm' : algorithm, 'score' : score, 'text' : text, 'prompt' : prompt_string} if len(text)!=len(score): msg = "Target and text lists must be same length." results['errors'].append(msg) log.exception(msg) return results try: #Create an essay set object that encapsulates all the essays and alternate representations (tokens, etc) e_set = model_creator.create_essay_set(text, score, prompt_string) except: msg = "essay set creation failed." results['errors'].append(msg) log.exception(msg) try: #Gets features from the essay set and computes error feature_ext, classifier, cv_error_results = model_creator.extract_features_and_generate_model(e_set, algorithm = algorithm) results['cv_kappa']=cv_error_results['kappa'] results['cv_mean_absolute_error']=cv_error_results['mae'] results['feature_ext']=feature_ext results['classifier']=classifier results['algorithm'] = algorithm results['success']=True except: msg = "feature extraction and model creation failed." results['errors'].append(msg) log.exception(msg) return results
[ "Creates", "a", "machine", "learning", "model", "from", "input", "text", "associated", "scores", "a", "prompt", "and", "a", "path", "to", "the", "model", "TODO", ":", "Remove", "model", "path", "argument", "it", "is", "needed", "for", "now", "to", "support...
edx/ease
python
https://github.com/edx/ease/blob/a7890ed403da94d03726b0639cd8ebda45af6bbb/ease/create.py#L42-L87
[ "def", "create", "(", "text", ",", "score", ",", "prompt_string", ",", "dump_data", "=", "False", ")", ":", "if", "dump_data", ":", "dump_input_data", "(", "text", ",", "score", ")", "algorithm", "=", "select_algorithm", "(", "score", ")", "#Initialize a res...
a7890ed403da94d03726b0639cd8ebda45af6bbb
valid
create_generic
Creates a model from a generic list numeric values and text values numeric_values - A list of lists that are the predictors textual_values - A list of lists that are the predictors (each item in textual_values corresponds to the similarly indexed counterpart in numeric_values) target - The variable that we are trying to predict. A list of integers. algorithm - the type of algorithm that will be used
ease/create.py
def create_generic(numeric_values, textual_values, target, algorithm = util_functions.AlgorithmTypes.regression): """ Creates a model from a generic list numeric values and text values numeric_values - A list of lists that are the predictors textual_values - A list of lists that are the predictors (each item in textual_values corresponds to the similarly indexed counterpart in numeric_values) target - The variable that we are trying to predict. A list of integers. algorithm - the type of algorithm that will be used """ algorithm = select_algorithm(target) #Initialize a result dictionary to return. results = {'errors': [],'success' : False, 'cv_kappa' : 0, 'cv_mean_absolute_error': 0, 'feature_ext' : "", 'classifier' : "", 'algorithm' : algorithm} if len(numeric_values)!=len(textual_values) or len(numeric_values)!=len(target): msg = "Target, numeric features, and text features must all be the same length." results['errors'].append(msg) log.exception(msg) return results try: #Initialize a predictor set object that encapsulates all of the text and numeric predictors pset = predictor_set.PredictorSet(essaytype="train") for i in xrange(0, len(numeric_values)): pset.add_row(numeric_values[i], textual_values[i], target[i]) except: msg = "predictor set creation failed." results['errors'].append(msg) log.exception(msg) try: #Extract all features and then train a classifier with the features feature_ext, classifier, cv_error_results = model_creator.extract_features_and_generate_model_predictors(pset, algorithm) results['cv_kappa']=cv_error_results['kappa'] results['cv_mean_absolute_error']=cv_error_results['mae'] results['feature_ext']=feature_ext results['classifier']=classifier results['success']=True except: msg = "feature extraction and model creation failed." results['errors'].append(msg) log.exception(msg) return results
def create_generic(numeric_values, textual_values, target, algorithm = util_functions.AlgorithmTypes.regression): """ Creates a model from a generic list numeric values and text values numeric_values - A list of lists that are the predictors textual_values - A list of lists that are the predictors (each item in textual_values corresponds to the similarly indexed counterpart in numeric_values) target - The variable that we are trying to predict. A list of integers. algorithm - the type of algorithm that will be used """ algorithm = select_algorithm(target) #Initialize a result dictionary to return. results = {'errors': [],'success' : False, 'cv_kappa' : 0, 'cv_mean_absolute_error': 0, 'feature_ext' : "", 'classifier' : "", 'algorithm' : algorithm} if len(numeric_values)!=len(textual_values) or len(numeric_values)!=len(target): msg = "Target, numeric features, and text features must all be the same length." results['errors'].append(msg) log.exception(msg) return results try: #Initialize a predictor set object that encapsulates all of the text and numeric predictors pset = predictor_set.PredictorSet(essaytype="train") for i in xrange(0, len(numeric_values)): pset.add_row(numeric_values[i], textual_values[i], target[i]) except: msg = "predictor set creation failed." results['errors'].append(msg) log.exception(msg) try: #Extract all features and then train a classifier with the features feature_ext, classifier, cv_error_results = model_creator.extract_features_and_generate_model_predictors(pset, algorithm) results['cv_kappa']=cv_error_results['kappa'] results['cv_mean_absolute_error']=cv_error_results['mae'] results['feature_ext']=feature_ext results['classifier']=classifier results['success']=True except: msg = "feature extraction and model creation failed." results['errors'].append(msg) log.exception(msg) return results
[ "Creates", "a", "model", "from", "a", "generic", "list", "numeric", "values", "and", "text", "values", "numeric_values", "-", "A", "list", "of", "lists", "that", "are", "the", "predictors", "textual_values", "-", "A", "list", "of", "lists", "that", "are", ...
edx/ease
python
https://github.com/edx/ease/blob/a7890ed403da94d03726b0639cd8ebda45af6bbb/ease/create.py#L90-L134
[ "def", "create_generic", "(", "numeric_values", ",", "textual_values", ",", "target", ",", "algorithm", "=", "util_functions", ".", "AlgorithmTypes", ".", "regression", ")", ":", "algorithm", "=", "select_algorithm", "(", "target", ")", "#Initialize a result dictionar...
a7890ed403da94d03726b0639cd8ebda45af6bbb
valid
create_essay_set
Creates an essay set from given data. Text should be a list of strings corresponding to essay text. Score should be a list of scores where score[n] corresponds to text[n] Prompt string is just a string containing the essay prompt. Generate_additional indicates whether to generate additional essays at the minimum score point or not.
ease/model_creator.py
def create_essay_set(text, score, prompt_string, generate_additional=True): """ Creates an essay set from given data. Text should be a list of strings corresponding to essay text. Score should be a list of scores where score[n] corresponds to text[n] Prompt string is just a string containing the essay prompt. Generate_additional indicates whether to generate additional essays at the minimum score point or not. """ x = EssaySet() for i in xrange(0, len(text)): x.add_essay(text[i], score[i]) if score[i] == min(score) and generate_additional == True: x.generate_additional_essays(x._clean_text[len(x._clean_text) - 1], score[i]) x.update_prompt(prompt_string) return x
def create_essay_set(text, score, prompt_string, generate_additional=True): """ Creates an essay set from given data. Text should be a list of strings corresponding to essay text. Score should be a list of scores where score[n] corresponds to text[n] Prompt string is just a string containing the essay prompt. Generate_additional indicates whether to generate additional essays at the minimum score point or not. """ x = EssaySet() for i in xrange(0, len(text)): x.add_essay(text[i], score[i]) if score[i] == min(score) and generate_additional == True: x.generate_additional_essays(x._clean_text[len(x._clean_text) - 1], score[i]) x.update_prompt(prompt_string) return x
[ "Creates", "an", "essay", "set", "from", "given", "data", ".", "Text", "should", "be", "a", "list", "of", "strings", "corresponding", "to", "essay", "text", ".", "Score", "should", "be", "a", "list", "of", "scores", "where", "score", "[", "n", "]", "co...
edx/ease
python
https://github.com/edx/ease/blob/a7890ed403da94d03726b0639cd8ebda45af6bbb/ease/model_creator.py#L71-L87
[ "def", "create_essay_set", "(", "text", ",", "score", ",", "prompt_string", ",", "generate_additional", "=", "True", ")", ":", "x", "=", "EssaySet", "(", ")", "for", "i", "in", "xrange", "(", "0", ",", "len", "(", "text", ")", ")", ":", "x", ".", "...
a7890ed403da94d03726b0639cd8ebda45af6bbb
valid
get_cv_error
Gets cross validated error for a given classifier, set of features, and scores clf - classifier feats - features to feed into the classified and cross validate over scores - scores associated with the features -- feature row 1 associates with score 1, etc.
ease/model_creator.py
def get_cv_error(clf,feats,scores): """ Gets cross validated error for a given classifier, set of features, and scores clf - classifier feats - features to feed into the classified and cross validate over scores - scores associated with the features -- feature row 1 associates with score 1, etc. """ results={'success' : False, 'kappa' : 0, 'mae' : 0} try: cv_preds=util_functions.gen_cv_preds(clf,feats,scores) err=numpy.mean(numpy.abs(numpy.array(cv_preds)-scores)) kappa=util_functions.quadratic_weighted_kappa(list(cv_preds),scores) results['mae']=err results['kappa']=kappa results['success']=True except ValueError as ex: # If this is hit, everything is fine. It is hard to explain why the error occurs, but it isn't a big deal. msg = u"Not enough classes (0,1,etc) in each cross validation fold: {ex}".format(ex=ex) log.debug(msg) except: log.exception("Error getting cv error estimates.") return results
def get_cv_error(clf,feats,scores): """ Gets cross validated error for a given classifier, set of features, and scores clf - classifier feats - features to feed into the classified and cross validate over scores - scores associated with the features -- feature row 1 associates with score 1, etc. """ results={'success' : False, 'kappa' : 0, 'mae' : 0} try: cv_preds=util_functions.gen_cv_preds(clf,feats,scores) err=numpy.mean(numpy.abs(numpy.array(cv_preds)-scores)) kappa=util_functions.quadratic_weighted_kappa(list(cv_preds),scores) results['mae']=err results['kappa']=kappa results['success']=True except ValueError as ex: # If this is hit, everything is fine. It is hard to explain why the error occurs, but it isn't a big deal. msg = u"Not enough classes (0,1,etc) in each cross validation fold: {ex}".format(ex=ex) log.debug(msg) except: log.exception("Error getting cv error estimates.") return results
[ "Gets", "cross", "validated", "error", "for", "a", "given", "classifier", "set", "of", "features", "and", "scores", "clf", "-", "classifier", "feats", "-", "features", "to", "feed", "into", "the", "classified", "and", "cross", "validate", "over", "scores", "...
edx/ease
python
https://github.com/edx/ease/blob/a7890ed403da94d03726b0639cd8ebda45af6bbb/ease/model_creator.py#L89-L111
[ "def", "get_cv_error", "(", "clf", ",", "feats", ",", "scores", ")", ":", "results", "=", "{", "'success'", ":", "False", ",", "'kappa'", ":", "0", ",", "'mae'", ":", "0", "}", "try", ":", "cv_preds", "=", "util_functions", ".", "gen_cv_preds", "(", ...
a7890ed403da94d03726b0639cd8ebda45af6bbb
valid
get_algorithms
Gets two classifiers for each type of algorithm, and returns them. First for predicting, second for cv error. type - one of util_functions.AlgorithmTypes
ease/model_creator.py
def get_algorithms(algorithm): """ Gets two classifiers for each type of algorithm, and returns them. First for predicting, second for cv error. type - one of util_functions.AlgorithmTypes """ if algorithm == util_functions.AlgorithmTypes.classification: clf = sklearn.ensemble.GradientBoostingClassifier(n_estimators=100, learn_rate=.05, max_depth=4, random_state=1,min_samples_leaf=3) clf2=sklearn.ensemble.GradientBoostingClassifier(n_estimators=100, learn_rate=.05, max_depth=4, random_state=1,min_samples_leaf=3) else: clf = sklearn.ensemble.GradientBoostingRegressor(n_estimators=100, learn_rate=.05, max_depth=4, random_state=1,min_samples_leaf=3) clf2=sklearn.ensemble.GradientBoostingRegressor(n_estimators=100, learn_rate=.05, max_depth=4, random_state=1,min_samples_leaf=3) return clf, clf2
def get_algorithms(algorithm): """ Gets two classifiers for each type of algorithm, and returns them. First for predicting, second for cv error. type - one of util_functions.AlgorithmTypes """ if algorithm == util_functions.AlgorithmTypes.classification: clf = sklearn.ensemble.GradientBoostingClassifier(n_estimators=100, learn_rate=.05, max_depth=4, random_state=1,min_samples_leaf=3) clf2=sklearn.ensemble.GradientBoostingClassifier(n_estimators=100, learn_rate=.05, max_depth=4, random_state=1,min_samples_leaf=3) else: clf = sklearn.ensemble.GradientBoostingRegressor(n_estimators=100, learn_rate=.05, max_depth=4, random_state=1,min_samples_leaf=3) clf2=sklearn.ensemble.GradientBoostingRegressor(n_estimators=100, learn_rate=.05, max_depth=4, random_state=1,min_samples_leaf=3) return clf, clf2
[ "Gets", "two", "classifiers", "for", "each", "type", "of", "algorithm", "and", "returns", "them", ".", "First", "for", "predicting", "second", "for", "cv", "error", ".", "type", "-", "one", "of", "util_functions", ".", "AlgorithmTypes" ]
edx/ease
python
https://github.com/edx/ease/blob/a7890ed403da94d03726b0639cd8ebda45af6bbb/ease/model_creator.py#L113-L128
[ "def", "get_algorithms", "(", "algorithm", ")", ":", "if", "algorithm", "==", "util_functions", ".", "AlgorithmTypes", ".", "classification", ":", "clf", "=", "sklearn", ".", "ensemble", ".", "GradientBoostingClassifier", "(", "n_estimators", "=", "100", ",", "l...
a7890ed403da94d03726b0639cd8ebda45af6bbb
valid
extract_features_and_generate_model_predictors
Extracts features and generates predictors based on a given predictor set predictor_set - a PredictorSet object that has been initialized with data type - one of util_functions.AlgorithmType
ease/model_creator.py
def extract_features_and_generate_model_predictors(predictor_set, algorithm=util_functions.AlgorithmTypes.regression): """ Extracts features and generates predictors based on a given predictor set predictor_set - a PredictorSet object that has been initialized with data type - one of util_functions.AlgorithmType """ if(algorithm not in [util_functions.AlgorithmTypes.regression, util_functions.AlgorithmTypes.classification]): algorithm = util_functions.AlgorithmTypes.regression f = predictor_extractor.PredictorExtractor() f.initialize_dictionaries(predictor_set) train_feats = f.gen_feats(predictor_set) clf,clf2 = get_algorithms(algorithm) cv_error_results=get_cv_error(clf2,train_feats,predictor_set._target) try: set_score = numpy.asarray(predictor_set._target, dtype=numpy.int) clf.fit(train_feats, set_score) except ValueError: log.exception("Not enough classes (0,1,etc) in sample.") set_score = predictor_set._target set_score[0]=1 set_score[1]=0 clf.fit(train_feats, set_score) return f, clf, cv_error_results
def extract_features_and_generate_model_predictors(predictor_set, algorithm=util_functions.AlgorithmTypes.regression): """ Extracts features and generates predictors based on a given predictor set predictor_set - a PredictorSet object that has been initialized with data type - one of util_functions.AlgorithmType """ if(algorithm not in [util_functions.AlgorithmTypes.regression, util_functions.AlgorithmTypes.classification]): algorithm = util_functions.AlgorithmTypes.regression f = predictor_extractor.PredictorExtractor() f.initialize_dictionaries(predictor_set) train_feats = f.gen_feats(predictor_set) clf,clf2 = get_algorithms(algorithm) cv_error_results=get_cv_error(clf2,train_feats,predictor_set._target) try: set_score = numpy.asarray(predictor_set._target, dtype=numpy.int) clf.fit(train_feats, set_score) except ValueError: log.exception("Not enough classes (0,1,etc) in sample.") set_score = predictor_set._target set_score[0]=1 set_score[1]=0 clf.fit(train_feats, set_score) return f, clf, cv_error_results
[ "Extracts", "features", "and", "generates", "predictors", "based", "on", "a", "given", "predictor", "set", "predictor_set", "-", "a", "PredictorSet", "object", "that", "has", "been", "initialized", "with", "data", "type", "-", "one", "of", "util_functions", ".",...
edx/ease
python
https://github.com/edx/ease/blob/a7890ed403da94d03726b0639cd8ebda45af6bbb/ease/model_creator.py#L131-L158
[ "def", "extract_features_and_generate_model_predictors", "(", "predictor_set", ",", "algorithm", "=", "util_functions", ".", "AlgorithmTypes", ".", "regression", ")", ":", "if", "(", "algorithm", "not", "in", "[", "util_functions", ".", "AlgorithmTypes", ".", "regress...
a7890ed403da94d03726b0639cd8ebda45af6bbb
valid
extract_features_and_generate_model
Feed in an essay set to get feature vector and classifier essays must be an essay set object additional array is an optional argument that can specify a numpy array of values to add in returns a trained FeatureExtractor object and a trained classifier
ease/model_creator.py
def extract_features_and_generate_model(essays, algorithm=util_functions.AlgorithmTypes.regression): """ Feed in an essay set to get feature vector and classifier essays must be an essay set object additional array is an optional argument that can specify a numpy array of values to add in returns a trained FeatureExtractor object and a trained classifier """ f = feature_extractor.FeatureExtractor() f.initialize_dictionaries(essays) train_feats = f.gen_feats(essays) set_score = numpy.asarray(essays._score, dtype=numpy.int) if len(util_functions.f7(list(set_score)))>5: algorithm = util_functions.AlgorithmTypes.regression else: algorithm = util_functions.AlgorithmTypes.classification clf,clf2 = get_algorithms(algorithm) cv_error_results=get_cv_error(clf2,train_feats,essays._score) try: clf.fit(train_feats, set_score) except ValueError: log.exception("Not enough classes (0,1,etc) in sample.") set_score[0]=1 set_score[1]=0 clf.fit(train_feats, set_score) return f, clf, cv_error_results
def extract_features_and_generate_model(essays, algorithm=util_functions.AlgorithmTypes.regression): """ Feed in an essay set to get feature vector and classifier essays must be an essay set object additional array is an optional argument that can specify a numpy array of values to add in returns a trained FeatureExtractor object and a trained classifier """ f = feature_extractor.FeatureExtractor() f.initialize_dictionaries(essays) train_feats = f.gen_feats(essays) set_score = numpy.asarray(essays._score, dtype=numpy.int) if len(util_functions.f7(list(set_score)))>5: algorithm = util_functions.AlgorithmTypes.regression else: algorithm = util_functions.AlgorithmTypes.classification clf,clf2 = get_algorithms(algorithm) cv_error_results=get_cv_error(clf2,train_feats,essays._score) try: clf.fit(train_feats, set_score) except ValueError: log.exception("Not enough classes (0,1,etc) in sample.") set_score[0]=1 set_score[1]=0 clf.fit(train_feats, set_score) return f, clf, cv_error_results
[ "Feed", "in", "an", "essay", "set", "to", "get", "feature", "vector", "and", "classifier", "essays", "must", "be", "an", "essay", "set", "object", "additional", "array", "is", "an", "optional", "argument", "that", "can", "specify", "a", "numpy", "array", "...
edx/ease
python
https://github.com/edx/ease/blob/a7890ed403da94d03726b0639cd8ebda45af6bbb/ease/model_creator.py#L161-L192
[ "def", "extract_features_and_generate_model", "(", "essays", ",", "algorithm", "=", "util_functions", ".", "AlgorithmTypes", ".", "regression", ")", ":", "f", "=", "feature_extractor", ".", "FeatureExtractor", "(", ")", "f", ".", "initialize_dictionaries", "(", "ess...
a7890ed403da94d03726b0639cd8ebda45af6bbb
valid
dump_model_to_file
Writes out a model to a file. prompt string is a string containing the prompt feature_ext is a trained FeatureExtractor object classifier is a trained classifier model_path is the path of write out the model file to
ease/model_creator.py
def dump_model_to_file(prompt_string, feature_ext, classifier, text, score, model_path): """ Writes out a model to a file. prompt string is a string containing the prompt feature_ext is a trained FeatureExtractor object classifier is a trained classifier model_path is the path of write out the model file to """ model_file = {'prompt': prompt_string, 'extractor': feature_ext, 'model': classifier, 'text' : text, 'score' : score} pickle.dump(model_file, file=open(model_path, "w"))
def dump_model_to_file(prompt_string, feature_ext, classifier, text, score, model_path): """ Writes out a model to a file. prompt string is a string containing the prompt feature_ext is a trained FeatureExtractor object classifier is a trained classifier model_path is the path of write out the model file to """ model_file = {'prompt': prompt_string, 'extractor': feature_ext, 'model': classifier, 'text' : text, 'score' : score} pickle.dump(model_file, file=open(model_path, "w"))
[ "Writes", "out", "a", "model", "to", "a", "file", ".", "prompt", "string", "is", "a", "string", "containing", "the", "prompt", "feature_ext", "is", "a", "trained", "FeatureExtractor", "object", "classifier", "is", "a", "trained", "classifier", "model_path", "i...
edx/ease
python
https://github.com/edx/ease/blob/a7890ed403da94d03726b0639cd8ebda45af6bbb/ease/model_creator.py#L194-L203
[ "def", "dump_model_to_file", "(", "prompt_string", ",", "feature_ext", ",", "classifier", ",", "text", ",", "score", ",", "model_path", ")", ":", "model_file", "=", "{", "'prompt'", ":", "prompt_string", ",", "'extractor'", ":", "feature_ext", ",", "'model'", ...
a7890ed403da94d03726b0639cd8ebda45af6bbb
valid
create_essay_set_and_dump_model
Function that creates essay set, extracts features, and writes out model See above functions for argument descriptions
ease/model_creator.py
def create_essay_set_and_dump_model(text,score,prompt,model_path,additional_array=None): """ Function that creates essay set, extracts features, and writes out model See above functions for argument descriptions """ essay_set=create_essay_set(text,score,prompt) feature_ext,clf=extract_features_and_generate_model(essay_set,additional_array) dump_model_to_file(prompt,feature_ext,clf,model_path)
def create_essay_set_and_dump_model(text,score,prompt,model_path,additional_array=None): """ Function that creates essay set, extracts features, and writes out model See above functions for argument descriptions """ essay_set=create_essay_set(text,score,prompt) feature_ext,clf=extract_features_and_generate_model(essay_set,additional_array) dump_model_to_file(prompt,feature_ext,clf,model_path)
[ "Function", "that", "creates", "essay", "set", "extracts", "features", "and", "writes", "out", "model", "See", "above", "functions", "for", "argument", "descriptions" ]
edx/ease
python
https://github.com/edx/ease/blob/a7890ed403da94d03726b0639cd8ebda45af6bbb/ease/model_creator.py#L205-L212
[ "def", "create_essay_set_and_dump_model", "(", "text", ",", "score", ",", "prompt", ",", "model_path", ",", "additional_array", "=", "None", ")", ":", "essay_set", "=", "create_essay_set", "(", "text", ",", "score", ",", "prompt", ")", "feature_ext", ",", "clf...
a7890ed403da94d03726b0639cd8ebda45af6bbb
valid
PredictorExtractor.initialize_dictionaries
Initialize dictionaries with the textual inputs in the PredictorSet object p_set - PredictorSet object that has had data fed in
ease/predictor_extractor.py
def initialize_dictionaries(self, p_set): """ Initialize dictionaries with the textual inputs in the PredictorSet object p_set - PredictorSet object that has had data fed in """ success = False if not (hasattr(p_set, '_type')): error_message = "needs to be an essay set of the train type." log.exception(error_message) raise util_functions.InputError(p_set, error_message) if not (p_set._type == "train"): error_message = "needs to be an essay set of the train type." log.exception(error_message) raise util_functions.InputError(p_set, error_message) div_length=len(p_set._essay_sets) if div_length==0: div_length=1 #Ensures that even with a large amount of input textual features, training time stays reasonable max_feats2 = int(math.floor(200/div_length)) for i in xrange(0,len(p_set._essay_sets)): self._extractors.append(FeatureExtractor()) self._extractors[i].initialize_dictionaries(p_set._essay_sets[i], max_feats2=max_feats2) self._initialized = True success = True return success
def initialize_dictionaries(self, p_set): """ Initialize dictionaries with the textual inputs in the PredictorSet object p_set - PredictorSet object that has had data fed in """ success = False if not (hasattr(p_set, '_type')): error_message = "needs to be an essay set of the train type." log.exception(error_message) raise util_functions.InputError(p_set, error_message) if not (p_set._type == "train"): error_message = "needs to be an essay set of the train type." log.exception(error_message) raise util_functions.InputError(p_set, error_message) div_length=len(p_set._essay_sets) if div_length==0: div_length=1 #Ensures that even with a large amount of input textual features, training time stays reasonable max_feats2 = int(math.floor(200/div_length)) for i in xrange(0,len(p_set._essay_sets)): self._extractors.append(FeatureExtractor()) self._extractors[i].initialize_dictionaries(p_set._essay_sets[i], max_feats2=max_feats2) self._initialized = True success = True return success
[ "Initialize", "dictionaries", "with", "the", "textual", "inputs", "in", "the", "PredictorSet", "object", "p_set", "-", "PredictorSet", "object", "that", "has", "had", "data", "fed", "in" ]
edx/ease
python
https://github.com/edx/ease/blob/a7890ed403da94d03726b0639cd8ebda45af6bbb/ease/predictor_extractor.py#L35-L62
[ "def", "initialize_dictionaries", "(", "self", ",", "p_set", ")", ":", "success", "=", "False", "if", "not", "(", "hasattr", "(", "p_set", ",", "'_type'", ")", ")", ":", "error_message", "=", "\"needs to be an essay set of the train type.\"", "log", ".", "except...
a7890ed403da94d03726b0639cd8ebda45af6bbb
valid
PredictorExtractor.gen_feats
Generates features based on an iput p_set p_set - PredictorSet
ease/predictor_extractor.py
def gen_feats(self, p_set): """ Generates features based on an iput p_set p_set - PredictorSet """ if self._initialized!=True: error_message = "Dictionaries have not been initialized." log.exception(error_message) raise util_functions.InputError(p_set, error_message) textual_features = [] for i in xrange(0,len(p_set._essay_sets)): textual_features.append(self._extractors[i].gen_feats(p_set._essay_sets[i])) textual_matrix = numpy.concatenate(textual_features, axis=1) predictor_matrix = numpy.array(p_set._numeric_features) print textual_matrix.shape print predictor_matrix.shape overall_matrix = numpy.concatenate((textual_matrix, predictor_matrix), axis=1) return overall_matrix.copy()
def gen_feats(self, p_set): """ Generates features based on an iput p_set p_set - PredictorSet """ if self._initialized!=True: error_message = "Dictionaries have not been initialized." log.exception(error_message) raise util_functions.InputError(p_set, error_message) textual_features = [] for i in xrange(0,len(p_set._essay_sets)): textual_features.append(self._extractors[i].gen_feats(p_set._essay_sets[i])) textual_matrix = numpy.concatenate(textual_features, axis=1) predictor_matrix = numpy.array(p_set._numeric_features) print textual_matrix.shape print predictor_matrix.shape overall_matrix = numpy.concatenate((textual_matrix, predictor_matrix), axis=1) return overall_matrix.copy()
[ "Generates", "features", "based", "on", "an", "iput", "p_set", "p_set", "-", "PredictorSet" ]
edx/ease
python
https://github.com/edx/ease/blob/a7890ed403da94d03726b0639cd8ebda45af6bbb/ease/predictor_extractor.py#L64-L86
[ "def", "gen_feats", "(", "self", ",", "p_set", ")", ":", "if", "self", ".", "_initialized", "!=", "True", ":", "error_message", "=", "\"Dictionaries have not been initialized.\"", "log", ".", "exception", "(", "error_message", ")", "raise", "util_functions", ".", ...
a7890ed403da94d03726b0639cd8ebda45af6bbb
valid
Lexer.t_text
r':\s*<text>
spdx/parsers/lexers/tagvalue.py
def t_text(self, t): r':\s*<text>' t.lexer.text_start = t.lexer.lexpos - len('<text>') t.lexer.begin('text')
def t_text(self, t): r':\s*<text>' t.lexer.text_start = t.lexer.lexpos - len('<text>') t.lexer.begin('text')
[ "r", ":", "\\", "s", "*", "<text", ">" ]
spdx/tools-python
python
https://github.com/spdx/tools-python/blob/301d72f6ae57c832c1da7f6402fa49b192de6810/spdx/parsers/lexers/tagvalue.py#L99-L102
[ "def", "t_text", "(", "self", ",", "t", ")", ":", "t", ".", "lexer", ".", "text_start", "=", "t", ".", "lexer", ".", "lexpos", "-", "len", "(", "'<text>'", ")", "t", ".", "lexer", ".", "begin", "(", "'text'", ")" ]
301d72f6ae57c832c1da7f6402fa49b192de6810
valid
Lexer.t_text_end
r'</text>\s*
spdx/parsers/lexers/tagvalue.py
def t_text_end(self, t): r'</text>\s*' t.type = 'TEXT' t.value = t.lexer.lexdata[ t.lexer.text_start:t.lexer.lexpos] t.lexer.lineno += t.value.count('\n') t.value = t.value.strip() t.lexer.begin('INITIAL') return t
def t_text_end(self, t): r'</text>\s*' t.type = 'TEXT' t.value = t.lexer.lexdata[ t.lexer.text_start:t.lexer.lexpos] t.lexer.lineno += t.value.count('\n') t.value = t.value.strip() t.lexer.begin('INITIAL') return t
[ "r", "<", "/", "text", ">", "\\", "s", "*" ]
spdx/tools-python
python
https://github.com/spdx/tools-python/blob/301d72f6ae57c832c1da7f6402fa49b192de6810/spdx/parsers/lexers/tagvalue.py#L104-L112
[ "def", "t_text_end", "(", "self", ",", "t", ")", ":", "t", ".", "type", "=", "'TEXT'", "t", ".", "value", "=", "t", ".", "lexer", ".", "lexdata", "[", "t", ".", "lexer", ".", "text_start", ":", "t", ".", "lexer", ".", "lexpos", "]", "t", ".", ...
301d72f6ae57c832c1da7f6402fa49b192de6810
valid
Lexer.t_KEYWORD_AS_TAG
r'[a-zA-Z]+
spdx/parsers/lexers/tagvalue.py
def t_KEYWORD_AS_TAG(self, t): r'[a-zA-Z]+' t.type = self.reserved.get(t.value, 'UNKNOWN_TAG') t.value = t.value.strip() return t
def t_KEYWORD_AS_TAG(self, t): r'[a-zA-Z]+' t.type = self.reserved.get(t.value, 'UNKNOWN_TAG') t.value = t.value.strip() return t
[ "r", "[", "a", "-", "zA", "-", "Z", "]", "+" ]
spdx/tools-python
python
https://github.com/spdx/tools-python/blob/301d72f6ae57c832c1da7f6402fa49b192de6810/spdx/parsers/lexers/tagvalue.py#L161-L165
[ "def", "t_KEYWORD_AS_TAG", "(", "self", ",", "t", ")", ":", "t", ".", "type", "=", "self", ".", "reserved", ".", "get", "(", "t", ".", "value", ",", "'UNKNOWN_TAG'", ")", "t", ".", "value", "=", "t", ".", "value", ".", "strip", "(", ")", "return"...
301d72f6ae57c832c1da7f6402fa49b192de6810
valid
Lexer.t_LINE_OR_KEYWORD_VALUE
r':.+
spdx/parsers/lexers/tagvalue.py
def t_LINE_OR_KEYWORD_VALUE(self, t): r':.+' t.value = t.value[1:].strip() if t.value in self.reserved.keys(): t.type = self.reserved[t.value] else: t.type = 'LINE' return t
def t_LINE_OR_KEYWORD_VALUE(self, t): r':.+' t.value = t.value[1:].strip() if t.value in self.reserved.keys(): t.type = self.reserved[t.value] else: t.type = 'LINE' return t
[ "r", ":", ".", "+" ]
spdx/tools-python
python
https://github.com/spdx/tools-python/blob/301d72f6ae57c832c1da7f6402fa49b192de6810/spdx/parsers/lexers/tagvalue.py#L167-L174
[ "def", "t_LINE_OR_KEYWORD_VALUE", "(", "self", ",", "t", ")", ":", "t", ".", "value", "=", "t", ".", "value", "[", "1", ":", "]", ".", "strip", "(", ")", "if", "t", ".", "value", "in", "self", ".", "reserved", ".", "keys", "(", ")", ":", "t", ...
301d72f6ae57c832c1da7f6402fa49b192de6810
valid
tv_to_rdf
Convert a SPDX file from tag/value format to RDF format. Return True on sucess, False otherwise.
spdx/tv_to_rdf.py
def tv_to_rdf(infile_name, outfile_name): """ Convert a SPDX file from tag/value format to RDF format. Return True on sucess, False otherwise. """ parser = Parser(Builder(), StandardLogger()) parser.build() with open(infile_name) as infile: data = infile.read() document, error = parser.parse(data) if not error: with open(outfile_name, mode='w') as outfile: write_document(document, outfile) return True else: print('Errors encountered while parsing RDF file.') messages = [] document.validate(messages) print('\n'.join(messages)) return False
def tv_to_rdf(infile_name, outfile_name): """ Convert a SPDX file from tag/value format to RDF format. Return True on sucess, False otherwise. """ parser = Parser(Builder(), StandardLogger()) parser.build() with open(infile_name) as infile: data = infile.read() document, error = parser.parse(data) if not error: with open(outfile_name, mode='w') as outfile: write_document(document, outfile) return True else: print('Errors encountered while parsing RDF file.') messages = [] document.validate(messages) print('\n'.join(messages)) return False
[ "Convert", "a", "SPDX", "file", "from", "tag", "/", "value", "format", "to", "RDF", "format", ".", "Return", "True", "on", "sucess", "False", "otherwise", "." ]
spdx/tools-python
python
https://github.com/spdx/tools-python/blob/301d72f6ae57c832c1da7f6402fa49b192de6810/spdx/tv_to_rdf.py#L27-L46
[ "def", "tv_to_rdf", "(", "infile_name", ",", "outfile_name", ")", ":", "parser", "=", "Parser", "(", "Builder", "(", ")", ",", "StandardLogger", "(", ")", ")", "parser", ".", "build", "(", ")", "with", "open", "(", "infile_name", ")", "as", "infile", "...
301d72f6ae57c832c1da7f6402fa49b192de6810
valid
Parser.order_error
Reports an OrderError. Error message will state that first_tag came before second_tag.
spdx/parsers/tagvalue.py
def order_error(self, first_tag, second_tag, line): """Reports an OrderError. Error message will state that first_tag came before second_tag. """ self.error = True msg = ERROR_MESSAGES['A_BEFORE_B'].format(first_tag, second_tag, line) self.logger.log(msg)
def order_error(self, first_tag, second_tag, line): """Reports an OrderError. Error message will state that first_tag came before second_tag. """ self.error = True msg = ERROR_MESSAGES['A_BEFORE_B'].format(first_tag, second_tag, line) self.logger.log(msg)
[ "Reports", "an", "OrderError", ".", "Error", "message", "will", "state", "that", "first_tag", "came", "before", "second_tag", "." ]
spdx/tools-python
python
https://github.com/spdx/tools-python/blob/301d72f6ae57c832c1da7f6402fa49b192de6810/spdx/parsers/tagvalue.py#L186-L192
[ "def", "order_error", "(", "self", ",", "first_tag", ",", "second_tag", ",", "line", ")", ":", "self", ".", "error", "=", "True", "msg", "=", "ERROR_MESSAGES", "[", "'A_BEFORE_B'", "]", ".", "format", "(", "first_tag", ",", "second_tag", ",", "line", ")"...
301d72f6ae57c832c1da7f6402fa49b192de6810
valid
Parser.p_lic_xref_1
lic_xref : LICS_CRS_REF LINE
spdx/parsers/tagvalue.py
def p_lic_xref_1(self, p): """lic_xref : LICS_CRS_REF LINE""" try: if six.PY2: value = p[2].decode(encoding='utf-8') else: value = p[2] self.builder.add_lic_xref(self.document, value) except OrderError: self.order_error('LicenseCrossReference', 'LicenseName', p.lineno(1))
def p_lic_xref_1(self, p): """lic_xref : LICS_CRS_REF LINE""" try: if six.PY2: value = p[2].decode(encoding='utf-8') else: value = p[2] self.builder.add_lic_xref(self.document, value) except OrderError: self.order_error('LicenseCrossReference', 'LicenseName', p.lineno(1))
[ "lic_xref", ":", "LICS_CRS_REF", "LINE" ]
spdx/tools-python
python
https://github.com/spdx/tools-python/blob/301d72f6ae57c832c1da7f6402fa49b192de6810/spdx/parsers/tagvalue.py#L194-L203
[ "def", "p_lic_xref_1", "(", "self", ",", "p", ")", ":", "try", ":", "if", "six", ".", "PY2", ":", "value", "=", "p", "[", "2", "]", ".", "decode", "(", "encoding", "=", "'utf-8'", ")", "else", ":", "value", "=", "p", "[", "2", "]", "self", "....
301d72f6ae57c832c1da7f6402fa49b192de6810
valid
Parser.p_lic_comment_1
lic_comment : LICS_COMMENT TEXT
spdx/parsers/tagvalue.py
def p_lic_comment_1(self, p): """lic_comment : LICS_COMMENT TEXT""" try: if six.PY2: value = p[2].decode(encoding='utf-8') else: value = p[2] self.builder.set_lic_comment(self.document, value) except OrderError: self.order_error('LicenseComment', 'LicenseID', p.lineno(1)) except CardinalityError: self.more_than_one_error('LicenseComment', p.lineno(1))
def p_lic_comment_1(self, p): """lic_comment : LICS_COMMENT TEXT""" try: if six.PY2: value = p[2].decode(encoding='utf-8') else: value = p[2] self.builder.set_lic_comment(self.document, value) except OrderError: self.order_error('LicenseComment', 'LicenseID', p.lineno(1)) except CardinalityError: self.more_than_one_error('LicenseComment', p.lineno(1))
[ "lic_comment", ":", "LICS_COMMENT", "TEXT" ]
spdx/tools-python
python
https://github.com/spdx/tools-python/blob/301d72f6ae57c832c1da7f6402fa49b192de6810/spdx/parsers/tagvalue.py#L211-L222
[ "def", "p_lic_comment_1", "(", "self", ",", "p", ")", ":", "try", ":", "if", "six", ".", "PY2", ":", "value", "=", "p", "[", "2", "]", ".", "decode", "(", "encoding", "=", "'utf-8'", ")", "else", ":", "value", "=", "p", "[", "2", "]", "self", ...
301d72f6ae57c832c1da7f6402fa49b192de6810
valid
Parser.p_extr_lic_name_1
extr_lic_name : LICS_NAME extr_lic_name_value
spdx/parsers/tagvalue.py
def p_extr_lic_name_1(self, p): """extr_lic_name : LICS_NAME extr_lic_name_value""" try: self.builder.set_lic_name(self.document, p[2]) except OrderError: self.order_error('LicenseName', 'LicenseID', p.lineno(1)) except CardinalityError: self.more_than_one_error('LicenseName', p.lineno(1))
def p_extr_lic_name_1(self, p): """extr_lic_name : LICS_NAME extr_lic_name_value""" try: self.builder.set_lic_name(self.document, p[2]) except OrderError: self.order_error('LicenseName', 'LicenseID', p.lineno(1)) except CardinalityError: self.more_than_one_error('LicenseName', p.lineno(1))
[ "extr_lic_name", ":", "LICS_NAME", "extr_lic_name_value" ]
spdx/tools-python
python
https://github.com/spdx/tools-python/blob/301d72f6ae57c832c1da7f6402fa49b192de6810/spdx/parsers/tagvalue.py#L230-L237
[ "def", "p_extr_lic_name_1", "(", "self", ",", "p", ")", ":", "try", ":", "self", ".", "builder", ".", "set_lic_name", "(", "self", ".", "document", ",", "p", "[", "2", "]", ")", "except", "OrderError", ":", "self", ".", "order_error", "(", "'LicenseNam...
301d72f6ae57c832c1da7f6402fa49b192de6810
valid
Parser.p_extr_lic_name_value_1
extr_lic_name_value : LINE
spdx/parsers/tagvalue.py
def p_extr_lic_name_value_1(self, p): """extr_lic_name_value : LINE""" if six.PY2: p[0] = p[1].decode(encoding='utf-8') else: p[0] = p[1]
def p_extr_lic_name_value_1(self, p): """extr_lic_name_value : LINE""" if six.PY2: p[0] = p[1].decode(encoding='utf-8') else: p[0] = p[1]
[ "extr_lic_name_value", ":", "LINE" ]
spdx/tools-python
python
https://github.com/spdx/tools-python/blob/301d72f6ae57c832c1da7f6402fa49b192de6810/spdx/parsers/tagvalue.py#L245-L250
[ "def", "p_extr_lic_name_value_1", "(", "self", ",", "p", ")", ":", "if", "six", ".", "PY2", ":", "p", "[", "0", "]", "=", "p", "[", "1", "]", ".", "decode", "(", "encoding", "=", "'utf-8'", ")", "else", ":", "p", "[", "0", "]", "=", "p", "[",...
301d72f6ae57c832c1da7f6402fa49b192de6810
valid
Parser.p_extr_lic_text_1
extr_lic_text : LICS_TEXT TEXT
spdx/parsers/tagvalue.py
def p_extr_lic_text_1(self, p): """extr_lic_text : LICS_TEXT TEXT""" try: if six.PY2: value = p[2].decode(encoding='utf-8') else: value = p[2] self.builder.set_lic_text(self.document, value) except OrderError: self.order_error('ExtractedText', 'LicenseID', p.lineno(1)) except CardinalityError: self.more_than_one_error('ExtractedText', p.lineno(1))
def p_extr_lic_text_1(self, p): """extr_lic_text : LICS_TEXT TEXT""" try: if six.PY2: value = p[2].decode(encoding='utf-8') else: value = p[2] self.builder.set_lic_text(self.document, value) except OrderError: self.order_error('ExtractedText', 'LicenseID', p.lineno(1)) except CardinalityError: self.more_than_one_error('ExtractedText', p.lineno(1))
[ "extr_lic_text", ":", "LICS_TEXT", "TEXT" ]
spdx/tools-python
python
https://github.com/spdx/tools-python/blob/301d72f6ae57c832c1da7f6402fa49b192de6810/spdx/parsers/tagvalue.py#L256-L267
[ "def", "p_extr_lic_text_1", "(", "self", ",", "p", ")", ":", "try", ":", "if", "six", ".", "PY2", ":", "value", "=", "p", "[", "2", "]", ".", "decode", "(", "encoding", "=", "'utf-8'", ")", "else", ":", "value", "=", "p", "[", "2", "]", "self",...
301d72f6ae57c832c1da7f6402fa49b192de6810
valid
Parser.p_extr_lic_id_1
extr_lic_id : LICS_ID LINE
spdx/parsers/tagvalue.py
def p_extr_lic_id_1(self, p): """extr_lic_id : LICS_ID LINE""" try: if six.PY2: value = p[2].decode(encoding='utf-8') else: value = p[2] self.builder.set_lic_id(self.document, value) except SPDXValueError: self.error = True msg = ERROR_MESSAGES['LICS_ID_VALUE'].format(p.lineno(1)) self.logger.log(msg)
def p_extr_lic_id_1(self, p): """extr_lic_id : LICS_ID LINE""" try: if six.PY2: value = p[2].decode(encoding='utf-8') else: value = p[2] self.builder.set_lic_id(self.document, value) except SPDXValueError: self.error = True msg = ERROR_MESSAGES['LICS_ID_VALUE'].format(p.lineno(1)) self.logger.log(msg)
[ "extr_lic_id", ":", "LICS_ID", "LINE" ]
spdx/tools-python
python
https://github.com/spdx/tools-python/blob/301d72f6ae57c832c1da7f6402fa49b192de6810/spdx/parsers/tagvalue.py#L275-L286
[ "def", "p_extr_lic_id_1", "(", "self", ",", "p", ")", ":", "try", ":", "if", "six", ".", "PY2", ":", "value", "=", "p", "[", "2", "]", ".", "decode", "(", "encoding", "=", "'utf-8'", ")", "else", ":", "value", "=", "p", "[", "2", "]", "self", ...
301d72f6ae57c832c1da7f6402fa49b192de6810
valid
Parser.p_prj_uri_art_3
prj_uri_art : ART_PRJ_URI error
spdx/parsers/tagvalue.py
def p_prj_uri_art_3(self, p): """prj_uri_art : ART_PRJ_URI error""" self.error = True msg = ERROR_MESSAGES['ART_PRJ_URI_VALUE'].format(p.lineno(1)) self.logger.log(msg)
def p_prj_uri_art_3(self, p): """prj_uri_art : ART_PRJ_URI error""" self.error = True msg = ERROR_MESSAGES['ART_PRJ_URI_VALUE'].format(p.lineno(1)) self.logger.log(msg)
[ "prj_uri_art", ":", "ART_PRJ_URI", "error" ]
spdx/tools-python
python
https://github.com/spdx/tools-python/blob/301d72f6ae57c832c1da7f6402fa49b192de6810/spdx/parsers/tagvalue.py#L339-L343
[ "def", "p_prj_uri_art_3", "(", "self", ",", "p", ")", ":", "self", ".", "error", "=", "True", "msg", "=", "ERROR_MESSAGES", "[", "'ART_PRJ_URI_VALUE'", "]", ".", "format", "(", "p", ".", "lineno", "(", "1", ")", ")", "self", ".", "logger", ".", "log"...
301d72f6ae57c832c1da7f6402fa49b192de6810
valid
Parser.p_prj_home_art_1
prj_home_art : ART_PRJ_HOME LINE
spdx/parsers/tagvalue.py
def p_prj_home_art_1(self, p): """prj_home_art : ART_PRJ_HOME LINE""" try: self.builder.set_file_atrificat_of_project(self.document, 'home', p[2]) except OrderError: self.order_error('ArtificatOfProjectHomePage', 'FileName', p.lineno(1))
def p_prj_home_art_1(self, p): """prj_home_art : ART_PRJ_HOME LINE""" try: self.builder.set_file_atrificat_of_project(self.document, 'home', p[2]) except OrderError: self.order_error('ArtificatOfProjectHomePage', 'FileName', p.lineno(1))
[ "prj_home_art", ":", "ART_PRJ_HOME", "LINE" ]
spdx/tools-python
python
https://github.com/spdx/tools-python/blob/301d72f6ae57c832c1da7f6402fa49b192de6810/spdx/parsers/tagvalue.py#L345-L350
[ "def", "p_prj_home_art_1", "(", "self", ",", "p", ")", ":", "try", ":", "self", ".", "builder", ".", "set_file_atrificat_of_project", "(", "self", ".", "document", ",", "'home'", ",", "p", "[", "2", "]", ")", "except", "OrderError", ":", "self", ".", "...
301d72f6ae57c832c1da7f6402fa49b192de6810
valid
Parser.p_prj_home_art_2
prj_home_art : ART_PRJ_HOME UN_KNOWN
spdx/parsers/tagvalue.py
def p_prj_home_art_2(self, p): """prj_home_art : ART_PRJ_HOME UN_KNOWN""" try: self.builder.set_file_atrificat_of_project(self.document, 'home', utils.UnKnown()) except OrderError: self.order_error('ArtifactOfProjectName', 'FileName', p.lineno(1))
def p_prj_home_art_2(self, p): """prj_home_art : ART_PRJ_HOME UN_KNOWN""" try: self.builder.set_file_atrificat_of_project(self.document, 'home', utils.UnKnown()) except OrderError: self.order_error('ArtifactOfProjectName', 'FileName', p.lineno(1))
[ "prj_home_art", ":", "ART_PRJ_HOME", "UN_KNOWN" ]
spdx/tools-python
python
https://github.com/spdx/tools-python/blob/301d72f6ae57c832c1da7f6402fa49b192de6810/spdx/parsers/tagvalue.py#L352-L358
[ "def", "p_prj_home_art_2", "(", "self", ",", "p", ")", ":", "try", ":", "self", ".", "builder", ".", "set_file_atrificat_of_project", "(", "self", ".", "document", ",", "'home'", ",", "utils", ".", "UnKnown", "(", ")", ")", "except", "OrderError", ":", "...
301d72f6ae57c832c1da7f6402fa49b192de6810
valid
Parser.p_prj_name_art_1
prj_name_art : ART_PRJ_NAME LINE
spdx/parsers/tagvalue.py
def p_prj_name_art_1(self, p): """prj_name_art : ART_PRJ_NAME LINE""" try: if six.PY2: value = p[2].decode(encoding='utf-8') else: value = p[2] self.builder.set_file_atrificat_of_project(self.document, 'name', value) except OrderError: self.order_error('ArtifactOfProjectName', 'FileName', p.lineno(1))
def p_prj_name_art_1(self, p): """prj_name_art : ART_PRJ_NAME LINE""" try: if six.PY2: value = p[2].decode(encoding='utf-8') else: value = p[2] self.builder.set_file_atrificat_of_project(self.document, 'name', value) except OrderError: self.order_error('ArtifactOfProjectName', 'FileName', p.lineno(1))
[ "prj_name_art", ":", "ART_PRJ_NAME", "LINE" ]
spdx/tools-python
python
https://github.com/spdx/tools-python/blob/301d72f6ae57c832c1da7f6402fa49b192de6810/spdx/parsers/tagvalue.py#L366-L375
[ "def", "p_prj_name_art_1", "(", "self", ",", "p", ")", ":", "try", ":", "if", "six", ".", "PY2", ":", "value", "=", "p", "[", "2", "]", ".", "decode", "(", "encoding", "=", "'utf-8'", ")", "else", ":", "value", "=", "p", "[", "2", "]", "self", ...
301d72f6ae57c832c1da7f6402fa49b192de6810
valid
Parser.p_file_dep_1
file_dep : FILE_DEP LINE
spdx/parsers/tagvalue.py
def p_file_dep_1(self, p): """file_dep : FILE_DEP LINE""" try: if six.PY2: value = p[2].decode(encoding='utf-8') else: value = p[2] self.builder.add_file_dep(self.document, value) except OrderError: self.order_error('FileDependency', 'FileName', p.lineno(1))
def p_file_dep_1(self, p): """file_dep : FILE_DEP LINE""" try: if six.PY2: value = p[2].decode(encoding='utf-8') else: value = p[2] self.builder.add_file_dep(self.document, value) except OrderError: self.order_error('FileDependency', 'FileName', p.lineno(1))
[ "file_dep", ":", "FILE_DEP", "LINE" ]
spdx/tools-python
python
https://github.com/spdx/tools-python/blob/301d72f6ae57c832c1da7f6402fa49b192de6810/spdx/parsers/tagvalue.py#L383-L392
[ "def", "p_file_dep_1", "(", "self", ",", "p", ")", ":", "try", ":", "if", "six", ".", "PY2", ":", "value", "=", "p", "[", "2", "]", ".", "decode", "(", "encoding", "=", "'utf-8'", ")", "else", ":", "value", "=", "p", "[", "2", "]", "self", "....
301d72f6ae57c832c1da7f6402fa49b192de6810
valid
Parser.p_file_contrib_1
file_contrib : FILE_CONTRIB LINE
spdx/parsers/tagvalue.py
def p_file_contrib_1(self, p): """file_contrib : FILE_CONTRIB LINE""" try: if six.PY2: value = p[2].decode(encoding='utf-8') else: value = p[2] self.builder.add_file_contribution(self.document, value) except OrderError: self.order_error('FileContributor', 'FileName', p.lineno(1))
def p_file_contrib_1(self, p): """file_contrib : FILE_CONTRIB LINE""" try: if six.PY2: value = p[2].decode(encoding='utf-8') else: value = p[2] self.builder.add_file_contribution(self.document, value) except OrderError: self.order_error('FileContributor', 'FileName', p.lineno(1))
[ "file_contrib", ":", "FILE_CONTRIB", "LINE" ]
spdx/tools-python
python
https://github.com/spdx/tools-python/blob/301d72f6ae57c832c1da7f6402fa49b192de6810/spdx/parsers/tagvalue.py#L400-L409
[ "def", "p_file_contrib_1", "(", "self", ",", "p", ")", ":", "try", ":", "if", "six", ".", "PY2", ":", "value", "=", "p", "[", "2", "]", ".", "decode", "(", "encoding", "=", "'utf-8'", ")", "else", ":", "value", "=", "p", "[", "2", "]", "self", ...
301d72f6ae57c832c1da7f6402fa49b192de6810
valid
Parser.p_file_notice_1
file_notice : FILE_NOTICE TEXT
spdx/parsers/tagvalue.py
def p_file_notice_1(self, p): """file_notice : FILE_NOTICE TEXT""" try: if six.PY2: value = p[2].decode(encoding='utf-8') else: value = p[2] self.builder.set_file_notice(self.document, value) except OrderError: self.order_error('FileNotice', 'FileName', p.lineno(1)) except CardinalityError: self.more_than_one_error('FileNotice', p.lineno(1))
def p_file_notice_1(self, p): """file_notice : FILE_NOTICE TEXT""" try: if six.PY2: value = p[2].decode(encoding='utf-8') else: value = p[2] self.builder.set_file_notice(self.document, value) except OrderError: self.order_error('FileNotice', 'FileName', p.lineno(1)) except CardinalityError: self.more_than_one_error('FileNotice', p.lineno(1))
[ "file_notice", ":", "FILE_NOTICE", "TEXT" ]
spdx/tools-python
python
https://github.com/spdx/tools-python/blob/301d72f6ae57c832c1da7f6402fa49b192de6810/spdx/parsers/tagvalue.py#L417-L428
[ "def", "p_file_notice_1", "(", "self", ",", "p", ")", ":", "try", ":", "if", "six", ".", "PY2", ":", "value", "=", "p", "[", "2", "]", ".", "decode", "(", "encoding", "=", "'utf-8'", ")", "else", ":", "value", "=", "p", "[", "2", "]", "self", ...
301d72f6ae57c832c1da7f6402fa49b192de6810
valid
Parser.p_file_cr_text_1
file_cr_text : FILE_CR_TEXT file_cr_value
spdx/parsers/tagvalue.py
def p_file_cr_text_1(self, p): """file_cr_text : FILE_CR_TEXT file_cr_value""" try: self.builder.set_file_copyright(self.document, p[2]) except OrderError: self.order_error('FileCopyrightText', 'FileName', p.lineno(1)) except CardinalityError: self.more_than_one_error('FileCopyrightText', p.lineno(1))
def p_file_cr_text_1(self, p): """file_cr_text : FILE_CR_TEXT file_cr_value""" try: self.builder.set_file_copyright(self.document, p[2]) except OrderError: self.order_error('FileCopyrightText', 'FileName', p.lineno(1)) except CardinalityError: self.more_than_one_error('FileCopyrightText', p.lineno(1))
[ "file_cr_text", ":", "FILE_CR_TEXT", "file_cr_value" ]
spdx/tools-python
python
https://github.com/spdx/tools-python/blob/301d72f6ae57c832c1da7f6402fa49b192de6810/spdx/parsers/tagvalue.py#L436-L443
[ "def", "p_file_cr_text_1", "(", "self", ",", "p", ")", ":", "try", ":", "self", ".", "builder", ".", "set_file_copyright", "(", "self", ".", "document", ",", "p", "[", "2", "]", ")", "except", "OrderError", ":", "self", ".", "order_error", "(", "'FileC...
301d72f6ae57c832c1da7f6402fa49b192de6810
valid
Parser.p_file_cr_value_1
file_cr_value : TEXT
spdx/parsers/tagvalue.py
def p_file_cr_value_1(self, p): """file_cr_value : TEXT""" if six.PY2: p[0] = p[1].decode(encoding='utf-8') else: p[0] = p[1]
def p_file_cr_value_1(self, p): """file_cr_value : TEXT""" if six.PY2: p[0] = p[1].decode(encoding='utf-8') else: p[0] = p[1]
[ "file_cr_value", ":", "TEXT" ]
spdx/tools-python
python
https://github.com/spdx/tools-python/blob/301d72f6ae57c832c1da7f6402fa49b192de6810/spdx/parsers/tagvalue.py#L451-L456
[ "def", "p_file_cr_value_1", "(", "self", ",", "p", ")", ":", "if", "six", ".", "PY2", ":", "p", "[", "0", "]", "=", "p", "[", "1", "]", ".", "decode", "(", "encoding", "=", "'utf-8'", ")", "else", ":", "p", "[", "0", "]", "=", "p", "[", "1"...
301d72f6ae57c832c1da7f6402fa49b192de6810
valid
Parser.p_file_lics_comment_1
file_lics_comment : FILE_LICS_COMMENT TEXT
spdx/parsers/tagvalue.py
def p_file_lics_comment_1(self, p): """file_lics_comment : FILE_LICS_COMMENT TEXT""" try: if six.PY2: value = p[2].decode(encoding='utf-8') else: value = p[2] self.builder.set_file_license_comment(self.document, value) except OrderError: self.order_error('LicenseComments', 'FileName', p.lineno(1)) except CardinalityError: self.more_than_one_error('LicenseComments', p.lineno(1))
def p_file_lics_comment_1(self, p): """file_lics_comment : FILE_LICS_COMMENT TEXT""" try: if six.PY2: value = p[2].decode(encoding='utf-8') else: value = p[2] self.builder.set_file_license_comment(self.document, value) except OrderError: self.order_error('LicenseComments', 'FileName', p.lineno(1)) except CardinalityError: self.more_than_one_error('LicenseComments', p.lineno(1))
[ "file_lics_comment", ":", "FILE_LICS_COMMENT", "TEXT" ]
spdx/tools-python
python
https://github.com/spdx/tools-python/blob/301d72f6ae57c832c1da7f6402fa49b192de6810/spdx/parsers/tagvalue.py#L466-L477
[ "def", "p_file_lics_comment_1", "(", "self", ",", "p", ")", ":", "try", ":", "if", "six", ".", "PY2", ":", "value", "=", "p", "[", "2", "]", ".", "decode", "(", "encoding", "=", "'utf-8'", ")", "else", ":", "value", "=", "p", "[", "2", "]", "se...
301d72f6ae57c832c1da7f6402fa49b192de6810
valid
Parser.p_file_lics_info_1
file_lics_info : FILE_LICS_INFO file_lic_info_value
spdx/parsers/tagvalue.py
def p_file_lics_info_1(self, p): """file_lics_info : FILE_LICS_INFO file_lic_info_value""" try: self.builder.set_file_license_in_file(self.document, p[2]) except OrderError: self.order_error('LicenseInfoInFile', 'FileName', p.lineno(1)) except SPDXValueError: self.error = True msg = ERROR_MESSAGES['FILE_LICS_INFO_VALUE'].format(p.lineno(1)) self.logger.log(msg)
def p_file_lics_info_1(self, p): """file_lics_info : FILE_LICS_INFO file_lic_info_value""" try: self.builder.set_file_license_in_file(self.document, p[2]) except OrderError: self.order_error('LicenseInfoInFile', 'FileName', p.lineno(1)) except SPDXValueError: self.error = True msg = ERROR_MESSAGES['FILE_LICS_INFO_VALUE'].format(p.lineno(1)) self.logger.log(msg)
[ "file_lics_info", ":", "FILE_LICS_INFO", "file_lic_info_value" ]
spdx/tools-python
python
https://github.com/spdx/tools-python/blob/301d72f6ae57c832c1da7f6402fa49b192de6810/spdx/parsers/tagvalue.py#L485-L494
[ "def", "p_file_lics_info_1", "(", "self", ",", "p", ")", ":", "try", ":", "self", ".", "builder", ".", "set_file_license_in_file", "(", "self", ".", "document", ",", "p", "[", "2", "]", ")", "except", "OrderError", ":", "self", ".", "order_error", "(", ...
301d72f6ae57c832c1da7f6402fa49b192de6810
valid
Parser.p_conc_license_3
conc_license : LINE
spdx/parsers/tagvalue.py
def p_conc_license_3(self, p): """conc_license : LINE""" if six.PY2: value = p[1].decode(encoding='utf-8') else: value = p[1] ref_re = re.compile('LicenseRef-.+', re.UNICODE) if (p[1] in config.LICENSE_MAP.keys()) or (ref_re.match(p[1]) is not None): p[0] = document.License.from_identifier(value) else: p[0] = self.license_list_parser.parse(value)
def p_conc_license_3(self, p): """conc_license : LINE""" if six.PY2: value = p[1].decode(encoding='utf-8') else: value = p[1] ref_re = re.compile('LicenseRef-.+', re.UNICODE) if (p[1] in config.LICENSE_MAP.keys()) or (ref_re.match(p[1]) is not None): p[0] = document.License.from_identifier(value) else: p[0] = self.license_list_parser.parse(value)
[ "conc_license", ":", "LINE" ]
spdx/tools-python
python
https://github.com/spdx/tools-python/blob/301d72f6ae57c832c1da7f6402fa49b192de6810/spdx/parsers/tagvalue.py#L527-L537
[ "def", "p_conc_license_3", "(", "self", ",", "p", ")", ":", "if", "six", ".", "PY2", ":", "value", "=", "p", "[", "1", "]", ".", "decode", "(", "encoding", "=", "'utf-8'", ")", "else", ":", "value", "=", "p", "[", "1", "]", "ref_re", "=", "re",...
301d72f6ae57c832c1da7f6402fa49b192de6810
valid
Parser.p_file_name_1
file_name : FILE_NAME LINE
spdx/parsers/tagvalue.py
def p_file_name_1(self, p): """file_name : FILE_NAME LINE""" try: if six.PY2: value = p[2].decode(encoding='utf-8') else: value = p[2] self.builder.set_file_name(self.document, value) except OrderError: self.order_error('FileName', 'PackageName', p.lineno(1))
def p_file_name_1(self, p): """file_name : FILE_NAME LINE""" try: if six.PY2: value = p[2].decode(encoding='utf-8') else: value = p[2] self.builder.set_file_name(self.document, value) except OrderError: self.order_error('FileName', 'PackageName', p.lineno(1))
[ "file_name", ":", "FILE_NAME", "LINE" ]
spdx/tools-python
python
https://github.com/spdx/tools-python/blob/301d72f6ae57c832c1da7f6402fa49b192de6810/spdx/parsers/tagvalue.py#L539-L548
[ "def", "p_file_name_1", "(", "self", ",", "p", ")", ":", "try", ":", "if", "six", ".", "PY2", ":", "value", "=", "p", "[", "2", "]", ".", "decode", "(", "encoding", "=", "'utf-8'", ")", "else", ":", "value", "=", "p", "[", "2", "]", "self", "...
301d72f6ae57c832c1da7f6402fa49b192de6810
valid
Parser.p_spdx_id
spdx_id : SPDX_ID LINE
spdx/parsers/tagvalue.py
def p_spdx_id(self, p): """spdx_id : SPDX_ID LINE""" if six.PY2: value = p[2].decode(encoding='utf-8') else: value = p[2] if not self.builder.doc_spdx_id_set: self.builder.set_doc_spdx_id(self.document, value) else: self.builder.set_file_spdx_id(self.document, value)
def p_spdx_id(self, p): """spdx_id : SPDX_ID LINE""" if six.PY2: value = p[2].decode(encoding='utf-8') else: value = p[2] if not self.builder.doc_spdx_id_set: self.builder.set_doc_spdx_id(self.document, value) else: self.builder.set_file_spdx_id(self.document, value)
[ "spdx_id", ":", "SPDX_ID", "LINE" ]
spdx/tools-python
python
https://github.com/spdx/tools-python/blob/301d72f6ae57c832c1da7f6402fa49b192de6810/spdx/parsers/tagvalue.py#L556-L565
[ "def", "p_spdx_id", "(", "self", ",", "p", ")", ":", "if", "six", ".", "PY2", ":", "value", "=", "p", "[", "2", "]", ".", "decode", "(", "encoding", "=", "'utf-8'", ")", "else", ":", "value", "=", "p", "[", "2", "]", "if", "not", "self", ".",...
301d72f6ae57c832c1da7f6402fa49b192de6810
valid
Parser.p_file_comment_1
file_comment : FILE_COMMENT TEXT
spdx/parsers/tagvalue.py
def p_file_comment_1(self, p): """file_comment : FILE_COMMENT TEXT""" try: if six.PY2: value = p[2].decode(encoding='utf-8') else: value = p[2] self.builder.set_file_comment(self.document, value) except OrderError: self.order_error('FileComment', 'FileName', p.lineno(1)) except CardinalityError: self.more_than_one_error('FileComment', p.lineno(1))
def p_file_comment_1(self, p): """file_comment : FILE_COMMENT TEXT""" try: if six.PY2: value = p[2].decode(encoding='utf-8') else: value = p[2] self.builder.set_file_comment(self.document, value) except OrderError: self.order_error('FileComment', 'FileName', p.lineno(1)) except CardinalityError: self.more_than_one_error('FileComment', p.lineno(1))
[ "file_comment", ":", "FILE_COMMENT", "TEXT" ]
spdx/tools-python
python
https://github.com/spdx/tools-python/blob/301d72f6ae57c832c1da7f6402fa49b192de6810/spdx/parsers/tagvalue.py#L567-L578
[ "def", "p_file_comment_1", "(", "self", ",", "p", ")", ":", "try", ":", "if", "six", ".", "PY2", ":", "value", "=", "p", "[", "2", "]", ".", "decode", "(", "encoding", "=", "'utf-8'", ")", "else", ":", "value", "=", "p", "[", "2", "]", "self", ...
301d72f6ae57c832c1da7f6402fa49b192de6810
valid
Parser.p_file_type_1
file_type : FILE_TYPE file_type_value
spdx/parsers/tagvalue.py
def p_file_type_1(self, p): """file_type : FILE_TYPE file_type_value""" try: self.builder.set_file_type(self.document, p[2]) except OrderError: self.order_error('FileType', 'FileName', p.lineno(1)) except CardinalityError: self.more_than_one_error('FileType', p.lineno(1))
def p_file_type_1(self, p): """file_type : FILE_TYPE file_type_value""" try: self.builder.set_file_type(self.document, p[2]) except OrderError: self.order_error('FileType', 'FileName', p.lineno(1)) except CardinalityError: self.more_than_one_error('FileType', p.lineno(1))
[ "file_type", ":", "FILE_TYPE", "file_type_value" ]
spdx/tools-python
python
https://github.com/spdx/tools-python/blob/301d72f6ae57c832c1da7f6402fa49b192de6810/spdx/parsers/tagvalue.py#L586-L593
[ "def", "p_file_type_1", "(", "self", ",", "p", ")", ":", "try", ":", "self", ".", "builder", ".", "set_file_type", "(", "self", ".", "document", ",", "p", "[", "2", "]", ")", "except", "OrderError", ":", "self", ".", "order_error", "(", "'FileType'", ...
301d72f6ae57c832c1da7f6402fa49b192de6810
valid
Parser.p_file_chksum_1
file_chksum : FILE_CHKSUM CHKSUM
spdx/parsers/tagvalue.py
def p_file_chksum_1(self, p): """file_chksum : FILE_CHKSUM CHKSUM""" try: if six.PY2: value = p[2].decode(encoding='utf-8') else: value = p[2] self.builder.set_file_chksum(self.document, value) except OrderError: self.order_error('FileChecksum', 'FileName', p.lineno(1)) except CardinalityError: self.more_than_one_error('FileChecksum', p.lineno(1))
def p_file_chksum_1(self, p): """file_chksum : FILE_CHKSUM CHKSUM""" try: if six.PY2: value = p[2].decode(encoding='utf-8') else: value = p[2] self.builder.set_file_chksum(self.document, value) except OrderError: self.order_error('FileChecksum', 'FileName', p.lineno(1)) except CardinalityError: self.more_than_one_error('FileChecksum', p.lineno(1))
[ "file_chksum", ":", "FILE_CHKSUM", "CHKSUM" ]
spdx/tools-python
python
https://github.com/spdx/tools-python/blob/301d72f6ae57c832c1da7f6402fa49b192de6810/spdx/parsers/tagvalue.py#L601-L612
[ "def", "p_file_chksum_1", "(", "self", ",", "p", ")", ":", "try", ":", "if", "six", ".", "PY2", ":", "value", "=", "p", "[", "2", "]", ".", "decode", "(", "encoding", "=", "'utf-8'", ")", "else", ":", "value", "=", "p", "[", "2", "]", "self", ...
301d72f6ae57c832c1da7f6402fa49b192de6810
valid
Parser.p_file_conc_1
file_conc : FILE_LICS_CONC conc_license
spdx/parsers/tagvalue.py
def p_file_conc_1(self, p): """file_conc : FILE_LICS_CONC conc_license""" try: self.builder.set_concluded_license(self.document, p[2]) except SPDXValueError: self.error = True msg = ERROR_MESSAGES['FILE_LICS_CONC_VALUE'].format(p.lineno(1)) self.logger.log(msg) except OrderError: self.order_error('LicenseConcluded', 'FileName', p.lineno(1)) except CardinalityError: self.more_than_one_error('LicenseConcluded', p.lineno(1))
def p_file_conc_1(self, p): """file_conc : FILE_LICS_CONC conc_license""" try: self.builder.set_concluded_license(self.document, p[2]) except SPDXValueError: self.error = True msg = ERROR_MESSAGES['FILE_LICS_CONC_VALUE'].format(p.lineno(1)) self.logger.log(msg) except OrderError: self.order_error('LicenseConcluded', 'FileName', p.lineno(1)) except CardinalityError: self.more_than_one_error('LicenseConcluded', p.lineno(1))
[ "file_conc", ":", "FILE_LICS_CONC", "conc_license" ]
spdx/tools-python
python
https://github.com/spdx/tools-python/blob/301d72f6ae57c832c1da7f6402fa49b192de6810/spdx/parsers/tagvalue.py#L620-L631
[ "def", "p_file_conc_1", "(", "self", ",", "p", ")", ":", "try", ":", "self", ".", "builder", ".", "set_concluded_license", "(", "self", ".", "document", ",", "p", "[", "2", "]", ")", "except", "SPDXValueError", ":", "self", ".", "error", "=", "True", ...
301d72f6ae57c832c1da7f6402fa49b192de6810
valid
Parser.p_file_type_value
file_type_value : OTHER | SOURCE | ARCHIVE | BINARY
spdx/parsers/tagvalue.py
def p_file_type_value(self, p): """file_type_value : OTHER | SOURCE | ARCHIVE | BINARY """ if six.PY2: p[0] = p[1].decode(encoding='utf-8') else: p[0] = p[1]
def p_file_type_value(self, p): """file_type_value : OTHER | SOURCE | ARCHIVE | BINARY """ if six.PY2: p[0] = p[1].decode(encoding='utf-8') else: p[0] = p[1]
[ "file_type_value", ":", "OTHER", "|", "SOURCE", "|", "ARCHIVE", "|", "BINARY" ]
spdx/tools-python
python
https://github.com/spdx/tools-python/blob/301d72f6ae57c832c1da7f6402fa49b192de6810/spdx/parsers/tagvalue.py#L639-L648
[ "def", "p_file_type_value", "(", "self", ",", "p", ")", ":", "if", "six", ".", "PY2", ":", "p", "[", "0", "]", "=", "p", "[", "1", "]", ".", "decode", "(", "encoding", "=", "'utf-8'", ")", "else", ":", "p", "[", "0", "]", "=", "p", "[", "1"...
301d72f6ae57c832c1da7f6402fa49b192de6810
valid
Parser.p_pkg_desc_1
pkg_desc : PKG_DESC TEXT
spdx/parsers/tagvalue.py
def p_pkg_desc_1(self, p): """pkg_desc : PKG_DESC TEXT""" try: if six.PY2: value = p[2].decode(encoding='utf-8') else: value = p[2] self.builder.set_pkg_desc(self.document, value) except CardinalityError: self.more_than_one_error('PackageDescription', p.lineno(1)) except OrderError: self.order_error('PackageDescription', 'PackageFileName', p.lineno(1))
def p_pkg_desc_1(self, p): """pkg_desc : PKG_DESC TEXT""" try: if six.PY2: value = p[2].decode(encoding='utf-8') else: value = p[2] self.builder.set_pkg_desc(self.document, value) except CardinalityError: self.more_than_one_error('PackageDescription', p.lineno(1)) except OrderError: self.order_error('PackageDescription', 'PackageFileName', p.lineno(1))
[ "pkg_desc", ":", "PKG_DESC", "TEXT" ]
spdx/tools-python
python
https://github.com/spdx/tools-python/blob/301d72f6ae57c832c1da7f6402fa49b192de6810/spdx/parsers/tagvalue.py#L650-L661
[ "def", "p_pkg_desc_1", "(", "self", ",", "p", ")", ":", "try", ":", "if", "six", ".", "PY2", ":", "value", "=", "p", "[", "2", "]", ".", "decode", "(", "encoding", "=", "'utf-8'", ")", "else", ":", "value", "=", "p", "[", "2", "]", "self", "....
301d72f6ae57c832c1da7f6402fa49b192de6810
valid
Parser.p_pkg_summary_1
pkg_summary : PKG_SUM TEXT
spdx/parsers/tagvalue.py
def p_pkg_summary_1(self, p): """pkg_summary : PKG_SUM TEXT""" try: if six.PY2: value = p[2].decode(encoding='utf-8') else: value = p[2] self.builder.set_pkg_summary(self.document, value) except OrderError: self.order_error('PackageSummary', 'PackageFileName', p.lineno(1)) except CardinalityError: self.more_than_one_error('PackageSummary', p.lineno(1))
def p_pkg_summary_1(self, p): """pkg_summary : PKG_SUM TEXT""" try: if six.PY2: value = p[2].decode(encoding='utf-8') else: value = p[2] self.builder.set_pkg_summary(self.document, value) except OrderError: self.order_error('PackageSummary', 'PackageFileName', p.lineno(1)) except CardinalityError: self.more_than_one_error('PackageSummary', p.lineno(1))
[ "pkg_summary", ":", "PKG_SUM", "TEXT" ]
spdx/tools-python
python
https://github.com/spdx/tools-python/blob/301d72f6ae57c832c1da7f6402fa49b192de6810/spdx/parsers/tagvalue.py#L669-L680
[ "def", "p_pkg_summary_1", "(", "self", ",", "p", ")", ":", "try", ":", "if", "six", ".", "PY2", ":", "value", "=", "p", "[", "2", "]", ".", "decode", "(", "encoding", "=", "'utf-8'", ")", "else", ":", "value", "=", "p", "[", "2", "]", "self", ...
301d72f6ae57c832c1da7f6402fa49b192de6810
valid
Parser.p_pkg_cr_text_1
pkg_cr_text : PKG_CPY_TEXT pkg_cr_text_value
spdx/parsers/tagvalue.py
def p_pkg_cr_text_1(self, p): """pkg_cr_text : PKG_CPY_TEXT pkg_cr_text_value""" try: self.builder.set_pkg_cr_text(self.document, p[2]) except OrderError: self.order_error('PackageCopyrightText', 'PackageFileName', p.lineno(1)) except CardinalityError: self.more_than_one_error('PackageCopyrightText', p.lineno(1))
def p_pkg_cr_text_1(self, p): """pkg_cr_text : PKG_CPY_TEXT pkg_cr_text_value""" try: self.builder.set_pkg_cr_text(self.document, p[2]) except OrderError: self.order_error('PackageCopyrightText', 'PackageFileName', p.lineno(1)) except CardinalityError: self.more_than_one_error('PackageCopyrightText', p.lineno(1))
[ "pkg_cr_text", ":", "PKG_CPY_TEXT", "pkg_cr_text_value" ]
spdx/tools-python
python
https://github.com/spdx/tools-python/blob/301d72f6ae57c832c1da7f6402fa49b192de6810/spdx/parsers/tagvalue.py#L688-L695
[ "def", "p_pkg_cr_text_1", "(", "self", ",", "p", ")", ":", "try", ":", "self", ".", "builder", ".", "set_pkg_cr_text", "(", "self", ".", "document", ",", "p", "[", "2", "]", ")", "except", "OrderError", ":", "self", ".", "order_error", "(", "'PackageCo...
301d72f6ae57c832c1da7f6402fa49b192de6810
valid
Parser.p_pkg_cr_text_value_1
pkg_cr_text_value : TEXT
spdx/parsers/tagvalue.py
def p_pkg_cr_text_value_1(self, p): """pkg_cr_text_value : TEXT""" if six.PY2: p[0] = p[1].decode(encoding='utf-8') else: p[0] = p[1]
def p_pkg_cr_text_value_1(self, p): """pkg_cr_text_value : TEXT""" if six.PY2: p[0] = p[1].decode(encoding='utf-8') else: p[0] = p[1]
[ "pkg_cr_text_value", ":", "TEXT" ]
spdx/tools-python
python
https://github.com/spdx/tools-python/blob/301d72f6ae57c832c1da7f6402fa49b192de6810/spdx/parsers/tagvalue.py#L703-L708
[ "def", "p_pkg_cr_text_value_1", "(", "self", ",", "p", ")", ":", "if", "six", ".", "PY2", ":", "p", "[", "0", "]", "=", "p", "[", "1", "]", ".", "decode", "(", "encoding", "=", "'utf-8'", ")", "else", ":", "p", "[", "0", "]", "=", "p", "[", ...
301d72f6ae57c832c1da7f6402fa49b192de6810
valid
Parser.p_pkg_lic_comment_1
pkg_lic_comment : PKG_LICS_COMMENT TEXT
spdx/parsers/tagvalue.py
def p_pkg_lic_comment_1(self, p): """pkg_lic_comment : PKG_LICS_COMMENT TEXT""" try: if six.PY2: value = p[2].decode(encoding='utf-8') else: value = p[2] self.builder.set_pkg_license_comment(self.document, value) except OrderError: self.order_error('PackageLicenseComments', 'PackageFileName', p.lineno(1)) except CardinalityError: self.more_than_one_error('PackageLicenseComments', p.lineno(1))
def p_pkg_lic_comment_1(self, p): """pkg_lic_comment : PKG_LICS_COMMENT TEXT""" try: if six.PY2: value = p[2].decode(encoding='utf-8') else: value = p[2] self.builder.set_pkg_license_comment(self.document, value) except OrderError: self.order_error('PackageLicenseComments', 'PackageFileName', p.lineno(1)) except CardinalityError: self.more_than_one_error('PackageLicenseComments', p.lineno(1))
[ "pkg_lic_comment", ":", "PKG_LICS_COMMENT", "TEXT" ]
spdx/tools-python
python
https://github.com/spdx/tools-python/blob/301d72f6ae57c832c1da7f6402fa49b192de6810/spdx/parsers/tagvalue.py#L718-L729
[ "def", "p_pkg_lic_comment_1", "(", "self", ",", "p", ")", ":", "try", ":", "if", "six", ".", "PY2", ":", "value", "=", "p", "[", "2", "]", ".", "decode", "(", "encoding", "=", "'utf-8'", ")", "else", ":", "value", "=", "p", "[", "2", "]", "self...
301d72f6ae57c832c1da7f6402fa49b192de6810
valid
Parser.p_pkg_lic_ff_1
pkg_lic_ff : PKG_LICS_FFILE pkg_lic_ff_value
spdx/parsers/tagvalue.py
def p_pkg_lic_ff_1(self, p): """pkg_lic_ff : PKG_LICS_FFILE pkg_lic_ff_value""" try: self.builder.set_pkg_license_from_file(self.document, p[2]) except OrderError: self.order_error('PackageLicenseInfoFromFiles', 'PackageName', p.lineno(1)) except SPDXValueError: self.error = True msg = ERROR_MESSAGES['PKG_LIC_FFILE_VALUE'].format(p.lineno(1)) self.logger.log(msg)
def p_pkg_lic_ff_1(self, p): """pkg_lic_ff : PKG_LICS_FFILE pkg_lic_ff_value""" try: self.builder.set_pkg_license_from_file(self.document, p[2]) except OrderError: self.order_error('PackageLicenseInfoFromFiles', 'PackageName', p.lineno(1)) except SPDXValueError: self.error = True msg = ERROR_MESSAGES['PKG_LIC_FFILE_VALUE'].format(p.lineno(1)) self.logger.log(msg)
[ "pkg_lic_ff", ":", "PKG_LICS_FFILE", "pkg_lic_ff_value" ]
spdx/tools-python
python
https://github.com/spdx/tools-python/blob/301d72f6ae57c832c1da7f6402fa49b192de6810/spdx/parsers/tagvalue.py#L756-L765
[ "def", "p_pkg_lic_ff_1", "(", "self", ",", "p", ")", ":", "try", ":", "self", ".", "builder", ".", "set_pkg_license_from_file", "(", "self", ".", "document", ",", "p", "[", "2", "]", ")", "except", "OrderError", ":", "self", ".", "order_error", "(", "'...
301d72f6ae57c832c1da7f6402fa49b192de6810
valid
Parser.p_pkg_lic_ff_value_3
pkg_lic_ff_value : LINE
spdx/parsers/tagvalue.py
def p_pkg_lic_ff_value_3(self, p): """pkg_lic_ff_value : LINE""" if six.PY2: value = p[1].decode(encoding='utf-8') else: value = p[1] p[0] = document.License.from_identifier(value)
def p_pkg_lic_ff_value_3(self, p): """pkg_lic_ff_value : LINE""" if six.PY2: value = p[1].decode(encoding='utf-8') else: value = p[1] p[0] = document.License.from_identifier(value)
[ "pkg_lic_ff_value", ":", "LINE" ]
spdx/tools-python
python
https://github.com/spdx/tools-python/blob/301d72f6ae57c832c1da7f6402fa49b192de6810/spdx/parsers/tagvalue.py#L775-L781
[ "def", "p_pkg_lic_ff_value_3", "(", "self", ",", "p", ")", ":", "if", "six", ".", "PY2", ":", "value", "=", "p", "[", "1", "]", ".", "decode", "(", "encoding", "=", "'utf-8'", ")", "else", ":", "value", "=", "p", "[", "1", "]", "p", "[", "0", ...
301d72f6ae57c832c1da7f6402fa49b192de6810
valid
Parser.p_pkg_src_info_1
pkg_src_info : PKG_SRC_INFO TEXT
spdx/parsers/tagvalue.py
def p_pkg_src_info_1(self, p): """pkg_src_info : PKG_SRC_INFO TEXT""" try: if six.PY2: value = p[2].decode(encoding='utf-8') else: value = p[2] self.builder.set_pkg_source_info(self.document, value) except CardinalityError: self.more_than_one_error('PackageSourceInfo', p.lineno(1)) except OrderError: self.order_error('PackageSourceInfo', 'PackageFileName', p.lineno(1))
def p_pkg_src_info_1(self, p): """pkg_src_info : PKG_SRC_INFO TEXT""" try: if six.PY2: value = p[2].decode(encoding='utf-8') else: value = p[2] self.builder.set_pkg_source_info(self.document, value) except CardinalityError: self.more_than_one_error('PackageSourceInfo', p.lineno(1)) except OrderError: self.order_error('PackageSourceInfo', 'PackageFileName', p.lineno(1))
[ "pkg_src_info", ":", "PKG_SRC_INFO", "TEXT" ]
spdx/tools-python
python
https://github.com/spdx/tools-python/blob/301d72f6ae57c832c1da7f6402fa49b192de6810/spdx/parsers/tagvalue.py#L808-L819
[ "def", "p_pkg_src_info_1", "(", "self", ",", "p", ")", ":", "try", ":", "if", "six", ".", "PY2", ":", "value", "=", "p", "[", "2", "]", ".", "decode", "(", "encoding", "=", "'utf-8'", ")", "else", ":", "value", "=", "p", "[", "2", "]", "self", ...
301d72f6ae57c832c1da7f6402fa49b192de6810
valid
Parser.p_pkg_chksum_1
pkg_chksum : PKG_CHKSUM CHKSUM
spdx/parsers/tagvalue.py
def p_pkg_chksum_1(self, p): """pkg_chksum : PKG_CHKSUM CHKSUM""" try: if six.PY2: value = p[2].decode(encoding='utf-8') else: value = p[2] self.builder.set_pkg_chk_sum(self.document, value) except OrderError: self.order_error('PackageChecksum', 'PackageFileName', p.lineno(1)) except CardinalityError: self.more_than_one_error('PackageChecksum', p.lineno(1))
def p_pkg_chksum_1(self, p): """pkg_chksum : PKG_CHKSUM CHKSUM""" try: if six.PY2: value = p[2].decode(encoding='utf-8') else: value = p[2] self.builder.set_pkg_chk_sum(self.document, value) except OrderError: self.order_error('PackageChecksum', 'PackageFileName', p.lineno(1)) except CardinalityError: self.more_than_one_error('PackageChecksum', p.lineno(1))
[ "pkg_chksum", ":", "PKG_CHKSUM", "CHKSUM" ]
spdx/tools-python
python
https://github.com/spdx/tools-python/blob/301d72f6ae57c832c1da7f6402fa49b192de6810/spdx/parsers/tagvalue.py#L827-L838
[ "def", "p_pkg_chksum_1", "(", "self", ",", "p", ")", ":", "try", ":", "if", "six", ".", "PY2", ":", "value", "=", "p", "[", "2", "]", ".", "decode", "(", "encoding", "=", "'utf-8'", ")", "else", ":", "value", "=", "p", "[", "2", "]", "self", ...
301d72f6ae57c832c1da7f6402fa49b192de6810
valid
Parser.p_pkg_home_1
pkg_home : PKG_HOME pkg_home_value
spdx/parsers/tagvalue.py
def p_pkg_home_1(self, p): """pkg_home : PKG_HOME pkg_home_value""" try: self.builder.set_pkg_down_location(self.document, p[2]) except OrderError: self.order_error('PackageHomePage', 'PackageName', p.lineno(1)) except CardinalityError: self.more_than_one_error('PackageHomePage', p.lineno(1))
def p_pkg_home_1(self, p): """pkg_home : PKG_HOME pkg_home_value""" try: self.builder.set_pkg_down_location(self.document, p[2]) except OrderError: self.order_error('PackageHomePage', 'PackageName', p.lineno(1)) except CardinalityError: self.more_than_one_error('PackageHomePage', p.lineno(1))
[ "pkg_home", ":", "PKG_HOME", "pkg_home_value" ]
spdx/tools-python
python
https://github.com/spdx/tools-python/blob/301d72f6ae57c832c1da7f6402fa49b192de6810/spdx/parsers/tagvalue.py#L869-L876
[ "def", "p_pkg_home_1", "(", "self", ",", "p", ")", ":", "try", ":", "self", ".", "builder", ".", "set_pkg_down_location", "(", "self", ".", "document", ",", "p", "[", "2", "]", ")", "except", "OrderError", ":", "self", ".", "order_error", "(", "'Packag...
301d72f6ae57c832c1da7f6402fa49b192de6810
valid
Parser.p_pkg_home_2
pkg_home : PKG_HOME error
spdx/parsers/tagvalue.py
def p_pkg_home_2(self, p): """pkg_home : PKG_HOME error""" self.error = True msg = ERROR_MESSAGES['PKG_HOME_VALUE'] self.logger.log(msg)
def p_pkg_home_2(self, p): """pkg_home : PKG_HOME error""" self.error = True msg = ERROR_MESSAGES['PKG_HOME_VALUE'] self.logger.log(msg)
[ "pkg_home", ":", "PKG_HOME", "error" ]
spdx/tools-python
python
https://github.com/spdx/tools-python/blob/301d72f6ae57c832c1da7f6402fa49b192de6810/spdx/parsers/tagvalue.py#L878-L882
[ "def", "p_pkg_home_2", "(", "self", ",", "p", ")", ":", "self", ".", "error", "=", "True", "msg", "=", "ERROR_MESSAGES", "[", "'PKG_HOME_VALUE'", "]", "self", ".", "logger", ".", "log", "(", "msg", ")" ]
301d72f6ae57c832c1da7f6402fa49b192de6810
valid
Parser.p_pkg_home_value_1
pkg_home_value : LINE
spdx/parsers/tagvalue.py
def p_pkg_home_value_1(self, p): """pkg_home_value : LINE""" if six.PY2: p[0] = p[1].decode(encoding='utf-8') else: p[0] = p[1]
def p_pkg_home_value_1(self, p): """pkg_home_value : LINE""" if six.PY2: p[0] = p[1].decode(encoding='utf-8') else: p[0] = p[1]
[ "pkg_home_value", ":", "LINE" ]
spdx/tools-python
python
https://github.com/spdx/tools-python/blob/301d72f6ae57c832c1da7f6402fa49b192de6810/spdx/parsers/tagvalue.py#L884-L889
[ "def", "p_pkg_home_value_1", "(", "self", ",", "p", ")", ":", "if", "six", ".", "PY2", ":", "p", "[", "0", "]", "=", "p", "[", "1", "]", ".", "decode", "(", "encoding", "=", "'utf-8'", ")", "else", ":", "p", "[", "0", "]", "=", "p", "[", "1...
301d72f6ae57c832c1da7f6402fa49b192de6810
valid
Parser.p_pkg_down_value_1
pkg_down_value : LINE
spdx/parsers/tagvalue.py
def p_pkg_down_value_1(self, p): """pkg_down_value : LINE """ if six.PY2: p[0] = p[1].decode(encoding='utf-8') else: p[0] = p[1]
def p_pkg_down_value_1(self, p): """pkg_down_value : LINE """ if six.PY2: p[0] = p[1].decode(encoding='utf-8') else: p[0] = p[1]
[ "pkg_down_value", ":", "LINE" ]
spdx/tools-python
python
https://github.com/spdx/tools-python/blob/301d72f6ae57c832c1da7f6402fa49b192de6810/spdx/parsers/tagvalue.py#L914-L919
[ "def", "p_pkg_down_value_1", "(", "self", ",", "p", ")", ":", "if", "six", ".", "PY2", ":", "p", "[", "0", "]", "=", "p", "[", "1", "]", ".", "decode", "(", "encoding", "=", "'utf-8'", ")", "else", ":", "p", "[", "0", "]", "=", "p", "[", "1...
301d72f6ae57c832c1da7f6402fa49b192de6810
valid
Parser.p_pkg_file_name
pkg_file_name : PKG_FILE_NAME LINE
spdx/parsers/tagvalue.py
def p_pkg_file_name(self, p): """pkg_file_name : PKG_FILE_NAME LINE""" try: if six.PY2: value = p[2].decode(encoding='utf-8') else: value = p[2] self.builder.set_pkg_file_name(self.document, value) except OrderError: self.order_error('PackageFileName', 'PackageName', p.lineno(1)) except CardinalityError: self.more_than_one_error('PackageFileName', p.lineno(1))
def p_pkg_file_name(self, p): """pkg_file_name : PKG_FILE_NAME LINE""" try: if six.PY2: value = p[2].decode(encoding='utf-8') else: value = p[2] self.builder.set_pkg_file_name(self.document, value) except OrderError: self.order_error('PackageFileName', 'PackageName', p.lineno(1)) except CardinalityError: self.more_than_one_error('PackageFileName', p.lineno(1))
[ "pkg_file_name", ":", "PKG_FILE_NAME", "LINE" ]
spdx/tools-python
python
https://github.com/spdx/tools-python/blob/301d72f6ae57c832c1da7f6402fa49b192de6810/spdx/parsers/tagvalue.py#L975-L986
[ "def", "p_pkg_file_name", "(", "self", ",", "p", ")", ":", "try", ":", "if", "six", ".", "PY2", ":", "value", "=", "p", "[", "2", "]", ".", "decode", "(", "encoding", "=", "'utf-8'", ")", "else", ":", "value", "=", "p", "[", "2", "]", "self", ...
301d72f6ae57c832c1da7f6402fa49b192de6810
valid
Parser.p_package_version_1
package_version : PKG_VERSION LINE
spdx/parsers/tagvalue.py
def p_package_version_1(self, p): """package_version : PKG_VERSION LINE""" try: if six.PY2: value = p[2].decode(encoding='utf-8') else: value = p[2] self.builder.set_pkg_vers(self.document, value) except OrderError: self.order_error('PackageVersion', 'PackageName', p.lineno(1)) except CardinalityError: self.more_than_one_error('PackageVersion', p.lineno(1))
def p_package_version_1(self, p): """package_version : PKG_VERSION LINE""" try: if six.PY2: value = p[2].decode(encoding='utf-8') else: value = p[2] self.builder.set_pkg_vers(self.document, value) except OrderError: self.order_error('PackageVersion', 'PackageName', p.lineno(1)) except CardinalityError: self.more_than_one_error('PackageVersion', p.lineno(1))
[ "package_version", ":", "PKG_VERSION", "LINE" ]
spdx/tools-python
python
https://github.com/spdx/tools-python/blob/301d72f6ae57c832c1da7f6402fa49b192de6810/spdx/parsers/tagvalue.py#L994-L1005
[ "def", "p_package_version_1", "(", "self", ",", "p", ")", ":", "try", ":", "if", "six", ".", "PY2", ":", "value", "=", "p", "[", "2", "]", ".", "decode", "(", "encoding", "=", "'utf-8'", ")", "else", ":", "value", "=", "p", "[", "2", "]", "self...
301d72f6ae57c832c1da7f6402fa49b192de6810
valid
Parser.p_package_name
package_name : PKG_NAME LINE
spdx/parsers/tagvalue.py
def p_package_name(self, p): """package_name : PKG_NAME LINE""" try: if six.PY2: value = p[2].decode(encoding='utf-8') else: value = p[2] self.builder.create_package(self.document, value) except CardinalityError: self.more_than_one_error('PackageName', p.lineno(1))
def p_package_name(self, p): """package_name : PKG_NAME LINE""" try: if six.PY2: value = p[2].decode(encoding='utf-8') else: value = p[2] self.builder.create_package(self.document, value) except CardinalityError: self.more_than_one_error('PackageName', p.lineno(1))
[ "package_name", ":", "PKG_NAME", "LINE" ]
spdx/tools-python
python
https://github.com/spdx/tools-python/blob/301d72f6ae57c832c1da7f6402fa49b192de6810/spdx/parsers/tagvalue.py#L1013-L1022
[ "def", "p_package_name", "(", "self", ",", "p", ")", ":", "try", ":", "if", "six", ".", "PY2", ":", "value", "=", "p", "[", "2", "]", ".", "decode", "(", "encoding", "=", "'utf-8'", ")", "else", ":", "value", "=", "p", "[", "2", "]", "self", ...
301d72f6ae57c832c1da7f6402fa49b192de6810
valid
Parser.p_review_date_1
review_date : REVIEW_DATE DATE
spdx/parsers/tagvalue.py
def p_review_date_1(self, p): """review_date : REVIEW_DATE DATE""" try: if six.PY2: value = p[2].decode(encoding='utf-8') else: value = p[2] self.builder.add_review_date(self.document, value) except CardinalityError: self.more_than_one_error('ReviewDate', p.lineno(1)) except OrderError: self.order_error('ReviewDate', 'Reviewer', p.lineno(1))
def p_review_date_1(self, p): """review_date : REVIEW_DATE DATE""" try: if six.PY2: value = p[2].decode(encoding='utf-8') else: value = p[2] self.builder.add_review_date(self.document, value) except CardinalityError: self.more_than_one_error('ReviewDate', p.lineno(1)) except OrderError: self.order_error('ReviewDate', 'Reviewer', p.lineno(1))
[ "review_date", ":", "REVIEW_DATE", "DATE" ]
spdx/tools-python
python
https://github.com/spdx/tools-python/blob/301d72f6ae57c832c1da7f6402fa49b192de6810/spdx/parsers/tagvalue.py#L1040-L1051
[ "def", "p_review_date_1", "(", "self", ",", "p", ")", ":", "try", ":", "if", "six", ".", "PY2", ":", "value", "=", "p", "[", "2", "]", ".", "decode", "(", "encoding", "=", "'utf-8'", ")", "else", ":", "value", "=", "p", "[", "2", "]", "self", ...
301d72f6ae57c832c1da7f6402fa49b192de6810
valid
Parser.p_review_comment_1
review_comment : REVIEW_COMMENT TEXT
spdx/parsers/tagvalue.py
def p_review_comment_1(self, p): """review_comment : REVIEW_COMMENT TEXT""" try: if six.PY2: value = p[2].decode(encoding='utf-8') else: value = p[2] self.builder.add_review_comment(self.document, value) except CardinalityError: self.more_than_one_error('ReviewComment', p.lineno(1)) except OrderError: self.order_error('ReviewComment', 'Reviewer', p.lineno(1))
def p_review_comment_1(self, p): """review_comment : REVIEW_COMMENT TEXT""" try: if six.PY2: value = p[2].decode(encoding='utf-8') else: value = p[2] self.builder.add_review_comment(self.document, value) except CardinalityError: self.more_than_one_error('ReviewComment', p.lineno(1)) except OrderError: self.order_error('ReviewComment', 'Reviewer', p.lineno(1))
[ "review_comment", ":", "REVIEW_COMMENT", "TEXT" ]
spdx/tools-python
python
https://github.com/spdx/tools-python/blob/301d72f6ae57c832c1da7f6402fa49b192de6810/spdx/parsers/tagvalue.py#L1059-L1070
[ "def", "p_review_comment_1", "(", "self", ",", "p", ")", ":", "try", ":", "if", "six", ".", "PY2", ":", "value", "=", "p", "[", "2", "]", ".", "decode", "(", "encoding", "=", "'utf-8'", ")", "else", ":", "value", "=", "p", "[", "2", "]", "self"...
301d72f6ae57c832c1da7f6402fa49b192de6810
valid
Parser.p_annotation_date_1
annotation_date : ANNOTATION_DATE DATE
spdx/parsers/tagvalue.py
def p_annotation_date_1(self, p): """annotation_date : ANNOTATION_DATE DATE""" try: if six.PY2: value = p[2].decode(encoding='utf-8') else: value = p[2] self.builder.add_annotation_date(self.document, value) except CardinalityError: self.more_than_one_error('AnnotationDate', p.lineno(1)) except OrderError: self.order_error('AnnotationDate', 'Annotator', p.lineno(1))
def p_annotation_date_1(self, p): """annotation_date : ANNOTATION_DATE DATE""" try: if six.PY2: value = p[2].decode(encoding='utf-8') else: value = p[2] self.builder.add_annotation_date(self.document, value) except CardinalityError: self.more_than_one_error('AnnotationDate', p.lineno(1)) except OrderError: self.order_error('AnnotationDate', 'Annotator', p.lineno(1))
[ "annotation_date", ":", "ANNOTATION_DATE", "DATE" ]
spdx/tools-python
python
https://github.com/spdx/tools-python/blob/301d72f6ae57c832c1da7f6402fa49b192de6810/spdx/parsers/tagvalue.py#L1088-L1099
[ "def", "p_annotation_date_1", "(", "self", ",", "p", ")", ":", "try", ":", "if", "six", ".", "PY2", ":", "value", "=", "p", "[", "2", "]", ".", "decode", "(", "encoding", "=", "'utf-8'", ")", "else", ":", "value", "=", "p", "[", "2", "]", "self...
301d72f6ae57c832c1da7f6402fa49b192de6810
valid
Parser.p_annotation_comment_1
annotation_comment : ANNOTATION_COMMENT TEXT
spdx/parsers/tagvalue.py
def p_annotation_comment_1(self, p): """annotation_comment : ANNOTATION_COMMENT TEXT""" try: if six.PY2: value = p[2].decode(encoding='utf-8') else: value = p[2] self.builder.add_annotation_comment(self.document, value) except CardinalityError: self.more_than_one_error('AnnotationComment', p.lineno(1)) except OrderError: self.order_error('AnnotationComment', 'Annotator', p.lineno(1))
def p_annotation_comment_1(self, p): """annotation_comment : ANNOTATION_COMMENT TEXT""" try: if six.PY2: value = p[2].decode(encoding='utf-8') else: value = p[2] self.builder.add_annotation_comment(self.document, value) except CardinalityError: self.more_than_one_error('AnnotationComment', p.lineno(1)) except OrderError: self.order_error('AnnotationComment', 'Annotator', p.lineno(1))
[ "annotation_comment", ":", "ANNOTATION_COMMENT", "TEXT" ]
spdx/tools-python
python
https://github.com/spdx/tools-python/blob/301d72f6ae57c832c1da7f6402fa49b192de6810/spdx/parsers/tagvalue.py#L1107-L1118
[ "def", "p_annotation_comment_1", "(", "self", ",", "p", ")", ":", "try", ":", "if", "six", ".", "PY2", ":", "value", "=", "p", "[", "2", "]", ".", "decode", "(", "encoding", "=", "'utf-8'", ")", "else", ":", "value", "=", "p", "[", "2", "]", "s...
301d72f6ae57c832c1da7f6402fa49b192de6810
valid
Parser.p_annotation_type_1
annotation_type : ANNOTATION_TYPE LINE
spdx/parsers/tagvalue.py
def p_annotation_type_1(self, p): """annotation_type : ANNOTATION_TYPE LINE""" try: if six.PY2: value = p[2].decode(encoding='utf-8') else: value = p[2] self.builder.add_annotation_type(self.document, value) except CardinalityError: self.more_than_one_error('AnnotationType', p.lineno(1)) except SPDXValueError: self.error = True msg = ERROR_MESSAGES['ANNOTATION_TYPE_VALUE'].format(p.lineno(1)) self.logger.log(msg) except OrderError: self.order_error('AnnotationType', 'Annotator', p.lineno(1))
def p_annotation_type_1(self, p): """annotation_type : ANNOTATION_TYPE LINE""" try: if six.PY2: value = p[2].decode(encoding='utf-8') else: value = p[2] self.builder.add_annotation_type(self.document, value) except CardinalityError: self.more_than_one_error('AnnotationType', p.lineno(1)) except SPDXValueError: self.error = True msg = ERROR_MESSAGES['ANNOTATION_TYPE_VALUE'].format(p.lineno(1)) self.logger.log(msg) except OrderError: self.order_error('AnnotationType', 'Annotator', p.lineno(1))
[ "annotation_type", ":", "ANNOTATION_TYPE", "LINE" ]
spdx/tools-python
python
https://github.com/spdx/tools-python/blob/301d72f6ae57c832c1da7f6402fa49b192de6810/spdx/parsers/tagvalue.py#L1126-L1141
[ "def", "p_annotation_type_1", "(", "self", ",", "p", ")", ":", "try", ":", "if", "six", ".", "PY2", ":", "value", "=", "p", "[", "2", "]", ".", "decode", "(", "encoding", "=", "'utf-8'", ")", "else", ":", "value", "=", "p", "[", "2", "]", "self...
301d72f6ae57c832c1da7f6402fa49b192de6810
valid
Parser.p_annotation_spdx_id_1
annotation_spdx_id : ANNOTATION_SPDX_ID LINE
spdx/parsers/tagvalue.py
def p_annotation_spdx_id_1(self, p): """annotation_spdx_id : ANNOTATION_SPDX_ID LINE""" try: if six.PY2: value = p[2].decode(encoding='utf-8') else: value = p[2] self.builder.set_annotation_spdx_id(self.document, value) except CardinalityError: self.more_than_one_error('SPDXREF', p.lineno(1)) except OrderError: self.order_error('SPDXREF', 'Annotator', p.lineno(1))
def p_annotation_spdx_id_1(self, p): """annotation_spdx_id : ANNOTATION_SPDX_ID LINE""" try: if six.PY2: value = p[2].decode(encoding='utf-8') else: value = p[2] self.builder.set_annotation_spdx_id(self.document, value) except CardinalityError: self.more_than_one_error('SPDXREF', p.lineno(1)) except OrderError: self.order_error('SPDXREF', 'Annotator', p.lineno(1))
[ "annotation_spdx_id", ":", "ANNOTATION_SPDX_ID", "LINE" ]
spdx/tools-python
python
https://github.com/spdx/tools-python/blob/301d72f6ae57c832c1da7f6402fa49b192de6810/spdx/parsers/tagvalue.py#L1150-L1161
[ "def", "p_annotation_spdx_id_1", "(", "self", ",", "p", ")", ":", "try", ":", "if", "six", ".", "PY2", ":", "value", "=", "p", "[", "2", "]", ".", "decode", "(", "encoding", "=", "'utf-8'", ")", "else", ":", "value", "=", "p", "[", "2", "]", "s...
301d72f6ae57c832c1da7f6402fa49b192de6810
valid
Parser.p_doc_comment_1
doc_comment : DOC_COMMENT TEXT
spdx/parsers/tagvalue.py
def p_doc_comment_1(self, p): """doc_comment : DOC_COMMENT TEXT""" try: if six.PY2: value = p[2].decode(encoding='utf-8') else: value = p[2] self.builder.set_doc_comment(self.document, value) except CardinalityError: self.more_than_one_error('DocumentComment', p.lineno(1))
def p_doc_comment_1(self, p): """doc_comment : DOC_COMMENT TEXT""" try: if six.PY2: value = p[2].decode(encoding='utf-8') else: value = p[2] self.builder.set_doc_comment(self.document, value) except CardinalityError: self.more_than_one_error('DocumentComment', p.lineno(1))
[ "doc_comment", ":", "DOC_COMMENT", "TEXT" ]
spdx/tools-python
python
https://github.com/spdx/tools-python/blob/301d72f6ae57c832c1da7f6402fa49b192de6810/spdx/parsers/tagvalue.py#L1192-L1201
[ "def", "p_doc_comment_1", "(", "self", ",", "p", ")", ":", "try", ":", "if", "six", ".", "PY2", ":", "value", "=", "p", "[", "2", "]", ".", "decode", "(", "encoding", "=", "'utf-8'", ")", "else", ":", "value", "=", "p", "[", "2", "]", "self", ...
301d72f6ae57c832c1da7f6402fa49b192de6810
valid
Parser.p_doc_name_1
doc_name : DOC_NAME LINE
spdx/parsers/tagvalue.py
def p_doc_name_1(self, p): """doc_name : DOC_NAME LINE""" try: if six.PY2: value = p[2].decode(encoding='utf-8') else: value = p[2] self.builder.set_doc_name(self.document, value) except CardinalityError: self.more_than_one_error('DocumentName', p.lineno(1))
def p_doc_name_1(self, p): """doc_name : DOC_NAME LINE""" try: if six.PY2: value = p[2].decode(encoding='utf-8') else: value = p[2] self.builder.set_doc_name(self.document, value) except CardinalityError: self.more_than_one_error('DocumentName', p.lineno(1))
[ "doc_name", ":", "DOC_NAME", "LINE" ]
spdx/tools-python
python
https://github.com/spdx/tools-python/blob/301d72f6ae57c832c1da7f6402fa49b192de6810/spdx/parsers/tagvalue.py#L1251-L1260
[ "def", "p_doc_name_1", "(", "self", ",", "p", ")", ":", "try", ":", "if", "six", ".", "PY2", ":", "value", "=", "p", "[", "2", "]", ".", "decode", "(", "encoding", "=", "'utf-8'", ")", "else", ":", "value", "=", "p", "[", "2", "]", "self", "....
301d72f6ae57c832c1da7f6402fa49b192de6810
valid
Parser.p_ext_doc_refs_1
ext_doc_ref : EXT_DOC_REF DOC_REF_ID DOC_URI EXT_DOC_REF_CHKSUM
spdx/parsers/tagvalue.py
def p_ext_doc_refs_1(self, p): """ext_doc_ref : EXT_DOC_REF DOC_REF_ID DOC_URI EXT_DOC_REF_CHKSUM""" try: if six.PY2: doc_ref_id = p[2].decode(encoding='utf-8') doc_uri = p[3].decode(encoding='utf-8') ext_doc_chksum = p[4].decode(encoding='utf-8') else: doc_ref_id = p[2] doc_uri = p[3] ext_doc_chksum = p[4] self.builder.add_ext_doc_refs(self.document, doc_ref_id, doc_uri, ext_doc_chksum) except SPDXValueError: self.error = True msg = ERROR_MESSAGES['EXT_DOC_REF_VALUE'].format(p.lineno(2)) self.logger.log(msg)
def p_ext_doc_refs_1(self, p): """ext_doc_ref : EXT_DOC_REF DOC_REF_ID DOC_URI EXT_DOC_REF_CHKSUM""" try: if six.PY2: doc_ref_id = p[2].decode(encoding='utf-8') doc_uri = p[3].decode(encoding='utf-8') ext_doc_chksum = p[4].decode(encoding='utf-8') else: doc_ref_id = p[2] doc_uri = p[3] ext_doc_chksum = p[4] self.builder.add_ext_doc_refs(self.document, doc_ref_id, doc_uri, ext_doc_chksum) except SPDXValueError: self.error = True msg = ERROR_MESSAGES['EXT_DOC_REF_VALUE'].format(p.lineno(2)) self.logger.log(msg)
[ "ext_doc_ref", ":", "EXT_DOC_REF", "DOC_REF_ID", "DOC_URI", "EXT_DOC_REF_CHKSUM" ]
spdx/tools-python
python
https://github.com/spdx/tools-python/blob/301d72f6ae57c832c1da7f6402fa49b192de6810/spdx/parsers/tagvalue.py#L1268-L1285
[ "def", "p_ext_doc_refs_1", "(", "self", ",", "p", ")", ":", "try", ":", "if", "six", ".", "PY2", ":", "doc_ref_id", "=", "p", "[", "2", "]", ".", "decode", "(", "encoding", "=", "'utf-8'", ")", "doc_uri", "=", "p", "[", "3", "]", ".", "decode", ...
301d72f6ae57c832c1da7f6402fa49b192de6810
valid
Parser.p_spdx_version_1
spdx_version : DOC_VERSION LINE
spdx/parsers/tagvalue.py
def p_spdx_version_1(self, p): """spdx_version : DOC_VERSION LINE""" try: if six.PY2: value = p[2].decode(encoding='utf-8') else: value = p[2] self.builder.set_doc_version(self.document, value) except CardinalityError: self.more_than_one_error('SPDXVersion', p.lineno(1)) except SPDXValueError: self.error = True msg = ERROR_MESSAGES['DOC_VERSION_VALUE'].format(p[2], p.lineno(1)) self.logger.log(msg)
def p_spdx_version_1(self, p): """spdx_version : DOC_VERSION LINE""" try: if six.PY2: value = p[2].decode(encoding='utf-8') else: value = p[2] self.builder.set_doc_version(self.document, value) except CardinalityError: self.more_than_one_error('SPDXVersion', p.lineno(1)) except SPDXValueError: self.error = True msg = ERROR_MESSAGES['DOC_VERSION_VALUE'].format(p[2], p.lineno(1)) self.logger.log(msg)
[ "spdx_version", ":", "DOC_VERSION", "LINE" ]
spdx/tools-python
python
https://github.com/spdx/tools-python/blob/301d72f6ae57c832c1da7f6402fa49b192de6810/spdx/parsers/tagvalue.py#L1293-L1306
[ "def", "p_spdx_version_1", "(", "self", ",", "p", ")", ":", "try", ":", "if", "six", ".", "PY2", ":", "value", "=", "p", "[", "2", "]", ".", "decode", "(", "encoding", "=", "'utf-8'", ")", "else", ":", "value", "=", "p", "[", "2", "]", "self", ...
301d72f6ae57c832c1da7f6402fa49b192de6810
valid
Parser.p_creator_comment_1
creator_comment : CREATOR_COMMENT TEXT
spdx/parsers/tagvalue.py
def p_creator_comment_1(self, p): """creator_comment : CREATOR_COMMENT TEXT""" try: if six.PY2: value = p[2].decode(encoding='utf-8') else: value = p[2] self.builder.set_creation_comment(self.document, value) except CardinalityError: self.more_than_one_error('CreatorComment', p.lineno(1))
def p_creator_comment_1(self, p): """creator_comment : CREATOR_COMMENT TEXT""" try: if six.PY2: value = p[2].decode(encoding='utf-8') else: value = p[2] self.builder.set_creation_comment(self.document, value) except CardinalityError: self.more_than_one_error('CreatorComment', p.lineno(1))
[ "creator_comment", ":", "CREATOR_COMMENT", "TEXT" ]
spdx/tools-python
python
https://github.com/spdx/tools-python/blob/301d72f6ae57c832c1da7f6402fa49b192de6810/spdx/parsers/tagvalue.py#L1314-L1323
[ "def", "p_creator_comment_1", "(", "self", ",", "p", ")", ":", "try", ":", "if", "six", ".", "PY2", ":", "value", "=", "p", "[", "2", "]", ".", "decode", "(", "encoding", "=", "'utf-8'", ")", "else", ":", "value", "=", "p", "[", "2", "]", "self...
301d72f6ae57c832c1da7f6402fa49b192de6810
valid
Parser.p_created_1
created : CREATED DATE
spdx/parsers/tagvalue.py
def p_created_1(self, p): """created : CREATED DATE""" try: if six.PY2: value = p[2].decode(encoding='utf-8') else: value = p[2] self.builder.set_created_date(self.document, value) except CardinalityError: self.more_than_one_error('Created', p.lineno(1))
def p_created_1(self, p): """created : CREATED DATE""" try: if six.PY2: value = p[2].decode(encoding='utf-8') else: value = p[2] self.builder.set_created_date(self.document, value) except CardinalityError: self.more_than_one_error('Created', p.lineno(1))
[ "created", ":", "CREATED", "DATE" ]
spdx/tools-python
python
https://github.com/spdx/tools-python/blob/301d72f6ae57c832c1da7f6402fa49b192de6810/spdx/parsers/tagvalue.py#L1341-L1350
[ "def", "p_created_1", "(", "self", ",", "p", ")", ":", "try", ":", "if", "six", ".", "PY2", ":", "value", "=", "p", "[", "2", "]", ".", "decode", "(", "encoding", "=", "'utf-8'", ")", "else", ":", "value", "=", "p", "[", "2", "]", "self", "."...
301d72f6ae57c832c1da7f6402fa49b192de6810
valid
write_creation_info
Write the creation info to out.
spdx/writers/tagvalue.py
def write_creation_info(creation_info, out): """ Write the creation info to out. """ out.write('# Creation Info\n\n') # Write sorted creators for creator in sorted(creation_info.creators): write_value('Creator', creator, out) # write created write_value('Created', creation_info.created_iso_format, out) # possible comment if creation_info.has_comment: write_text_value('CreatorComment', creation_info.comment, out)
def write_creation_info(creation_info, out): """ Write the creation info to out. """ out.write('# Creation Info\n\n') # Write sorted creators for creator in sorted(creation_info.creators): write_value('Creator', creator, out) # write created write_value('Created', creation_info.created_iso_format, out) # possible comment if creation_info.has_comment: write_text_value('CreatorComment', creation_info.comment, out)
[ "Write", "the", "creation", "info", "to", "out", "." ]
spdx/tools-python
python
https://github.com/spdx/tools-python/blob/301d72f6ae57c832c1da7f6402fa49b192de6810/spdx/writers/tagvalue.py#L51-L64
[ "def", "write_creation_info", "(", "creation_info", ",", "out", ")", ":", "out", ".", "write", "(", "'# Creation Info\\n\\n'", ")", "# Write sorted creators", "for", "creator", "in", "sorted", "(", "creation_info", ".", "creators", ")", ":", "write_value", "(", ...
301d72f6ae57c832c1da7f6402fa49b192de6810
valid
write_review
Write the fields of a single review to out.
spdx/writers/tagvalue.py
def write_review(review, out): """ Write the fields of a single review to out. """ out.write('# Review\n\n') write_value('Reviewer', review.reviewer, out) write_value('ReviewDate', review.review_date_iso_format, out) if review.has_comment: write_text_value('ReviewComment', review.comment, out)
def write_review(review, out): """ Write the fields of a single review to out. """ out.write('# Review\n\n') write_value('Reviewer', review.reviewer, out) write_value('ReviewDate', review.review_date_iso_format, out) if review.has_comment: write_text_value('ReviewComment', review.comment, out)
[ "Write", "the", "fields", "of", "a", "single", "review", "to", "out", "." ]
spdx/tools-python
python
https://github.com/spdx/tools-python/blob/301d72f6ae57c832c1da7f6402fa49b192de6810/spdx/writers/tagvalue.py#L67-L75
[ "def", "write_review", "(", "review", ",", "out", ")", ":", "out", ".", "write", "(", "'# Review\\n\\n'", ")", "write_value", "(", "'Reviewer'", ",", "review", ".", "reviewer", ",", "out", ")", "write_value", "(", "'ReviewDate'", ",", "review", ".", "revie...
301d72f6ae57c832c1da7f6402fa49b192de6810
valid
write_annotation
Write the fields of a single annotation to out.
spdx/writers/tagvalue.py
def write_annotation(annotation, out): """ Write the fields of a single annotation to out. """ out.write('# Annotation\n\n') write_value('Annotator', annotation.annotator, out) write_value('AnnotationDate', annotation.annotation_date_iso_format, out) if annotation.has_comment: write_text_value('AnnotationComment', annotation.comment, out) write_value('AnnotationType', annotation.annotation_type, out) write_value('SPDXREF', annotation.spdx_id, out)
def write_annotation(annotation, out): """ Write the fields of a single annotation to out. """ out.write('# Annotation\n\n') write_value('Annotator', annotation.annotator, out) write_value('AnnotationDate', annotation.annotation_date_iso_format, out) if annotation.has_comment: write_text_value('AnnotationComment', annotation.comment, out) write_value('AnnotationType', annotation.annotation_type, out) write_value('SPDXREF', annotation.spdx_id, out)
[ "Write", "the", "fields", "of", "a", "single", "annotation", "to", "out", "." ]
spdx/tools-python
python
https://github.com/spdx/tools-python/blob/301d72f6ae57c832c1da7f6402fa49b192de6810/spdx/writers/tagvalue.py#L78-L88
[ "def", "write_annotation", "(", "annotation", ",", "out", ")", ":", "out", ".", "write", "(", "'# Annotation\\n\\n'", ")", "write_value", "(", "'Annotator'", ",", "annotation", ".", "annotator", ",", "out", ")", "write_value", "(", "'AnnotationDate'", ",", "an...
301d72f6ae57c832c1da7f6402fa49b192de6810
valid
write_file
Write a file fields to out.
spdx/writers/tagvalue.py
def write_file(spdx_file, out): """ Write a file fields to out. """ out.write('# File\n\n') write_value('FileName', spdx_file.name, out) write_value('SPDXID', spdx_file.spdx_id, out) if spdx_file.has_optional_field('type'): write_file_type(spdx_file.type, out) write_value('FileChecksum', spdx_file.chk_sum.to_tv(), out) if isinstance(spdx_file.conc_lics, (document.LicenseConjunction, document.LicenseDisjunction)): write_value('LicenseConcluded', u'({0})'.format(spdx_file.conc_lics), out) else: write_value('LicenseConcluded', spdx_file.conc_lics, out) # write sorted list for lics in sorted(spdx_file.licenses_in_file): write_value('LicenseInfoInFile', lics, out) if isinstance(spdx_file.copyright, six.string_types): write_text_value('FileCopyrightText', spdx_file.copyright, out) else: write_value('FileCopyrightText', spdx_file.copyright, out) if spdx_file.has_optional_field('license_comment'): write_text_value('LicenseComments', spdx_file.license_comment, out) if spdx_file.has_optional_field('comment'): write_text_value('FileComment', spdx_file.comment, out) if spdx_file.has_optional_field('notice'): write_text_value('FileNotice', spdx_file.notice, out) for contributor in sorted(spdx_file.contributors): write_value('FileContributor', contributor, out) for dependency in sorted(spdx_file.dependencies): write_value('FileDependency', dependency, out) names = spdx_file.artifact_of_project_name homepages = spdx_file.artifact_of_project_home uris = spdx_file.artifact_of_project_uri for name, homepage, uri in sorted(zip_longest(names, homepages, uris)): write_value('ArtifactOfProjectName', name, out) if homepage is not None: write_value('ArtifactOfProjectHomePage', homepage, out) if uri is not None: write_value('ArtifactOfProjectURI', uri, out)
def write_file(spdx_file, out): """ Write a file fields to out. """ out.write('# File\n\n') write_value('FileName', spdx_file.name, out) write_value('SPDXID', spdx_file.spdx_id, out) if spdx_file.has_optional_field('type'): write_file_type(spdx_file.type, out) write_value('FileChecksum', spdx_file.chk_sum.to_tv(), out) if isinstance(spdx_file.conc_lics, (document.LicenseConjunction, document.LicenseDisjunction)): write_value('LicenseConcluded', u'({0})'.format(spdx_file.conc_lics), out) else: write_value('LicenseConcluded', spdx_file.conc_lics, out) # write sorted list for lics in sorted(spdx_file.licenses_in_file): write_value('LicenseInfoInFile', lics, out) if isinstance(spdx_file.copyright, six.string_types): write_text_value('FileCopyrightText', spdx_file.copyright, out) else: write_value('FileCopyrightText', spdx_file.copyright, out) if spdx_file.has_optional_field('license_comment'): write_text_value('LicenseComments', spdx_file.license_comment, out) if spdx_file.has_optional_field('comment'): write_text_value('FileComment', spdx_file.comment, out) if spdx_file.has_optional_field('notice'): write_text_value('FileNotice', spdx_file.notice, out) for contributor in sorted(spdx_file.contributors): write_value('FileContributor', contributor, out) for dependency in sorted(spdx_file.dependencies): write_value('FileDependency', dependency, out) names = spdx_file.artifact_of_project_name homepages = spdx_file.artifact_of_project_home uris = spdx_file.artifact_of_project_uri for name, homepage, uri in sorted(zip_longest(names, homepages, uris)): write_value('ArtifactOfProjectName', name, out) if homepage is not None: write_value('ArtifactOfProjectHomePage', homepage, out) if uri is not None: write_value('ArtifactOfProjectURI', uri, out)
[ "Write", "a", "file", "fields", "to", "out", "." ]
spdx/tools-python
python
https://github.com/spdx/tools-python/blob/301d72f6ae57c832c1da7f6402fa49b192de6810/spdx/writers/tagvalue.py#L101-L149
[ "def", "write_file", "(", "spdx_file", ",", "out", ")", ":", "out", ".", "write", "(", "'# File\\n\\n'", ")", "write_value", "(", "'FileName'", ",", "spdx_file", ".", "name", ",", "out", ")", "write_value", "(", "'SPDXID'", ",", "spdx_file", ".", "spdx_id"...
301d72f6ae57c832c1da7f6402fa49b192de6810
valid
write_package
Write a package fields to out.
spdx/writers/tagvalue.py
def write_package(package, out): """ Write a package fields to out. """ out.write('# Package\n\n') write_value('PackageName', package.name, out) if package.has_optional_field('version'): write_value('PackageVersion', package.version, out) write_value('PackageDownloadLocation', package.download_location, out) if package.has_optional_field('summary'): write_text_value('PackageSummary', package.summary, out) if package.has_optional_field('source_info'): write_text_value('PackageSourceInfo', package.source_info, out) if package.has_optional_field('file_name'): write_value('PackageFileName', package.file_name, out) if package.has_optional_field('supplier'): write_value('PackageSupplier', package.supplier, out) if package.has_optional_field('originator'): write_value('PackageOriginator', package.originator, out) if package.has_optional_field('check_sum'): write_value('PackageChecksum', package.check_sum.to_tv(), out) write_value('PackageVerificationCode', format_verif_code(package), out) if package.has_optional_field('description'): write_text_value('PackageDescription', package.description, out) if isinstance(package.license_declared, (document.LicenseConjunction, document.LicenseDisjunction)): write_value('PackageLicenseDeclared', u'({0})'.format(package.license_declared), out) else: write_value('PackageLicenseDeclared', package.license_declared, out) if isinstance(package.conc_lics, (document.LicenseConjunction, document.LicenseDisjunction)): write_value('PackageLicenseConcluded', u'({0})'.format(package.conc_lics), out) else: write_value('PackageLicenseConcluded', package.conc_lics, out) # Write sorted list of licenses. for lics in sorted(package.licenses_from_files): write_value('PackageLicenseInfoFromFiles', lics, out) if package.has_optional_field('license_comment'): write_text_value('PackageLicenseComments', package.license_comment, out) # cr_text is either free form text or NONE or NOASSERTION. if isinstance(package.cr_text, six.string_types): write_text_value('PackageCopyrightText', package.cr_text, out) else: write_value('PackageCopyrightText', package.cr_text, out) if package.has_optional_field('homepage'): write_value('PackageHomePage', package.homepage, out) # Write sorted files. for spdx_file in sorted(package.files): write_separators(out) write_file(spdx_file, out)
def write_package(package, out): """ Write a package fields to out. """ out.write('# Package\n\n') write_value('PackageName', package.name, out) if package.has_optional_field('version'): write_value('PackageVersion', package.version, out) write_value('PackageDownloadLocation', package.download_location, out) if package.has_optional_field('summary'): write_text_value('PackageSummary', package.summary, out) if package.has_optional_field('source_info'): write_text_value('PackageSourceInfo', package.source_info, out) if package.has_optional_field('file_name'): write_value('PackageFileName', package.file_name, out) if package.has_optional_field('supplier'): write_value('PackageSupplier', package.supplier, out) if package.has_optional_field('originator'): write_value('PackageOriginator', package.originator, out) if package.has_optional_field('check_sum'): write_value('PackageChecksum', package.check_sum.to_tv(), out) write_value('PackageVerificationCode', format_verif_code(package), out) if package.has_optional_field('description'): write_text_value('PackageDescription', package.description, out) if isinstance(package.license_declared, (document.LicenseConjunction, document.LicenseDisjunction)): write_value('PackageLicenseDeclared', u'({0})'.format(package.license_declared), out) else: write_value('PackageLicenseDeclared', package.license_declared, out) if isinstance(package.conc_lics, (document.LicenseConjunction, document.LicenseDisjunction)): write_value('PackageLicenseConcluded', u'({0})'.format(package.conc_lics), out) else: write_value('PackageLicenseConcluded', package.conc_lics, out) # Write sorted list of licenses. for lics in sorted(package.licenses_from_files): write_value('PackageLicenseInfoFromFiles', lics, out) if package.has_optional_field('license_comment'): write_text_value('PackageLicenseComments', package.license_comment, out) # cr_text is either free form text or NONE or NOASSERTION. if isinstance(package.cr_text, six.string_types): write_text_value('PackageCopyrightText', package.cr_text, out) else: write_value('PackageCopyrightText', package.cr_text, out) if package.has_optional_field('homepage'): write_value('PackageHomePage', package.homepage, out) # Write sorted files. for spdx_file in sorted(package.files): write_separators(out) write_file(spdx_file, out)
[ "Write", "a", "package", "fields", "to", "out", "." ]
spdx/tools-python
python
https://github.com/spdx/tools-python/blob/301d72f6ae57c832c1da7f6402fa49b192de6810/spdx/writers/tagvalue.py#L152-L216
[ "def", "write_package", "(", "package", ",", "out", ")", ":", "out", ".", "write", "(", "'# Package\\n\\n'", ")", "write_value", "(", "'PackageName'", ",", "package", ".", "name", ",", "out", ")", "if", "package", ".", "has_optional_field", "(", "'version'",...
301d72f6ae57c832c1da7f6402fa49b192de6810
valid
write_extracted_licenses
Write extracted licenses fields to out.
spdx/writers/tagvalue.py
def write_extracted_licenses(lics, out): """ Write extracted licenses fields to out. """ write_value('LicenseID', lics.identifier, out) if lics.full_name is not None: write_value('LicenseName', lics.full_name, out) if lics.comment is not None: write_text_value('LicenseComment', lics.comment, out) for xref in sorted(lics.cross_ref): write_value('LicenseCrossReference', xref, out) write_text_value('ExtractedText', lics.text, out)
def write_extracted_licenses(lics, out): """ Write extracted licenses fields to out. """ write_value('LicenseID', lics.identifier, out) if lics.full_name is not None: write_value('LicenseName', lics.full_name, out) if lics.comment is not None: write_text_value('LicenseComment', lics.comment, out) for xref in sorted(lics.cross_ref): write_value('LicenseCrossReference', xref, out) write_text_value('ExtractedText', lics.text, out)
[ "Write", "extracted", "licenses", "fields", "to", "out", "." ]
spdx/tools-python
python
https://github.com/spdx/tools-python/blob/301d72f6ae57c832c1da7f6402fa49b192de6810/spdx/writers/tagvalue.py#L219-L234
[ "def", "write_extracted_licenses", "(", "lics", ",", "out", ")", ":", "write_value", "(", "'LicenseID'", ",", "lics", ".", "identifier", ",", "out", ")", "if", "lics", ".", "full_name", "is", "not", "None", ":", "write_value", "(", "'LicenseName'", ",", "l...
301d72f6ae57c832c1da7f6402fa49b192de6810
valid
write_document
Write an SPDX tag value document. - document - spdx.document instance. - out - file like object that will be written to. Optionally `validate` the document before writing and raise InvalidDocumentError if document.validate returns False.
spdx/writers/tagvalue.py
def write_document(document, out, validate=True): """ Write an SPDX tag value document. - document - spdx.document instance. - out - file like object that will be written to. Optionally `validate` the document before writing and raise InvalidDocumentError if document.validate returns False. """ messages = [] messages = document.validate(messages) if validate and messages: raise InvalidDocumentError(messages) # Write out document information out.write('# Document Information\n\n') write_value('SPDXVersion', str(document.version), out) write_value('DataLicense', document.data_license.identifier, out) write_value('DocumentName', document.name, out) write_value('SPDXID', 'SPDXRef-DOCUMENT', out) write_value('DocumentNamespace', document.namespace, out) if document.has_comment: write_text_value('DocumentComment', document.comment, out) for doc_ref in document.ext_document_references: doc_ref_str = ' '.join([doc_ref.external_document_id, doc_ref.spdx_document_uri, doc_ref.check_sum.identifier + ':' + doc_ref.check_sum.value]) write_value('ExternalDocumentRef', doc_ref_str, out) write_separators(out) # Write out creation info write_creation_info(document.creation_info, out) write_separators(out) # Writesorted reviews for review in sorted(document.reviews): write_review(review, out) write_separators(out) #Write sorted annotations for annotation in sorted(document.annotations): write_annotation(annotation, out) write_separators(out) # Write out package info write_package(document.package, out) write_separators(out) out.write('# Extracted Licenses\n\n') for lic in sorted(document.extracted_licenses): write_extracted_licenses(lic, out) write_separators(out)
def write_document(document, out, validate=True): """ Write an SPDX tag value document. - document - spdx.document instance. - out - file like object that will be written to. Optionally `validate` the document before writing and raise InvalidDocumentError if document.validate returns False. """ messages = [] messages = document.validate(messages) if validate and messages: raise InvalidDocumentError(messages) # Write out document information out.write('# Document Information\n\n') write_value('SPDXVersion', str(document.version), out) write_value('DataLicense', document.data_license.identifier, out) write_value('DocumentName', document.name, out) write_value('SPDXID', 'SPDXRef-DOCUMENT', out) write_value('DocumentNamespace', document.namespace, out) if document.has_comment: write_text_value('DocumentComment', document.comment, out) for doc_ref in document.ext_document_references: doc_ref_str = ' '.join([doc_ref.external_document_id, doc_ref.spdx_document_uri, doc_ref.check_sum.identifier + ':' + doc_ref.check_sum.value]) write_value('ExternalDocumentRef', doc_ref_str, out) write_separators(out) # Write out creation info write_creation_info(document.creation_info, out) write_separators(out) # Writesorted reviews for review in sorted(document.reviews): write_review(review, out) write_separators(out) #Write sorted annotations for annotation in sorted(document.annotations): write_annotation(annotation, out) write_separators(out) # Write out package info write_package(document.package, out) write_separators(out) out.write('# Extracted Licenses\n\n') for lic in sorted(document.extracted_licenses): write_extracted_licenses(lic, out) write_separators(out)
[ "Write", "an", "SPDX", "tag", "value", "document", ".", "-", "document", "-", "spdx", ".", "document", "instance", ".", "-", "out", "-", "file", "like", "object", "that", "will", "be", "written", "to", ".", "Optionally", "validate", "the", "document", "b...
spdx/tools-python
python
https://github.com/spdx/tools-python/blob/301d72f6ae57c832c1da7f6402fa49b192de6810/spdx/writers/tagvalue.py#L237-L287
[ "def", "write_document", "(", "document", ",", "out", ",", "validate", "=", "True", ")", ":", "messages", "=", "[", "]", "messages", "=", "document", ".", "validate", "(", "messages", ")", "if", "validate", "and", "messages", ":", "raise", "InvalidDocument...
301d72f6ae57c832c1da7f6402fa49b192de6810
valid
checksum_from_sha1
Return an spdx.checksum.Algorithm instance representing the SHA1 checksum or None if does not match CHECKSUM_RE.
spdx/parsers/tagvaluebuilders.py
def checksum_from_sha1(value): """ Return an spdx.checksum.Algorithm instance representing the SHA1 checksum or None if does not match CHECKSUM_RE. """ # More constrained regex at lexer level CHECKSUM_RE = re.compile('SHA1:\s*([\S]+)', re.UNICODE) match = CHECKSUM_RE.match(value) if match: return checksum.Algorithm(identifier='SHA1', value=match.group(1)) else: return None
def checksum_from_sha1(value): """ Return an spdx.checksum.Algorithm instance representing the SHA1 checksum or None if does not match CHECKSUM_RE. """ # More constrained regex at lexer level CHECKSUM_RE = re.compile('SHA1:\s*([\S]+)', re.UNICODE) match = CHECKSUM_RE.match(value) if match: return checksum.Algorithm(identifier='SHA1', value=match.group(1)) else: return None
[ "Return", "an", "spdx", ".", "checksum", ".", "Algorithm", "instance", "representing", "the", "SHA1", "checksum", "or", "None", "if", "does", "not", "match", "CHECKSUM_RE", "." ]
spdx/tools-python
python
https://github.com/spdx/tools-python/blob/301d72f6ae57c832c1da7f6402fa49b192de6810/spdx/parsers/tagvaluebuilders.py#L38-L49
[ "def", "checksum_from_sha1", "(", "value", ")", ":", "# More constrained regex at lexer level", "CHECKSUM_RE", "=", "re", ".", "compile", "(", "'SHA1:\\s*([\\S]+)'", ",", "re", ".", "UNICODE", ")", "match", "=", "CHECKSUM_RE", ".", "match", "(", "value", ")", "i...
301d72f6ae57c832c1da7f6402fa49b192de6810
valid
str_from_text
Return content of a free form text block as a string.
spdx/parsers/tagvaluebuilders.py
def str_from_text(text): """ Return content of a free form text block as a string. """ REGEX = re.compile('<text>((.|\n)+)</text>', re.UNICODE) match = REGEX.match(text) if match: return match.group(1) else: return None
def str_from_text(text): """ Return content of a free form text block as a string. """ REGEX = re.compile('<text>((.|\n)+)</text>', re.UNICODE) match = REGEX.match(text) if match: return match.group(1) else: return None
[ "Return", "content", "of", "a", "free", "form", "text", "block", "as", "a", "string", "." ]
spdx/tools-python
python
https://github.com/spdx/tools-python/blob/301d72f6ae57c832c1da7f6402fa49b192de6810/spdx/parsers/tagvaluebuilders.py#L52-L61
[ "def", "str_from_text", "(", "text", ")", ":", "REGEX", "=", "re", ".", "compile", "(", "'<text>((.|\\n)+)</text>'", ",", "re", ".", "UNICODE", ")", "match", "=", "REGEX", ".", "match", "(", "text", ")", "if", "match", ":", "return", "match", ".", "gro...
301d72f6ae57c832c1da7f6402fa49b192de6810
valid
DocBuilder.set_doc_version
Set the document version. Raise SPDXValueError if malformed value, CardinalityError if already defined
spdx/parsers/tagvaluebuilders.py
def set_doc_version(self, doc, value): """ Set the document version. Raise SPDXValueError if malformed value, CardinalityError if already defined """ if not self.doc_version_set: self.doc_version_set = True m = self.VERS_STR_REGEX.match(value) if m is None: raise SPDXValueError('Document::Version') else: doc.version = version.Version(major=int(m.group(1)), minor=int(m.group(2))) return True else: raise CardinalityError('Document::Version')
def set_doc_version(self, doc, value): """ Set the document version. Raise SPDXValueError if malformed value, CardinalityError if already defined """ if not self.doc_version_set: self.doc_version_set = True m = self.VERS_STR_REGEX.match(value) if m is None: raise SPDXValueError('Document::Version') else: doc.version = version.Version(major=int(m.group(1)), minor=int(m.group(2))) return True else: raise CardinalityError('Document::Version')
[ "Set", "the", "document", "version", ".", "Raise", "SPDXValueError", "if", "malformed", "value", "CardinalityError", "if", "already", "defined" ]
spdx/tools-python
python
https://github.com/spdx/tools-python/blob/301d72f6ae57c832c1da7f6402fa49b192de6810/spdx/parsers/tagvaluebuilders.py#L74-L90
[ "def", "set_doc_version", "(", "self", ",", "doc", ",", "value", ")", ":", "if", "not", "self", ".", "doc_version_set", ":", "self", ".", "doc_version_set", "=", "True", "m", "=", "self", ".", "VERS_STR_REGEX", ".", "match", "(", "value", ")", "if", "m...
301d72f6ae57c832c1da7f6402fa49b192de6810
valid
DocBuilder.set_doc_data_lics
Sets the document data license. Raises value error if malformed value, CardinalityError if already defined.
spdx/parsers/tagvaluebuilders.py
def set_doc_data_lics(self, doc, lics): """Sets the document data license. Raises value error if malformed value, CardinalityError if already defined. """ if not self.doc_data_lics_set: self.doc_data_lics_set = True if validations.validate_data_lics(lics): doc.data_license = document.License.from_identifier(lics) return True else: raise SPDXValueError('Document::DataLicense') else: raise CardinalityError('Document::DataLicense')
def set_doc_data_lics(self, doc, lics): """Sets the document data license. Raises value error if malformed value, CardinalityError if already defined. """ if not self.doc_data_lics_set: self.doc_data_lics_set = True if validations.validate_data_lics(lics): doc.data_license = document.License.from_identifier(lics) return True else: raise SPDXValueError('Document::DataLicense') else: raise CardinalityError('Document::DataLicense')
[ "Sets", "the", "document", "data", "license", ".", "Raises", "value", "error", "if", "malformed", "value", "CardinalityError", "if", "already", "defined", "." ]
spdx/tools-python
python
https://github.com/spdx/tools-python/blob/301d72f6ae57c832c1da7f6402fa49b192de6810/spdx/parsers/tagvaluebuilders.py#L92-L105
[ "def", "set_doc_data_lics", "(", "self", ",", "doc", ",", "lics", ")", ":", "if", "not", "self", ".", "doc_data_lics_set", ":", "self", ".", "doc_data_lics_set", "=", "True", "if", "validations", ".", "validate_data_lics", "(", "lics", ")", ":", "doc", "."...
301d72f6ae57c832c1da7f6402fa49b192de6810
valid
DocBuilder.set_doc_name
Sets the document name. Raises CardinalityError if already defined.
spdx/parsers/tagvaluebuilders.py
def set_doc_name(self, doc, name): """Sets the document name. Raises CardinalityError if already defined. """ if not self.doc_name_set: doc.name = name self.doc_name_set = True return True else: raise CardinalityError('Document::Name')
def set_doc_name(self, doc, name): """Sets the document name. Raises CardinalityError if already defined. """ if not self.doc_name_set: doc.name = name self.doc_name_set = True return True else: raise CardinalityError('Document::Name')
[ "Sets", "the", "document", "name", ".", "Raises", "CardinalityError", "if", "already", "defined", "." ]
spdx/tools-python
python
https://github.com/spdx/tools-python/blob/301d72f6ae57c832c1da7f6402fa49b192de6810/spdx/parsers/tagvaluebuilders.py#L107-L116
[ "def", "set_doc_name", "(", "self", ",", "doc", ",", "name", ")", ":", "if", "not", "self", ".", "doc_name_set", ":", "doc", ".", "name", "=", "name", "self", ".", "doc_name_set", "=", "True", "return", "True", "else", ":", "raise", "CardinalityError", ...
301d72f6ae57c832c1da7f6402fa49b192de6810