blob_id
stringlengths
40
40
directory_id
stringlengths
40
40
path
stringlengths
2
616
content_id
stringlengths
40
40
detected_licenses
listlengths
0
69
license_type
stringclasses
2 values
repo_name
stringlengths
5
118
snapshot_id
stringlengths
40
40
revision_id
stringlengths
40
40
branch_name
stringlengths
4
63
visit_date
timestamp[us]
revision_date
timestamp[us]
committer_date
timestamp[us]
github_id
int64
2.91k
686M
star_events_count
int64
0
209k
fork_events_count
int64
0
110k
gha_license_id
stringclasses
23 values
gha_event_created_at
timestamp[us]
gha_created_at
timestamp[us]
gha_language
stringclasses
213 values
src_encoding
stringclasses
30 values
language
stringclasses
1 value
is_vendor
bool
2 classes
is_generated
bool
2 classes
length_bytes
int64
2
10.3M
extension
stringclasses
246 values
content
stringlengths
2
10.3M
authors
listlengths
1
1
author_id
stringlengths
0
212
53ff9198dfe47a214fb9a2f6bd7b7738da26a9cf
dfe78d0296c2f2497d323bc6d0516e3d03f6e5c3
/learning_templates/app1/urls.py
ec7eefe116f7b638f3fae7bd3beb5c4ebd257c51
[]
no_license
tekam47/django-deployement-trial
49ac22a6683d0156d694ec67432bdc8fa21e2514
c96b4bb8ef7104b6dbab2bd33b4fce65c627925a
refs/heads/master
2022-11-21T14:47:16.097166
2020-07-22T10:54:03
2020-07-22T10:54:03
281,641,203
0
0
null
null
null
null
UTF-8
Python
false
false
211
py
from app1 import views from django.conf.urls import url #TEMPLATE TAGGING app_name = "app1" urlpatterns = [ url(r'^other',views.other,name="other"), url(r'^relative',views.relative,name="relative"), ]
[ "kounatzepi@gmail.com" ]
kounatzepi@gmail.com
e514273c815a754c9cc94115339b8f0ab6c3c284
58e8117c418a8931d56fe57ba0edd38214c79642
/ate/context.py
ed59e79081a32350eef9099c194c3270b9fc91da
[ "MIT" ]
permissive
229051923/ApiTestEngine
90b495ee7be037a35f63ffa3b2c95f6ba24f43ce
4a529068f451c8880681286518ff6a643ecf9067
refs/heads/master
2021-01-01T16:09:10.737245
2017-07-19T15:29:29
2017-07-19T15:29:29
97,780,015
2
0
null
2017-07-20T02:09:36
2017-07-20T02:09:36
null
UTF-8
Python
false
false
7,621
py
import copy import importlib import re import types from collections import OrderedDict from ate import exception, testcase, utils def is_function(tup): """ Takes (name, object) tuple, returns True if it is a function. """ name, item = tup return isinstance(item, types.FunctionType) class Context(object): """ Manages context functions and variables. context has two levels, testset and testcase. """ def __init__(self): self.testset_config = {} self.testset_shared_variables_mapping = dict() self.testcase_config = {} self.testcase_variables_mapping = dict() self.init_context() def init_context(self, level='testset'): """ testset level context initializes when a file is loaded, testcase level context initializes when each testcase starts. """ if level == "testset": self.testset_config["functions"] = {} self.testset_config["variables"] = OrderedDict() self.testset_config["request"] = {} self.testset_shared_variables_mapping = {} self.testcase_config["functions"] = {} self.testcase_config["variables"] = OrderedDict() self.testcase_config["request"] = {} self.testcase_variables_mapping = copy.deepcopy(self.testset_shared_variables_mapping) def import_requires(self, modules): """ import required modules dynamicly """ for module_name in modules: globals()[module_name] = importlib.import_module(module_name) def bind_functions(self, function_binds, level="testcase"): """ Bind named functions within the context This allows for passing in self-defined functions in testing. e.g. function_binds: { "add_one": lambda x: x + 1, # lambda function "add_two_nums": "lambda x, y: x + y" # lambda function in string } """ eval_function_binds = {} for func_name, function in function_binds.items(): if isinstance(function, str): function = eval(function) eval_function_binds[func_name] = function self.__update_context_config(level, "functions", eval_function_binds) def import_module_functions(self, modules, level="testcase"): """ import modules and bind all functions within the context """ for module_name in modules: imported = importlib.import_module(module_name) imported_functions_dict = dict(filter(is_function, vars(imported).items())) self.__update_context_config(level, "functions", imported_functions_dict) def register_variables_config(self, variable_binds, level="testcase"): """ register variable configs @param (list) variable_binds, variable can be value or custom function e.g. [ {"TOKEN": "debugtalk"}, {"random": "${gen_random_string(5)}"}, {"json": {'name': 'user', 'password': '123456'}}, {"md5": "${gen_md5($TOKEN, $json, $random)}"} ] """ if level == "testset": for variable_bind in variable_binds: self.testset_config["variables"].update(variable_bind) elif level == "testcase": self.testcase_config["variables"] = copy.deepcopy(self.testset_config["variables"]) for variable_bind in variable_binds: self.testcase_config["variables"].update(variable_bind) def register_request(self, request_dict, level="testcase"): self.__update_context_config(level, "request", request_dict) def __update_context_config(self, level, config_type, config_mapping): """ @param level: testset or testcase @param config_type: functions, variables or request @param config_mapping: functions config mapping or variables config mapping """ if level == "testset": self.testset_config[config_type].update(config_mapping) elif level == "testcase": self.testcase_config[config_type].update(config_mapping) def get_parsed_request(self): """ get parsed request, with each variable replaced by bind value. testcase request shall inherit from testset request configs, but can not change testset configs, that's why we use copy.deepcopy here. """ testcase_request_config = utils.deep_update_dict( copy.deepcopy(self.testset_config["request"]), self.testcase_config["request"] ) parsed_request = testcase.parse_template( testcase_request_config, self._get_evaluated_testcase_variables() ) return parsed_request def bind_extracted_variables(self, variables_mapping): """ bind extracted variable to current testcase context and testset context. since extracted variable maybe used in current testcase and next testcases. """ self.testset_shared_variables_mapping.update(variables_mapping) self.testcase_variables_mapping.update(variables_mapping) def get_testcase_variables_mapping(self): return self.testcase_variables_mapping def _get_evaluated_testcase_variables(self): """ variables in variables_config will be evaluated each time """ testcase_functions_config = copy.deepcopy(self.testset_config["functions"]) testcase_functions_config.update(self.testcase_config["functions"]) self.testcase_config["functions"] = testcase_functions_config testcase_variables_config = copy.deepcopy(self.testset_config["variables"]) testcase_variables_config.update(self.testcase_config["variables"]) self.testcase_config["variables"] = testcase_variables_config for var_name, var_value in self.testcase_config["variables"].items(): self.testcase_variables_mapping[var_name] = self.get_eval_value(var_value) return self.testcase_variables_mapping def get_eval_value(self, data): """ evaluate data recursively, each variable in data will be evaluated. """ if isinstance(data, (list, tuple)): return [self.get_eval_value(item) for item in data] if isinstance(data, dict): evaluated_data = {} for key, value in data.items(): evaluated_data[key] = self.get_eval_value(value) return evaluated_data if isinstance(data, (int, float)): return data # data is in string format here data = "" if data is None else data.strip() if utils.is_variable(data): # variable marker: $var variable_name = utils.parse_variable(data) value = self.testcase_variables_mapping.get(variable_name) if value is None: raise exception.ParamsError( "%s is not defined in bind variables!" % variable_name) return value elif utils.is_functon(data): # function marker: ${func(1, 2, a=3, b=4)} fuction_meta = utils.parse_function(data) func_name = fuction_meta['func_name'] args = fuction_meta.get('args', []) kwargs = fuction_meta.get('kwargs', {}) args = self.get_eval_value(args) kwargs = self.get_eval_value(kwargs) return self.testcase_config["functions"][func_name](*args, **kwargs) else: return data
[ "mail@debugtalk.com" ]
mail@debugtalk.com
2272af86ec47659657698bd4b83b445ace287269
4b4ff2c0d135d3615caaeb80735c2ad6ee987914
/venv/bin/pip
f5f3cb0efa2f979b5fd036420c2b2dc08b10a062
[]
no_license
Nicolas-Turck/Tuto-deployement-heroku
23060837b47f195d9af2eb280a85836d1a8f8efd
54d104054c06070420ae36b6bbb45089492da286
refs/heads/master
2023-08-01T07:18:13.563988
2021-05-20T16:07:56
2021-05-20T16:07:56
257,563,781
0
0
null
2021-09-22T18:54:33
2020-04-21T10:46:23
Python
UTF-8
Python
false
false
263
#!/home/nicos/PycharmProjects/Tuto-heroku/venv/bin/python # -*- coding: utf-8 -*- import re import sys from pip._internal.cli.main import main if __name__ == '__main__': sys.argv[0] = re.sub(r'(-script\.pyw?|\.exe)?$', '', sys.argv[0]) sys.exit(main())
[ "nicolas.turck@gmail.com" ]
nicolas.turck@gmail.com
a1df8914c35f5e949416165a0782c85926e4e9f7
2cf9f165cb4d6e8e9009d74b43020fe2d5c1964f
/chat/migrations/0001_initial.py
a2db3bb2cfb40b358f72ef4113b611bc648018b0
[]
no_license
jimy1824/chat
29f5039c6284083b8328502932795bee586dec14
627ad4678c6215d37322737b38b3e5eb6d69696f
refs/heads/master
2023-04-27T15:10:51.316824
2019-11-08T05:00:12
2019-11-08T05:00:12
220,081,959
1
0
null
2023-04-21T20:40:47
2019-11-06T20:17:20
HTML
UTF-8
Python
false
false
1,194
py
# Generated by Django 2.2.6 on 2019-11-06 19:01 from django.conf import settings from django.db import migrations, models import django.db.models.deletion class Migration(migrations.Migration): initial = True dependencies = [ migrations.swappable_dependency(settings.AUTH_USER_MODEL), ] operations = [ migrations.CreateModel( name='Chat', fields=[ ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), ('is_active', models.BooleanField(default=True)), ('created_at', models.DateTimeField(auto_now_add=True)), ('updated_at', models.DateTimeField(auto_now=True)), ('message', models.TextField()), ('receiver', models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, related_name='message_receiver', to=settings.AUTH_USER_MODEL)), ('sender', models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, related_name='message_sender', to=settings.AUTH_USER_MODEL)), ], options={ 'abstract': False, }, ), ]
[ "jamshaid1824@gmail.com" ]
jamshaid1824@gmail.com
4357f72ea71e855c96174c243f3d51a84d1b6c8e
57a8986e7622e471f392ffefa1e1d959c6a1bb7c
/mazeSolver.py
1e522304e60f9ee925eee57e47874b7821dff42c
[]
no_license
pandyakaa/MazeProblem
394b5f99eef5ee61e0f4f206c29cc393ced2fa64
8ca7d6b5dd96697632326322ff1e3272d9d6fbea
refs/heads/master
2020-05-01T16:38:48.006105
2019-03-29T06:45:58
2019-03-29T06:45:58
177,577,738
0
0
null
null
null
null
UTF-8
Python
false
false
6,601
py
import sys from collections import deque from point import Point from priorityQueue import PriorityQueue # Fungsi inputMaze, dengan parameter sebuah file # digunakan untuk memasukkan matriks sebagai representasi dari maze # dari file eksternal dengan nama filename # Sekaligus mencari titik awal masuk dan keluar, disimpan dalam startb, startk, finishb dan finishk def inputMaze(filename) : arr = [] f = open("{}.txt".format(filename),"r") for line in f : arr.append([int(c) for c in line.strip()]) baris = len(arr) kolom = len(arr[0]) f.close() startb = -1 startk = -1 finishb = -1 finishk = -1 # Melakukan pencarian titik mulai dan akhir (case : kiri dan kanan) for i in range(baris) : if (arr[i][0] == 0) : startb = i startk = 0 if (arr[i][kolom-1] == 0) : finishb = i finishk = kolom-1 # Melakukan pencarian titik mulai dan akhir (case : atas dan bawah) for i in range(kolom) : if (arr[0][i] == 0) : startb = 0 startk = i if (arr[baris-1][i] == 0) : finishb = baris-1 finishk = i # Melakukan validasi apakah matriks bisa dimainkan atau tidak if ( startb != -1 and startk != -1 and finishb != -1 and finishk != -1 ) : valid = True else : valid = False return arr,startb,startk,finishb,finishk,valid # Fungsi printSolution, dengan parameter sebuah matriks m # digunakan untuk melakukan output sebuah maze yang sudah ada solved def printSolution(m) : for i in m : for j in i : if (j == 1 ) : print("# ",end = '') elif (j == 3 or j == 2) : print(" ",end = '') elif ( j == 4 ) : print("X ", end = '') else : print(" ", end = '') print() # Fungsi copy, dengan parameter sebuah matriks m1 # digunakan untuk melakukan DEEP COPY pada sebuah matriks # sehingga tidak perlu membaca dari file eksternal lagi def copy(m1) : m2 = [] for i in range(len(m1)) : temp = [] for j in range(len(m1[0])) : temp.append(m1[i][j]) m2.append(temp) return m2 # Fungsi isFeasible, dengan parameter sebuah matriks m, int x dan int y # digunakan untuk melakukan validasi, apakah koordinat (x,y) valid atau tidak # DEFINISI VALID : Lebih atau sama dengan 0 , dan kurang dari panjang atau kolom matriks def isFeasible(m,x,y) : if ( m[x][y]==0 and x >= 0 and x < len(m) and y >= 0 and y < len(m[0]) ) : return True return False # Fungsi BFS, dengan parameter maze maze, int x, int y, dan point fp # merupakan salah satu dari dua fungsi utama dalam program ini # Memanfaatkan sebuah type data DEQUE, dan melakukan proses Breadth-First Searching # Jika memiliki solusi, akan me-return sebuah point p def BFS(maze,x,y,fp) : de = deque() de.append(Point(x,y,None)) while ( not(len(de) == 0) ) : p = de.popleft() maze[p.x][p.y] = 3 if (p.isEqual(fp)) : return p if(isFeasible(maze,p.x-1,p.y)) : nextP = Point(p.x-1,p.y,p) de.append(nextP) if (isFeasible(maze,p.x+1,p.y)) : nextP = Point(p.x+1,p.y,p) de.append(nextP) if(isFeasible(maze,p.x,p.y+1)) : nextP = Point(p.x,p.y+1,p) de.append(nextP) if(isFeasible(maze,p.x,p.y-1)) : nextP = Point(p.x,p.y-1,p) de.append(nextP) # Fungsi manhattanDist, dengan parameter point point_start dan point point_finish # digunakan untuk mencari nilai h(n) pada algoritma A* # Menggunakan manhattan distance karena hanya bisa bergerak ke empat arah def manhattanDist(point_start,point_finish) : return (abs(point_start.x - point_finish.x) + abs(point_start.y - point_finish.y)) # Fungsi AStar, dengan parameter maze maze, int x, int y, dan point fpoint # merupakan salah satu dari dua fungsi utama dalam program ini # Memanfaatkan type data Priority Queue, yang telah dibuat kelas sendiri sebelumnya # Akan melakukan pencarian dengan algoritma AStar dengan : # f(n) = g(n) + h(n) # dengan g(n) adalah jarak sebenarnya sebuah titik ke titik akhir # dan h(n) adalah jarak heuristik dari sebuah titik ke titik akhir dengan memanfaatkan manhattanDist def AStar(maze,x,y,fpoint) : startPoint = Point(x,y,None) startPoint.f = startPoint.g = startPoint.h = 0 openList = PriorityQueue() openList.insert(startPoint) while ( not(openList.isEmpty()) ) : current_node = openList.delete() maze[current_node.x][current_node.y] = 3 if (current_node.isEqual(fpoint) ) : return current_node children = [] for pos in [(0, -1), (0, 1), (-1, 0), (1, 0)]: curr_x = current_node.x + pos[0] curr_y = current_node.y + pos[1] if (not(isFeasible(maze,curr_x,curr_y))) : continue child = Point(curr_x,curr_y,current_node) children.append(child) for child in children : child.g = current_node.g + 1 child.h = manhattanDist(child,fpoint) child.f = child.g + child.h openList.insert(child) # Fungsi main, akan dipanggil saat program ini dijalankan if __name__ == "__main__": # Melakukan input nama file dari pengguna, dan memanggil fungsi inputMaze untuk # memasukkannya ke dalam maze file = input("Masukkan nama file : ") maze, start_baris , start_kolom, finish_baris , finish_kolom , valid = inputMaze(file) maze2 = copy(maze) # Util yang diperlukan oleh fungsi-fungsi searching fp = Point(finish_baris,finish_kolom,None) if ( valid ) : # Melakukan pemanggilan fungsi-fungsi Searching p = BFS(maze,start_baris,start_kolom,fp) q = AStar(maze2,start_baris,start_kolom,fp) # Melakukan output dari algoritma DFS maze[start_baris][start_kolom] = 4 while (p.getParent() != None ) : maze[p.x][p.y] = 4 p = p.getParent() print("\n \t \t Solution with BFS : \n") printSolution(maze) # Melakukan output dari algoritma A* maze2[start_baris][start_kolom] = 4 while (q.getParent() != None ) : maze2[q.x][q.y] = 4 q = q.getParent() print("\n \t \t Solution with A-Star : \n") printSolution(maze2) else : print("NOT FOUND")
[ "pandyaka.aptanagi@gmail.com" ]
pandyaka.aptanagi@gmail.com
1b13c4cc42e2710f2aa2c196f9030cf2e7b9ae1d
7a2573f5539fe6de33fe4e355fde5df366cedb1f
/pl3.py
11eb70e73311b185b4cecea5fcc5e87fa406b57f
[]
no_license
anujasubbarao/anuja
508c2d5e777d854cdc1387560590f7c345f25134
c840cdcf99986f3e2d842ee24045178e72da6e9c
refs/heads/master
2021-07-06T01:52:17.420722
2019-03-06T14:26:53
2019-03-06T14:26:53
142,555,136
0
1
null
null
null
null
UTF-8
Python
false
false
101
py
n=int(raw_input()) rev=0 while n>0: rem=n%10 rev=(rev*10)+rem n=n//10 print rev
[ "noreply@github.com" ]
anujasubbarao.noreply@github.com
554c7aabf1e3a0c07792dec86ac77d8c958f7099
37faa935d3af283be72423821a25d17172f2df6a
/projects/analysing_baseball_data/isp_baseball_template.py
b091cf8da7961c40e7fd65bcc195823b729a3dae
[]
no_license
rogeriolaguilar/python
2b214fccbb32ca269c745b26fbc082cb20981c23
fef86ca360e5a47fcd86f9ab99c04c58d353705b
refs/heads/master
2020-03-19T11:00:51.367723
2018-09-16T12:56:39
2018-09-16T12:56:39
136,421,396
0
0
null
null
null
null
UTF-8
Python
false
false
11,560
py
""" Project for Week 4 of "Python Data Analysis". Processing CSV files with baseball stastics. Be sure to read the project description page for further information about the expected behavior of the program. """ import csv ## ## Provided code from Week 3 Project ## def read_csv_as_list_dict(filename, separator, quote): """ Inputs: filename - name of CSV file separator - character that separates fields quote - character used to optionally quote fields Output: Returns a list of dictionaries where each item in the list corresponds to a row in the CSV file. The dictionaries in the list map the field names to the field values for that row. """ table = [] with open(filename, newline='') as csvfile: csvreader = csv.DictReader(csvfile, delimiter=separator, quotechar=quote) for row in csvreader: table.append(row) return table def read_csv_as_nested_dict(filename, keyfield, separator, quote): """ Inputs: filename - name of CSV file keyfield - field to use as key for rows separator - character that separates fields quote - character used to optionally quote fields Output: Returns a dictionary of dictionaries where the outer dictionary maps the value in the key_field to the corresponding row in the CSV file. The inner dictionaries map the field names to the field values for that row. """ table = {} with open(filename, newline='') as csvfile: csvreader = csv.DictReader(csvfile, delimiter=separator, quotechar=quote) for row in csvreader: rowid = row[keyfield] table[rowid] = row return table ## ## Provided formulas for common batting statistics ## # Typical cutoff used for official statistics MINIMUM_AB = 500 def batting_average(info, batting_stats): """ Inputs: batting_stats - dictionary of batting statistics (values are strings) Output: Returns the batting average as a float """ hits = float(batting_stats[info["hits"]]) at_bats = float(batting_stats[info["atbats"]]) if at_bats >= MINIMUM_AB: return hits / at_bats else: return 0 def onbase_percentage(info, batting_stats): """ Inputs: batting_stats - dictionary of batting statistics (values are strings) Output: Returns the on-base percentage as a float """ hits = float(batting_stats[info["hits"]]) at_bats = float(batting_stats[info["atbats"]]) walks = float(batting_stats[info["walks"]]) if at_bats >= MINIMUM_AB: return (hits + walks) / (at_bats + walks) else: return 0 def slugging_percentage(info, batting_stats): """ Inputs: batting_stats - dictionary of batting statistics (values are strings) Output: Returns the slugging percentage as a float """ hits = float(batting_stats[info["hits"]]) doubles = float(batting_stats[info["doubles"]]) triples = float(batting_stats[info["triples"]]) home_runs = float(batting_stats[info["homeruns"]]) singles = hits - doubles - triples - home_runs at_bats = float(batting_stats[info["atbats"]]) if at_bats >= MINIMUM_AB: return (singles + 2 * doubles + 3 * triples + 4 * home_runs) / at_bats else: return 0 ## ## Part 1: Functions to compute top batting statistics by year ## def filter_by_year(statistics, year, yearid): """ Inputs: statistics - List of batting statistics dictionaries year - Year to filter by yearid - Year ID field in statistics Outputs: Returns a list of batting statistics dictionaries that are from the input year. """ return list(filter(lambda x: (x[yearid] == str(year)), statistics)) def top_player_ids(info, statistics, formula, numplayers): """ Inputs: info - Baseball data information dictionary statistics - List of batting statistics dictionaries formula - function that takes an info dictionary and a batting statistics dictionary as input and computes a compound statistic numplayers - Number of top players to return Outputs: Returns a list of tuples, player ID and compound statistic computed by formula, of the top numplayers players sorted in decreasing order of the computed statistic. """ table = [] for dic in statistics: table.append((dic[info['playerid']], formula(info, dic))) table = sorted(table, key=lambda x: x[1], reverse=True) return table[0:numplayers] def lookup_player_names(info, top_ids_and_stats): """ Inputs: info - Baseball data information dictionary top_ids_and_stats - list of tuples containing player IDs and computed statistics Outputs: List of strings of the form "x.xxx --- FirstName LastName", where "x.xxx" is a string conversion of the float stat in the input and "FirstName LastName" is the name of the player corresponding to the player ID in the input. """ stats = read_csv_as_list_dict(info['masterfile'], info['separator'], info['quote']) template = '{0:.3f} --- {1} {2}' table = [] for top in top_ids_and_stats: for dit in stats: if dit[info['playerid']] == top[0]: table.append(template.format(top[1], dit[info['firstname']], dit[info['lastname']])) return table def compute_top_stats_year(info, formula, numplayers, year): """ Inputs: info - Baseball data information dictionary formula - function that takes an info dictionary and a batting statistics dictionary as input and computes a compound statistic numplayers - Number of top players to return year - Year to filter by Outputs: Returns a list of strings for the top numplayers in the given year according to the given formula. """ b_stat = read_csv_as_list_dict(info['battingfile'], info['separator'], info['quote']) sep_year = filter_by_year(b_stat, year, info['yearid']) top_player = top_player_ids(info, sep_year, formula, numplayers) player_name = lookup_player_names(info, top_player) return player_name ## ## Part 2: Functions to compute top batting statistics by career ## def aggregate_by_player_id(statistics, playerid, fields): """ Inputs: statistics - List of batting statistics dictionaries playerid - Player ID field name fields - List of fields to aggregate Output: Returns a nested dictionary whose keys are player IDs and whose values are dictionaries of aggregated stats. Only the fields from the fields input will be aggregated in the aggregated stats dictionaries. """ dictt = {} for stat in statistics: dictt[stat[playerid]] = {} for field in fields: dictt[stat[playerid]][field] = 0 dictt[stat[playerid]][playerid] = stat[playerid] for key, _val in dictt.items(): for dit in statistics: if key == dit[playerid]: for stat in fields: dictt[key][stat] += int(dit[stat]) return dictt def compute_top_stats_career(info, formula, numplayers): """ Inputs: info - Baseball data information dictionary formula - function that takes an info dictionary and a batting statistics dictionary as input and computes a compound statistic numplayers - Number of top players to return """ bat_stat = read_csv_as_list_dict(info['battingfile'], info['separator'], info['quote']) fieldnames = info['battingfields'] indiv_player_stat = aggregate_by_player_id(bat_stat, info['playerid'], fieldnames) player_name = [k for k, v in indiv_player_stat.items()] compounded_stat = [] for player in player_name: compounded_stat.append((player, formula(info, indiv_player_stat[player]))) compounded_stat = sorted(compounded_stat, key=lambda x: x[1], reverse=True) return lookup_player_names(info, compounded_stat[0:numplayers]) ## ## Provided testing code ## def test_baseball_statistics(): """ Simple testing code. """ # # Dictionary containing information needed to access baseball statistics # This information is all tied to the format and contents of the CSV files # baseballdatainfo = {"masterfile": "Master_2016.csv", # Name of Master CSV file "battingfile": "Batting_2016.csv", # Name of Batting CSV file "separator": ",", # Separator character in CSV files "quote": '"', # Quote character in CSV files "playerid": "playerID", # Player ID field name "firstname": "nameFirst", # First name field name "lastname": "nameLast", # Last name field name "yearid": "yearID", # Year field name "atbats": "AB", # At bats field name "hits": "H", # Hits field name "doubles": "2B", # Doubles field name "triples": "3B", # Triples field name "homeruns": "HR", # Home runs field name "walks": "BB", # Walks field name "battingfields": ["AB", "H", "2B", "3B", "HR", "BB"]} print("Top 5 batting averages in 1923") top_batting_average_1923 = compute_top_stats_year(baseballdatainfo, batting_average, 5, 1923) for player in top_batting_average_1923: print(player) print("") print("Top 10 batting averages in 2010") top_batting_average_2010 = compute_top_stats_year(baseballdatainfo, batting_average, 10, 2010) for player in top_batting_average_2010: print(player) print("") print("Top 10 on-base percentage in 2010") top_onbase_2010 = compute_top_stats_year(baseballdatainfo, onbase_percentage, 10, 2010) for player in top_onbase_2010: print(player) print("") print("Top 10 slugging percentage in 2010") top_slugging_2010 = compute_top_stats_year(baseballdatainfo, slugging_percentage, 10, 2010) for player in top_slugging_2010: print(player) print("") # You can also use lambdas for the formula # This one computes onbase plus slugging percentage print("Top 10 OPS in 2010") top_ops_2010 = compute_top_stats_year(baseballdatainfo, lambda info, stats: (onbase_percentage(info, stats) + slugging_percentage(info, stats)), 10, 2010) for player in top_ops_2010: print(player) print("") print("Top 20 career batting averages") top_batting_average_career = compute_top_stats_career(baseballdatainfo, batting_average, 20) for player in top_batting_average_career: print(player) print("") # Make sure the following call to test_baseball_statistics is # commented out when submitting to OwlTest/CourseraTest. #test_baseball_statistics()
[ "rogerio.l.aguilar@gmail.com" ]
rogerio.l.aguilar@gmail.com
a6db9d0ebe8e9a8f0ab0a1cacff578441a2234ba
4fcfc6834f598954e069e9481e4f69d6f7205f3b
/Week1/day_3/Flask_Intro/first_flask_app/server.py
a7693151c6981a66648352d4cf8857f87de47de7
[]
no_license
sadieBoBadie/jan_2020_python_stack
dadc77a8a76fd4b900bc31ad83ed4680a0802aa1
b91da3da23ea57a27c086075f4b86c5bfec413d0
refs/heads/main
2023-03-02T22:27:46.877290
2021-01-26T19:04:57
2021-01-26T19:04:57
326,852,429
0
3
null
null
null
null
UTF-8
Python
false
false
543
py
from flask import Flask, render_template app = Flask(__name__) @app.route('/') def puppies(): return "<h1 style='color: red'>Puppies are cute!</h1>" @app.route('/<animal>/<color>') @app.route('/<animal>') @app.route('/<animal>/<color>/<int:num>') def display_animal(animal, color="blue", num=5): print(f"Animal: {animal}") print(f"Color: {color}") print("Type of the num var: ", type(num)) return render_template('index.html', animal=animal, color=color, num=num) if __name__=="__main__": app.run(debug=True)
[ "sflick@codingdojo.com" ]
sflick@codingdojo.com
25724b55aa4d29b7d7541545d4b7a257d6a3af48
a2accf55b7b57a376344689e1f8fa4d64acdd6be
/salaray_to_csv.py
692da5634ca9d92a5b99a5746186df6cf7f8d942
[]
no_license
Thybat/EulerProject
8e590455f2fbebceeec031e667ac77a97fd5d35c
54ee77fac34de4c13b33bf5459f6b6258d6cca1f
refs/heads/main
2022-12-27T23:19:25.058389
2020-10-09T09:57:54
2020-10-09T09:57:54
302,599,892
0
0
null
null
null
null
UTF-8
Python
false
false
495
py
import pandas as pd import matplotlib.pyplot as plt from sklearn import preprocessing from sklearn.linear_model import LinearRegression from sklearn.model_selection import train_test_split df = pd.read_csv("data_reg_age_salary.csv") x = df.values #returns a numpy array min_max_scaler = preprocessing.MinMaxScaler() pd.DataFrame(min_max_scaler.fit_transform(df.T), columns=df.columns, index=df.index) df = pd.DataFrame(x_scaled) df.to_csv("data_reg_age_salary.csv", index=False)
[ "noreply@github.com" ]
Thybat.noreply@github.com
3b930951163c7e85c90c1df23c7a762af58fc623
55c250525bd7198ac905b1f2f86d16a44f73e03a
/Python/Projects/pyinstaller/tests/unit/test_modulegraph/testpkg-relimport/pkg/relimport.py
0e76e39fedf777016e011f2daa0a233827236b57
[ "LicenseRef-scancode-other-permissive" ]
permissive
NateWeiler/Resources
213d18ba86f7cc9d845741b8571b9e2c2c6be916
bd4a8a82a3e83a381c97d19e5df42cbababfc66c
refs/heads/master
2023-09-03T17:50:31.937137
2023-08-28T23:50:57
2023-08-28T23:50:57
267,368,545
2
1
null
2022-09-08T15:20:18
2020-05-27T16:18:17
null
UTF-8
Python
false
false
127
py
version https://git-lfs.github.com/spec/v1 oid sha256:b1c982ab98ef0425093ad8bafe4c9deb7aad009260b384b762696e8fc4c0e375 size 22
[ "nateweiler84@gmail.com" ]
nateweiler84@gmail.com
0843c71fc383feb3fad8512aa89bc2e6a0434ce9
760b47b46574552c5371270036b391a6a3b7dc2f
/src/loggedfs/_core/fs.py
ced81b475209403f3dcd7fbc2e7e3a9c1a7796b4
[ "Apache-2.0" ]
permissive
pleiszenburg/loggedfs-python
7ae2ba90780f700b3ec5153b8b7447452bf643c7
1c0740f6f38e7795d6f834d08d9bab59d911b454
refs/heads/master
2023-04-06T05:00:50.457663
2020-07-11T08:33:41
2020-07-11T08:33:41
113,670,055
31
3
Apache-2.0
2023-03-31T02:14:20
2017-12-09T13:07:44
Python
UTF-8
Python
false
false
14,726
py
# -*- coding: utf-8 -*- """ LoggedFS-python Filesystem monitoring with Fuse and Python https://github.com/pleiszenburg/loggedfs-python src/loggedfs/_core/fs.py: File system core Copyright (C) 2017-2020 Sebastian M. Ernst <ernst@pleiszenburg.de> <LICENSE_BLOCK> The contents of this file are subject to the Apache License Version 2 ("License"). You may not use this file except in compliance with the License. You may obtain a copy of the License at https://www.apache.org/licenses/LICENSE-2.0 https://github.com/pleiszenburg/loggedfs-python/blob/master/LICENSE Software distributed under the License is distributed on an "AS IS" basis, WITHOUT WARRANTY OF ANY KIND, either express or implied. See the License for the specific language governing rights and limitations under the License. </LICENSE_BLOCK> """ # +++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++ # IMPORT # +++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++ import errno import os import stat from refuse.high import ( FUSE, fuse_get_context, FuseOSError, Operations ) from .defaults import ( FUSE_ALLOWOTHER_DEFAULT, FUSE_FOREGROUND_DEFAULT, LIB_MODE_DEFAULT, LOG_BUFFERS_DEFAULT, LOG_ENABLED_DEFAULT, LOG_JSON_DEFAULT, LOG_ONLYMODIFYOPERATIONS_DEFAULT, LOG_PRINTPROCESSNAME_DEFAULT, LOG_SYSLOG_DEFAULT ) from .filter import filter_pipeline_class from .log import get_logger, log_msg from .out import event from .timing import time # +++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++ # ROUTINES # +++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++ def loggedfs_factory(directory, **kwargs): if not isinstance(directory, str): raise TypeError('directory must be of type string') if not os.path.isdir(directory): raise ValueError('directory must be a path to an existing directory') if not isinstance(kwargs.get('fuse_foreground', FUSE_FOREGROUND_DEFAULT), bool): raise TypeError('fuse_foreground must be of type bool') if not isinstance(kwargs.get('fuse_allowother', FUSE_ALLOWOTHER_DEFAULT), bool): raise TypeError('fuse_allowother must be of type bool') return FUSE( _loggedfs( directory, **kwargs ), directory, raw_fi = True, nothreads = True, foreground = kwargs.get('fuse_foreground', FUSE_FOREGROUND_DEFAULT), allow_other = kwargs.get('fuse_allowother', FUSE_ALLOWOTHER_DEFAULT), default_permissions = kwargs.get('fuse_allowother', FUSE_ALLOWOTHER_DEFAULT), attr_timeout = 0, entry_timeout = 0, negative_timeout = 0, sync_read = False, # relying on fuse.Operations class defaults? # max_readahead = 0, # relying on fuse.Operations class defaults? # direct_io = True, # relying on fuse.Operations class defaults? nonempty = True, # common options taken from LoggedFS use_ino = True # common options taken from LoggedFS ) # +++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++ # CORE CLASS: Init and internal routines # +++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++ class _loggedfs(Operations): flag_utime_omit_ok = 1 use_ns = True _ST_FIELDS = tuple(i for i in dir(os.stat_result) if i.startswith('st_')) _STVFS_FIELDS = tuple(i for i in dir(os.statvfs_result) if i.startswith('f_')) def __init__(self, directory, fuse_foreground = FUSE_FOREGROUND_DEFAULT, fuse_allowother = FUSE_ALLOWOTHER_DEFAULT, lib_mode = LIB_MODE_DEFAULT, log_buffers = LOG_BUFFERS_DEFAULT, log_enabled = LOG_ENABLED_DEFAULT, log_file = None, log_filter = None, log_json = LOG_JSON_DEFAULT, log_only_modify_operations = LOG_ONLYMODIFYOPERATIONS_DEFAULT, log_printprocessname = LOG_PRINTPROCESSNAME_DEFAULT, log_syslog = LOG_SYSLOG_DEFAULT, **kwargs ): if log_filter is None: log_filter = filter_pipeline_class() if not isinstance(directory, str): raise TypeError('directory must be of type string') if not os.path.isdir(directory): raise ValueError('directory must be a path to an existing directory') if not os.access(directory, os.W_OK | os.R_OK): raise ValueError('not sufficient permissions on "directory"') if not isinstance(log_filter, filter_pipeline_class): raise TypeError('log_filter must either be None or of type filter_pipeline_class') if log_file is not None: if not os.path.isdir(os.path.dirname(log_file)): raise ValueError('path to logfile directory does not exist') if os.path.exists(log_file) and not os.path.isfile(log_file): raise ValueError('logfile exists and is not a file') if os.path.isfile(log_file) and not os.access(log_file, os.W_OK): raise ValueError('logfile exists and is not writeable') if not os.path.exists(log_file) and not os.access(directory, os.W_OK): raise ValueError('path to logfile directory is not writeable') if not isinstance(log_syslog, bool): raise TypeError('log_syslog must be of type bool') if not isinstance(log_enabled, bool): raise TypeError('log_enabled must be of type bool') if not isinstance(log_printprocessname, bool): raise TypeError('log_printprocessname must be of type bool') if not isinstance(log_json, bool): raise TypeError('log_json must be of type bool') if not isinstance(log_buffers, bool): raise TypeError('log_buffers must be of type bool') if not isinstance(lib_mode, bool): raise TypeError('lib_mode must be of type bool') if not isinstance(log_only_modify_operations, bool): raise TypeError('log_only_modify_operations must be of type bool') if not isinstance(fuse_foreground, bool): raise TypeError('fuse_foreground must be of type bool') if not isinstance(fuse_allowother, bool): raise TypeError('fuse_allowother must be of type bool') self._root_path = directory self._log_printprocessname = log_printprocessname self._log_json = log_json self._log_buffers = log_buffers self._log_filter = log_filter self._lib_mode = lib_mode self._log_only_modify_operations = log_only_modify_operations self._logger = get_logger('LoggedFS-python', log_enabled, log_file, log_syslog, self._log_json) if fuse_foreground: self._logger.info(log_msg(self._log_json, 'LoggedFS-python not running as a daemon')) if fuse_allowother: self._logger.info(log_msg(self._log_json, 'LoggedFS-python running as a public filesystem')) if log_file is not None: self._logger.info(log_msg(self._log_json, 'LoggedFS-python log file: %s' % log_file)) self._logger.info(log_msg(self._log_json, 'LoggedFS-python starting at %s' % directory)) try: self._root_path_fd = os.open(directory, os.O_RDONLY) except Exception as e: self._logger.exception('Directory access failed.') raise e log_configfile = kwargs.pop('_log_configfile', None) if log_configfile is not None: self._logger.info(log_msg(self._log_json, 'LoggedFS-python using configuration file %s' % log_configfile )) if len(kwargs) > 0: raise ValueError('unknown keyword argument(s)') def _full_path(self, partial_path): if partial_path.startswith('/'): partial_path = partial_path[1:] path = os.path.join(self._root_path, partial_path) return path @staticmethod def _rel_path(partial_path): if len(partial_path) == 0: return '.' elif partial_path == '/': return '.' elif partial_path.startswith('/'): return partial_path[1:] elif partial_path.startswith('./'): return partial_path[2:] else: return partial_path # +++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++ # CORE CLASS: Filesystem & file methods - STUBS # ... addressing https://github.com/fusepy/fusepy/issues/81 # +++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++ def create(self, path, mode, fi = None): raise FuseOSError(errno.ENOSYS) def flush(self, path, fip): raise FuseOSError(errno.ENOSYS) def fsync(self, path, datasync, fip): raise FuseOSError(errno.ENOSYS) # the original loggedfs just returns 0 def ioctl(self, path, cmd, arg, fh, flags, data): raise FuseOSError(errno.ENOSYS) def lock(self, path, fh, cmd, lock): raise FuseOSError(errno.ENOSYS) # +++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++ # CORE CLASS: Filesystem & file methods - IMPLEMENTATION # +++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++ @event(format_pattern = '{param_path}') def access(self, path, mode): if not os.access(self._rel_path(path), mode, dir_fd = self._root_path_fd): raise FuseOSError(errno.EACCES) @event(format_pattern = '{param_path} to {param_mode}') def chmod(self, path, mode): os.chmod(self._rel_path(path), mode, dir_fd = self._root_path_fd) @event(format_pattern = '{param_path} to {param_uid_name}({param_uid}):{param_gid_name}({param_gid})') def chown(self, path, uid, gid): os.chown(self._rel_path(path), uid, gid, dir_fd = self._root_path_fd, follow_symlinks = False) @event(format_pattern = '{param_path}') def destroy(self, path): os.close(self._root_path_fd) @event(format_pattern = '{param_path} (fh={param_fip})') def getattr(self, path, fip): if not fip: try: st = os.lstat(self._rel_path(path), dir_fd = self._root_path_fd) except FileNotFoundError: raise FuseOSError(errno.ENOENT) else: st = os.fstat(fip.fh) ret_dict = {key: getattr(st, key) for key in self._ST_FIELDS} for key in ['st_atime', 'st_ctime', 'st_mtime']: ret_dict[key] = ret_dict.pop(key + '_ns') return ret_dict @event(format_pattern = '{param_path}') def init(self, path): pass @event(format_pattern = '{param_source_path} to {param_target_path}') def link(self, target_path, source_path): target_rel_path = self._rel_path(target_path) os.link( self._rel_path(source_path), target_rel_path, src_dir_fd = self._root_path_fd, dst_dir_fd = self._root_path_fd ) uid, gid, pid = fuse_get_context() os.chown(target_rel_path, uid, gid, dir_fd = self._root_path_fd, follow_symlinks = False) @event(format_pattern = '{param_path} {param_mode}') def mkdir(self, path, mode): rel_path = self._rel_path(path) os.mkdir(rel_path, mode, dir_fd = self._root_path_fd) uid, gid, pid = fuse_get_context() os.chown(rel_path, uid, gid, dir_fd = self._root_path_fd, follow_symlinks = False) os.chmod(rel_path, mode, dir_fd = self._root_path_fd) # follow_symlinks = False @event(format_pattern = '{param_path} {param_mode}') def mknod(self, path, mode, dev): rel_path = self._rel_path(path) if stat.S_ISREG(mode): res = os.open( rel_path, os.O_CREAT | os.O_EXCL | os.O_WRONLY, mode, dir_fd = self._root_path_fd ) # TODO broken, applies umask to mode no matter what ... if res >= 0: os.close(res) elif stat.S_ISFIFO(mode): os.mkfifo(rel_path, mode, dir_fd = self._root_path_fd) else: os.mknod(rel_path, mode, dev, dir_fd = self._root_path_fd) uid, gid, pid = fuse_get_context() os.chown(rel_path, uid, gid, dir_fd = self._root_path_fd, follow_symlinks = False) os.chmod(rel_path, mode, dir_fd = self._root_path_fd) # follow_symlinks = False @event(format_pattern = '({param_fip}) {param_path} (fh={param_fip})') def open(self, path, fip): fip.fh = os.open(self._rel_path(path), fip.flags, dir_fd = self._root_path_fd) return 0 @event(format_pattern = '{param_length} bytes from {param_path} at offset {param_offset} (fh={param_fip})') def read(self, path, length, offset, fip): ret = os.pread(fip.fh, length, offset) return ret @event(format_pattern = '{param_path}') def readdir(self, path, fh): rel_path = self._rel_path(path) dirents = ['.', '..'] if stat.S_ISDIR(os.lstat(rel_path, dir_fd = self._root_path_fd).st_mode): dir_fd = os.open(rel_path, os.O_RDONLY, dir_fd = self._root_path_fd) dirents.extend(os.listdir(dir_fd)) os.close(dir_fd) return dirents @event(format_pattern = '{param_path}') def readlink(self, path): pathname = os.readlink(self._rel_path(path), dir_fd = self._root_path_fd) if pathname.startswith('/'): # TODO check this ... actually required? return os.path.relpath(pathname, self._root_path) else: return pathname @event(format_pattern = '{param_path} (fh={param_fip})') def release(self, path, fip): os.close(fip.fh) @event(format_pattern = '{param_old_path} to {param_new_path}') def rename(self, old_path, new_path): os.rename( self._rel_path(old_path), self._rel_path(new_path), src_dir_fd = self._root_path_fd, dst_dir_fd = self._root_path_fd ) @event(format_pattern = '{param_path}') def rmdir(self, path): os.rmdir(self._rel_path(path), dir_fd = self._root_path_fd) @event(format_pattern = '{param_path}') def statfs(self, path): fd = os.open(self._rel_path(path), os.O_RDONLY, dir_fd = self._root_path_fd) stv = os.statvfs(fd) os.close(fd) return {key: getattr(stv, key) for key in self._STVFS_FIELDS} @event(format_pattern = 'from {param_source_path} to {param_target_path_}') def symlink(self, target_path_, source_path): target_rel_path = self._rel_path(target_path_) os.symlink(source_path, target_rel_path, dir_fd = self._root_path_fd) uid, gid, pid = fuse_get_context() os.chown(target_rel_path, uid, gid, dir_fd = self._root_path_fd, follow_symlinks = False) @event(format_pattern = '{param_path} to {param_length} bytes (fh={param_fip})') def truncate(self, path, length, fip = None): if fip is None: fd = os.open(self._rel_path(path), flags = os.O_WRONLY, dir_fd = self._root_path_fd) ret = os.ftruncate(fd, length) os.close(fd) return ret else: return os.ftruncate(fip.fh, length) @event(format_pattern = '{param_path}') def unlink(self, path): os.unlink(self._rel_path(path), dir_fd = self._root_path_fd) @event(format_pattern = '{param_path}') def utimens(self, path, times = None): def _fix_time_(atime, mtime): if None in (atime, mtime): st = os.lstat(relpath, dir_fd = self._root_path_fd) if atime is None: atime = st.st_atime_ns if mtime is None: mtime = st.st_mtime_ns return (atime, mtime) relpath = self._rel_path(path) os.utime(relpath, ns = _fix_time_(*times), dir_fd = self._root_path_fd, follow_symlinks = False) @event(format_pattern = '{param_buf_len} bytes to {param_path} at offset {param_offset} (fh={param_fip})') def write(self, path, buf, offset, fip): res = os.pwrite(fip.fh, buf, offset) return res
[ "ernst@pleiszenburg.de" ]
ernst@pleiszenburg.de
f64c6ffb584cd043d80268c613a23fadf9f3d960
d0e268862f359bbeec426b00a0c45788f6fb0b4e
/lesson22-优化小实例/main.py
10fefdf9bbed196814d0bc20e94497752dbfa13d
[]
no_license
jpegbert/PyTorch
f87c2e38572c51842785de5ed1b39bb641402ac6
482421c76a093312ffeff7e5af4ecd3ab0cdcf30
refs/heads/master
2023-08-27T03:56:26.883297
2021-11-08T06:03:30
2021-11-08T06:03:30
326,677,679
2
1
null
null
null
null
UTF-8
Python
false
false
853
py
import numpy as np from matplotlib import pyplot as plt import torch def himmelblau(x): return (x[0] ** 2 + x[1] - 11) ** 2 + (x[0] + x[1] ** 2 - 7) ** 2 x = np.arange(-6, 6, 0.1) y = np.arange(-6, 6, 0.1) print('x,y range:', x.shape, y.shape) X, Y = np.meshgrid(x, y) print('X,Y maps:', X.shape, Y.shape) Z = himmelblau([X, Y]) fig = plt.figure('himmelblau') ax = fig.gca(projection='3d') ax.plot_surface(X, Y, Z) ax.view_init(60, -30) ax.set_xlabel('x') ax.set_ylabel('y') plt.show() # [1., 0.], [-4, 0.], [4, 0.] x = torch.tensor([-4., 0.], requires_grad=True) print(x) optimizer = torch.optim.Adam([x], lr=1e-3) for step in range(20000): pred = himmelblau(x) optimizer.zero_grad() pred.backward() optimizer.step() if step % 2000 == 0: print('step {}: x = {}, f(x) = {}'.format(step, x.tolist(), pred.item()))
[ "jiangpeng.jiang@zhaopin.com.cn" ]
jiangpeng.jiang@zhaopin.com.cn
853ea408b40f36263a1ac3f6bb6b89829d2bbd39
a639206e5849432d6613c9af093b14f97e9dd794
/manage.py
3cda76ad753a5eb70246fda28e6126466f0d11c0
[ "MIT" ]
permissive
Bazulenkov/primorsk-cup
6a37139b56b3b3df605f9fa49b1ef652a08d8020
13c227f9fb8289eb3e25d21e1de7a0fde8cb5f87
refs/heads/main
2023-06-10T00:21:27.939260
2021-06-30T12:36:04
2021-06-30T12:36:04
336,052,393
0
0
null
null
null
null
UTF-8
Python
false
false
664
py
#!/usr/bin/env python """Django's command-line utility for administrative tasks.""" import os import sys def main(): """Run administrative tasks.""" os.environ.setdefault("DJANGO_SETTINGS_MODULE", "primorsk.settings") try: from django.core.management import execute_from_command_line except ImportError as exc: raise ImportError( "Couldn't import Django. Are you sure it's installed and " "available on your PYTHONPATH environment variable? Did you " "forget to activate a virtual environment?" ) from exc execute_from_command_line(sys.argv) if __name__ == "__main__": main()
[ "anton@bazul.ru" ]
anton@bazul.ru
035504981c5c4ce873430a3439ea302e21244885
53c3462ff265b6273f4a4fa17f6d59688f69def0
/剑指offer/65_hasPath.py
fc4f3bf5633e263578be82adeacb409feea73759
[]
no_license
17764591637/jianzhi_offer
b76e69a3ecb2174676da2c8d8d3372a3fc27b5c4
27e420ee302d5ab6512ecfdb8d469b043fb7102d
refs/heads/master
2023-08-03T01:32:51.588472
2019-10-13T07:56:21
2019-10-13T07:56:21
197,692,548
2
0
null
null
null
null
UTF-8
Python
false
false
3,300
py
''' 请设计一个函数,用来判断在一个矩阵中是否存在一条包含某字符串所有字符的路径。 路径可以从矩阵中的任意一个格子开始,每一步可以在矩阵中向左,向右,向上,向下移动一个格子。 如果一条路径经过了矩阵中的某一个格子,则之后不能再次进入这个格子。 例如 a b c e s f c s a d e e 这样的3 X 4 矩阵中包含一条字符串"bcced"的路径,但是矩阵中不包含"abcb"路径, 因为字符串的第一个字符b占据了矩阵中的第一行第二个格子之后,路径不能再次进入该格子。 分析:回溯算法 这是一个可以用回朔法解决的典型题。首先,在矩阵中任选一个格子作为路径的起点。如果路径上的第i个字符不是ch, 那么这个格子不可能处在路径上的第i个位置。如果路径上的第i个字符正好是ch,那么往相邻的格子寻找路径上的第i+1个字符。 除在矩阵边界上的格子之外,其他格子都有4个相邻的格子。重复这个过程直到路径上的所有字符都在矩阵中找到相应的位置。   由于回朔法的递归特性,路径可以被开成一个栈。当在矩阵中定位了路径中前n个字符的位置之后,在与第n个字符对应的格子 的周围都没有找到第n+1个字符,这个时候只要在路径上回到第n-1个字符,重新定位第n个字符。   由于路径不能重复进入矩阵的格子,还需要定义和字符矩阵大小一样的布尔值矩阵,用来标识路径是否已经进入每个格子。 当矩阵中坐标为(row,col)的格子和路径字符串中相应的字符一样时,从4个相邻的格子(row,col-1),(row-1,col), (row,col+1)以及(row+1,col)中去定位路径字符串中下一个字符如果4个相邻的格子都没有匹配字符串中下一个的字符, 表明当前路径字符串中字符在矩阵中的定位不正确,我们需要回到前一个,然后重新定位。一直重复这个过程, 直到路径字符串上所有字符都在矩阵中找到合适的位置。 ''' class Solution: def hasPath(self, matrix, rows, cols, path): # write code here for i in range(rows): for j in range(cols): if matrix[i*cols+j] == path[0]: #print(i,j) if self.find(list(matrix),rows,cols,path[1:],i,j): return True return False def find(self,matrix,rows,cols,path,i,j): if not path: return True matrix[i*cols+j]='0'#记录是否已经走过,0表示已经走过 if j+1<cols and matrix[i*cols+j+1]==path[0]: return self.find(matrix,rows,cols,path[1:],i,j+1)#往右 elif j-1>=0 and matrix[i*cols+j-1]==path[0]: return self.find(matrix,rows,cols,path[1:],i,j-1)#往左 elif i+1<rows and matrix[(i+1)*cols+j]==path[0]: return self.find(matrix,rows,cols,path[1:],i+1,j)#往下 elif i-1>=0 and matrix[(i-1)*cols+j]==path[0]: return self.find(matrix,rows,cols,path[1:],i-1,j)#往上 else: return False s = Solution() res = s.hasPath(['a','b','c', 'e', 's' ,'f' ,'c', 's', 'a', 'd', 'e', 'e'],3,4,'bcced') print(res)
[ "17764591637@163.com" ]
17764591637@163.com
c672da054ce83d887363f44a14b4d22240729758
131d941a4d1df110afe09733fe010da816597bb4
/PASA/alameda/input/motion-detection-camera/mmal/motion-detected.py
5cf1a72a7454179e2769cc7877f1faf7f8313954
[]
no_license
onebeartoe/electronic-signs
386cfb7288c211634a3afad55d334d4602178a95
621099b7d81995a9c7e977ca683bb3c3d02c9d17
refs/heads/master
2020-12-24T11:06:57.754906
2018-11-01T16:40:35
2018-11-01T16:40:35
13,177,627
0
0
null
null
null
null
UTF-8
Python
false
false
120
py
#import sendint; print "nothing happened in motion-detected.py" #print "motion detected" #sendint.sendInt(ser, '2')
[ "onebeartoe@gmail.com" ]
onebeartoe@gmail.com
5c65745999dcd0d78e74ff5c2c30969e2bf57ef8
b75aedc25188b059189df7654dacd32818a86de2
/Facial Emotions/emotions.py
ef7d2650e5198354441345cbe25e5092c2c92730
[]
no_license
scorpiocodes/ComputerVision
2308c9dcfb07fc85fdb8c46f45891ae0e4b106fa
a346ce69c81ae1a74cbd94f1ad8749a50aa44fbd
refs/heads/master
2020-03-22T01:53:54.861885
2019-06-11T18:39:34
2019-06-11T18:39:34
139,337,035
0
0
null
null
null
null
UTF-8
Python
false
false
3,457
py
import cv2 import numpy as np from keras.models import load_model from statistics import mode from utils.datasets import get_labels from utils.inference import detect_faces from utils.inference import draw_text from utils.inference import draw_bounding_box from utils.inference import apply_offsets from utils.inference import load_detection_model from utils.preprocessor import preprocess_input USE_WEBCAM = True # If false, loads video file source # parameters for loading data and images emotion_model_path = './models/emotion_model.hdf5' emotion_labels = get_labels('fer2013') # hyper-parameters for bounding boxes shape frame_window = 10 emotion_offsets = (20, 40) # loading models face_cascade = cv2.CascadeClassifier('./models/haarcascade_frontalface_default.xml') emotion_classifier = load_model(emotion_model_path) # getting input model shapes for inference emotion_target_size = emotion_classifier.input_shape[1:3] # starting lists for calculating modes emotion_window = [] # starting video streaming cv2.namedWindow('window_frame') video_capture = cv2.VideoCapture(0) # Select video or webcam feed cap = None if (USE_WEBCAM == True): cap = cv2.VideoCapture(0) # Webcam source else: cap = cv2.VideoCapture('./demo/dinner.mp4') # Video file source while cap.isOpened(): # True: ret, bgr_image = cap.read() #bgr_image = video_capture.read()[1] gray_image = cv2.cvtColor(bgr_image, cv2.COLOR_BGR2GRAY) rgb_image = cv2.cvtColor(bgr_image, cv2.COLOR_BGR2RGB) faces = face_cascade.detectMultiScale(gray_image, scaleFactor=1.1, minNeighbors=5, minSize=(30, 30), flags=cv2.CASCADE_SCALE_IMAGE) for face_coordinates in faces: x1, x2, y1, y2 = apply_offsets(face_coordinates, emotion_offsets) gray_face = gray_image[y1:y2, x1:x2] try: gray_face = cv2.resize(gray_face, (emotion_target_size)) except: continue gray_face = preprocess_input(gray_face, True) gray_face = np.expand_dims(gray_face, 0) gray_face = np.expand_dims(gray_face, -1) emotion_prediction = emotion_classifier.predict(gray_face) emotion_probability = np.max(emotion_prediction) emotion_label_arg = np.argmax(emotion_prediction) emotion_text = emotion_labels[emotion_label_arg] emotion_window.append(emotion_text) if len(emotion_window) > frame_window: emotion_window.pop(0) try: emotion_mode = mode(emotion_window) except: continue if emotion_text == 'angry': color = emotion_probability * np.asarray((255, 0, 0)) elif emotion_text == 'sad': color = emotion_probability * np.asarray((0, 0, 255)) elif emotion_text == 'happy': color = emotion_probability * np.asarray((255, 255, 0)) elif emotion_text == 'surprise': color = emotion_probability * np.asarray((0, 255, 255)) else: color = emotion_probability * np.asarray((0, 255, 0)) color = color.astype(int) color = color.tolist() draw_bounding_box(face_coordinates, rgb_image, color) draw_text(face_coordinates, rgb_image, emotion_mode, color, 0, -45, 1, 1) bgr_image = cv2.cvtColor(rgb_image, cv2.COLOR_RGB2BGR) cv2.imshow('window_frame', bgr_image) if cv2.waitKey(1) & 0xFF == ord('q'): break cap.release() cv2.destroyAllWindows()
[ "noreply@github.com" ]
scorpiocodes.noreply@github.com
c8217fd5718d9e8ce56aa8984ec31a0f48ea85a3
d95cd6f18f9f83df625362dd28ab6b733fa98c90
/extract_bboxes.py
ca82fd055ca9f70678b1b8d0c449e0586a5e3a0a
[]
no_license
hassony2/yolo2-pytorch
477196349ee1d44ad09376b8f02cb08cced559a2
b9e1e905115694df5dd33b8a7485977263e19caf
refs/heads/master
2021-01-20T21:41:48.529189
2017-09-21T13:24:54
2017-09-21T13:24:54
101,783,945
1
0
null
2017-09-21T13:24:56
2017-08-29T16:40:30
Python
UTF-8
Python
false
false
5,011
py
import numpy as np import os from PIL import Image import torch from torch.autograd import Variable from torchvision import transforms from tqdm import tqdm import cfgs.config as cfg from darknet import Darknet19 from datasets.gteagazeplusimage import GTEAGazePlusImage from datasets.smthgimage import SmthgImage from datasets.utils.visualize import plot_bboxes import utils.network as net_utils import utils.yolo as yolo_utils def get_crop_params(bbox, img_shape, increase_ratio=2.2): """ returns x_min, y_min, x_max ,y_max crop coordinates according to rule 2.2 times max dimension of the bounding box Args: bbox(numpy.ndarray): x_min, y_min, x_max, y_max img_shape(tuple): original image shape increase_ratio(float): final bbox size / tight bbox size """ width = bbox[2] - bbox[0] height = bbox[3] - bbox[1] square_dim = max(width, height) final_dim = square_dim * increase_ratio center_x = bbox[0] + width / 2 center_y = bbox[1] + height / 2 new_x_min = int(center_x - final_dim / 2) new_x_max = int(center_x + final_dim / 2) new_y_min = int(center_y - final_dim / 2) new_y_max = int(center_y + final_dim / 2) if new_x_min >= 0 and new_y_min >= 0 and\ new_x_max <= img_shape[0] and new_y_max <= img_shape[1]: success = True else: success = False return success, (new_x_min, new_y_min, new_x_max, new_y_max) def test_net(net, dataset, transform=None, max_per_image=300, thresh=0.5, num_classes=1, vis=False, crop_folder=None): # Initialize counter for number of cropped hands extracted_hands = 0 # Run through dataset for i, (img, annots) in tqdm(enumerate(dataset)): original_img = img np_original_img = np.array(original_img) if transform is not None: img = transform(img) # Add batch dimension img = img.unsqueeze(0) # Create GPU variable img_var = Variable(img.type(torch.FloatTensor)) img_var = img_var.cuda() # Detect hands bbox_pred, iou_pred, prob_pred = net(img_var) # to numpy bbox_pred = bbox_pred.data.cpu().numpy() iou_pred = iou_pred.data.cpu().numpy() prob_pred = prob_pred.data.cpu().numpy() bboxes, scores, cls_inds = yolo_utils.postprocess( bbox_pred, iou_pred, prob_pred, np_original_img.shape[0:2], cfg, thresh) for class_idx in range(num_classes): # Extract class inds = np.where(cls_inds == class_idx)[0] class_bboxes = bboxes[inds] class_scores = scores[inds] class_scores = class_scores[:, np.newaxis] # Create class detections in format # [[x_min, y_min, x_max, y_max, score], ...] if vis: fig = plot_bboxes(np_original_img, class_bboxes, class_scores) fig.savefig('bboxes_{:03}.jpg'.format(i), bbox_inches='tight') # Save crops to (368x368) images if crop_folder is not None: for i, bbox in enumerate(class_bboxes): crop_success, crop_params = get_crop_params(bbox, (original_img.width, original_img.height)) if crop_success: crop = original_img.crop((crop_params)) crop_name = 'rendered_{:03d}.jpg'.format( extracted_hands) crop = crop.resize((368, 368)) # if bbox[2] - bbox[0] > 100 and bbox[3] - bbox[1] > 100: # Mirror left hands if cls_inds[i] == 0: crop = crop.transpose(Image.FLIP_LEFT_RIGHT) print('saving image') crop.save(os.path.join(crop_folder, crop_name)) extracted_hands += 1 if __name__ == "__main__": vis = True crop_folder = 'results/crops' # Initialize dataset dataset = GTEAGazePlusImage() # dataset = SmthgImage() # Initialize test image transform test_transform = transforms.Compose([ transforms.Scale(cfg.inp_size), transforms.ToTensor()]) # Initialise network # trained_model = 'models/training/darknet19_all_exp1/darknet19_all_exp1_64.h5' trained_model = 'models/training/darknet19_all_exp1/darknet19_all_exp1_15.h5' net = Darknet19() net_utils.load_net(trained_model, net) net.cuda() net.eval() # Extract bounding boxes test_net(net, dataset, transform=test_transform, vis=vis, thresh=0.5, crop_folder=None)
[ "yana.hasson@inria.fr" ]
yana.hasson@inria.fr
4b35f9e5c7f11f607452a11c9cd445ed2278b7b9
54e8ac0398bfa33d9a1d40e5a8d6477e3806bb17
/RaspberryPiCode/getHSV.py
f25d56d310c28ed15cd468d7255e72f710e71a4d
[]
no_license
RoboLions/frc2016-vision
ac46a15ba3c85f713f2d86619bce8b27aa996174
c32e559485e956a33794fa5a453c7202d042c27c
refs/heads/master
2021-04-30T22:59:30.402487
2016-03-30T03:26:17
2016-03-30T03:26:17
50,622,375
0
0
null
null
null
null
UTF-8
Python
false
false
1,193
py
import cv2 import numpy as np # move u* and l* sliders to find upper and lower end of hsv range respectively. # hit q to quit cap=cv2.VideoCapture(0) def nothing(x): pass cv2.namedWindow("result") h,s,v=100,100,100 cv2.createTrackbar('lh', 'result', 0, 255, nothing) cv2.createTrackbar('ls', 'result', 0, 255, nothing) cv2.createTrackbar('lv', 'result', 0, 255, nothing) cv2.createTrackbar('uh', 'result', 0, 255, nothing) cv2.createTrackbar('us', 'result', 0, 255, nothing) cv2.createTrackbar('uv', 'result', 0, 255, nothing) while True: _, frame = cap.read() hsv = cv2.cvtColor(frame, cv2.COLOR_BGR2HSV ) lh=cv2.getTrackbarPos('lh', 'result') ls=cv2.getTrackbarPos('ls', 'result') lv=cv2.getTrackbarPos('lv', 'result') uh=cv2.getTrackbarPos('uh', 'result') us=cv2.getTrackbarPos('us', 'result') uv=cv2.getTrackbarPos('uv', 'result') lower = np.array([lh,ls,lv]) upper = np.array([uh,us,uv]) mask = cv2.inRange(hsv, lower, upper) result = cv2.bitwise_and(frame, frame, mask = mask ) cv2.imshow('result', result ) key = cv2.waitKey(1) & 0xFF if key == ord("q"): break cap.release cv2.destroyAllWindows()
[ "alan.glaser@gnail.com" ]
alan.glaser@gnail.com
87e04c086ca8fcfe065781aaefdf79add1f73023
14a93d1114c26f1483b711793297cd00c1abd37f
/data/data_dav/export.py
2cc46ba94df22917591e14f70aef822629e3ec10
[]
no_license
csmfindling/bci_eeg
7c50eca0c150e760a0e70892c1182f8862df3902
a5a4ab8456f0354904cae8d3817017861ebd192d
refs/heads/master
2021-05-09T21:34:35.513477
2018-01-29T16:10:54
2018-01-29T16:10:54
118,730,858
0
0
null
null
null
null
UTF-8
Python
false
false
3,089
py
import pickle import sys import numpy as np from scipy.signal import butter, lfilter, detrend from numpy.fft import fft, fftfreq import time as timelib import pickle # functions def butter_bandpass(lowcut, highcut, fs, order=9): nyq = 0.5 * fs low = lowcut / nyq high = highcut / nyq b, a = butter(order, [low, high], btype='band') return b, a def butter_bandpass_filter(data, lowcut, highcut, fs, order=9): b, a = butter_bandpass(lowcut, highcut, fs, order=order) y = lfilter(b, a, data) return y def butter_lowpass(lowcut, fs, order=9): nyq = 0.5 * fs low = lowcut / nyq b, a = butter(order, low, btype='low') return b, a def butter_lowpass_filter(data, lowcut, fs, order=9): b, a = butter_lowpass(lowcut, fs, order=order) y = lfilter(b, a, data) return y # data ref = pickle.load(open('python_data/rawref.pkl')) close = pickle.load(open('python_data/rawclose.pkl')) eye = pickle.load(open('python_data/rawblink.pkl')) # parameters dt = 0.004 # sampling frequency is of 1/0.004 = 250Hz # write in csv for reference = 30sec nb_points = int(30 * 1./0.004) data = ref[:, -nb_points:] y = butter_lowpass_filter(data, 30., 1./dt) # low pass filter y = (y - np.mean(y))/np.std(y) # normalize concat = np.concatenate((data[:-1], y[:-1]), axis=0) import csv with open('csv_data/reference.csv', 'wb') as csvfile: spamwriter = csv.writer(csvfile, delimiter=',', quoting=csv.QUOTE_MINIMAL) spamwriter.writerow(['eye', 'channel 2', 'channel 3', 'alpha', 'eye filtered', 'channel 2 filtered', 'channel 3 filtered', 'alpha filtered']) for idx in range(concat.shape[-1]): spamwriter.writerow(concat[:,idx]) # write in csv for close_eye = 60 sec nb_points = int(60 * 1./0.004) data = close[:, -nb_points:] y = butter_lowpass_filter(data, 30., 1./dt) # low pass filter y = (y - np.mean(y))/np.std(y) # normalize concat = np.concatenate((data[:-1], y[:-1]), axis=0) import csv with open('csv_data/close_eye.csv', 'wb') as csvfile: spamwriter = csv.writer(csvfile, delimiter=',', quoting=csv.QUOTE_MINIMAL) spamwriter.writerow(['eye', 'channel 2', 'channel 3', 'alpha', 'eye filtered', 'channel 2 filtered', 'channel 3 filtered', 'alpha filtered']) for idx in range(concat.shape[-1]): spamwriter.writerow(concat[:,idx]) # write in csv for close_eye = 30 sec nb_points = int(30 * 1./0.004) data = eye[:, -nb_points:] y = butter_lowpass_filter(data, 30., 1./dt) # low pass filter y = (y - np.mean(y))/np.std(y) # normalize concat = np.concatenate((data[:-1], y[:-1]), axis=0) import csv with open('csv_data/blinks.csv', 'wb') as csvfile: spamwriter = csv.writer(csvfile, delimiter=',', quoting=csv.QUOTE_MINIMAL) spamwriter.writerow(['eye', 'channel 2', 'channel 3', 'alpha', 'eye filtered', 'channel 2 filtered', 'channel 3 filtered', 'alpha filtered']) for idx in range(concat.shape[-1]): spamwriter.writerow(concat[:,idx])
[ "charles.findling@gmail.com" ]
charles.findling@gmail.com
e06e1100601a1bacb795bb1f1efe4a2e96a3d781
221d5405763d1a6ab3c6755583e557c14b9f3742
/gusregon/gus.py
ca34d58334fd1bd2383ff69dfbb77899f224cc07
[ "BSD-2-Clause" ]
permissive
tpro/django-gusregon
4bd7253be9d43345376e36312d763d4653d0bbcd
75d4f291ae805bd986e1b4cb03b3b94e52a48076
refs/heads/master
2021-01-18T18:11:50.746453
2015-04-19T11:00:49
2015-04-19T11:00:49
null
0
0
null
null
null
null
UTF-8
Python
false
false
2,656
py
import requests import json GUS_API_URL = 'https://wyszukiwarkaregon.stat.gov.pl/wsBIR/UslugaBIRzewnPubl.svc/ajaxEndpoint/' LOGIN_ENDPOINT = 'Zaloguj' CAPTCHA_ENDPOINT = 'PobierzCaptcha' CHECK_CAPTCHA_ENDPOINT = 'SprawdzCaptcha' SEARCH_ENDPOINT = 'daneSzukaj' COMPANY_DETAILS_ENDPOINT = 'DanePobierzPelnyRaport' class GUS(object): sid = None report_type = { 'F': 'DaneRaportFizycznaPubl', 'P': 'DaneRaportPrawnaPubl'} prefix_data = { 'F': 'fiz_', 'P': 'praw_'} def __init__(self, sid=None): self.sid = sid def login(self): data = {'pKluczUzytkownika': 'aaaaaabbbbbcccccdddd'} self.sid = self._post(LOGIN_ENDPOINT, data=json.dumps(data)) return self.sid def get_captcha(self): return self._post(CAPTCHA_ENDPOINT) def check_captcha(self, captcha): data = {'pCaptcha': captcha} return self._post( CHECK_CAPTCHA_ENDPOINT, data=json.dumps(data)) def search(self, nip=None, regon=None, krs=None, detailed=True, no_prefix=True): if not any([nip, regon, krs]): raise AttributeError( 'At least one parameter (nip, regon, krs) is required.') if nip: search_params = {'Nip': nip} elif regon: search_params = {'Regon': regon} else: search_params = {'Krs': krs} data = {'pParametryWyszukiwania': search_params} basic_info = self._post( SEARCH_ENDPOINT, data=json.dumps(data)) if not detailed or not basic_info: return basic_info basic_info = json.loads(basic_info)[0] data = { 'pNazwaRaportu': self.report_type.get(basic_info['Typ']), 'pRegon': basic_info['Regon'], 'pSilosID': 1, } details = json.loads(self._post( COMPANY_DETAILS_ENDPOINT, data=json.dumps(data)))[0] if no_prefix: return self._remove_prefix(details) return details def _post(self, url, **kwargs): headers = {'Content-Type': 'application/json'} if self.sid: headers.update({'sid': self.sid}) url = '%s%s' % (GUS_API_URL, url) response = requests.post(url, headers=headers, **kwargs) return json.loads(response.content)['d'] def _remove_prefix(self, data): data_without_prefix = {} for key, value in data.iteritems(): if key.startswith(tuple(self.prefix_data.values())): key = key[key.find('_') + 1:] data_without_prefix[key] = value return data_without_prefix
[ "adam@bogdal.pl" ]
adam@bogdal.pl
3ec809e921148f10c35e5d7afb4098a0d5cfa51e
3a107debc014817103fd6cf18a346ba87cd4a896
/indexapp/views.py
19013ae175afb5a1629f37931d307a8c04c3468e
[]
no_license
shuang3322/itsm_new_hj
20c979bb682af084d347a10f773591dc090e3a8d
a99384243ec0dc667e04edda30754e9cead96bbf
refs/heads/master
2021-08-22T22:30:20.032674
2018-11-27T06:17:27
2018-11-27T06:17:27
147,472,768
0
0
null
null
null
null
UTF-8
Python
false
false
432
py
from django.shortcuts import render # Create your views here. from django.shortcuts import HttpResponse from indexapp.models import IPdata def visit(request): ip = request.META.get('REMOTE_ADDR') print(ip) all = IPdata.objects.all() # for item in request.META: # print(item,request.META.get(item)) return render(request, "test.html", {'current_user': all,'re_ip':ip}) # # def add_IP(request): # for
[ "shuang0528@hotmail.com" ]
shuang0528@hotmail.com
26c0479d73b4773d979b4de73b5383b4ceeb2883
ff5705d2813486da67b8aca48cfd5bf2c6cce068
/2_text_classification.py
0fa709be5b6d1b79dc77f3609155e88c78f26369
[]
no_license
m-nasiruddin/text_blob
61934763586413115054f1557c62e0301ab70f87
1f95f7ee0632ddc6bc09cc619a0510114a6a93c6
refs/heads/master
2022-07-15T03:48:56.591056
2018-07-20T10:48:13
2018-07-20T10:48:13
null
0
0
null
null
null
null
UTF-8
Python
false
false
2,714
py
from textblob.classifiers import NaiveBayesClassifier from textblob import TextBlob # creating a custom sentiment analyzer # loading data and creating a classifier train = [('I love this sandwich.', 'pos'), ('this is an amazing place!', 'pos'), ('I feel very good about these beers.', 'pos'), ('this is my best work.', 'pos'), ("what an awesome view", 'pos'), ('I do not like this restaurant', 'neg'), ('I am tired of this stuff.', 'neg'), ("I can't deal with this", 'neg'), ('he is my sworn enemy!', 'neg'), ('my boss is horrible.', 'neg')] test = [('the beer was good.', 'pos'), ('I do not enjoy my job', 'neg'), ("I ain't feeling dandy today.", 'neg'), ("I feel amazing!", 'pos'), ('Gary is a friend of mine.', 'pos'), ("I can't believe I'm doing this.", 'neg')] cl = NaiveBayesClassifier(train) # creating a naive bayes classifier # or, open from a file # with open('data/input/train.json', 'r') as fp: # cl = NaiveBayesClassifier(fp, format="json") # classifying text print(cl.classify("This is an amazing library!")) # get the label probability distribution prob_dist = cl.prob_classify("This one's a doozy.") print(prob_dist.max()) print(round(prob_dist.prob("pos"), 2)) print(round(prob_dist.prob("neg"), 2)) # classifying textblob blob = TextBlob("The beer is good. But the hangover is horrible.", classifier=cl) print(blob.classify()) for s in blob.sentences: print(s) print(s.classify()) # evaluating classifiers print(cl.accuracy(test)) print(cl.show_informative_features(5)) # displaying a listing of the most informative features # updating classifiers wth new data new_data = [('She is my best friend.', 'pos'), ("I'm happy to have a new friend.", 'pos'), ("Stay thirsty, my friend.", 'pos'), ("He ain't from around here.", 'neg')] print(cl.update(new_data)) print(cl.accuracy(test)) # feature extractors # creating a feature extractor that just uses the first and last words of a document as its features def end_word_extractor(document): tokens = document.split() first_word, last_word = tokens[0], tokens[-1] feats = {} feats["first({0})".format(first_word)] = True feats["last({0})".format(last_word)] = False return feats features = end_word_extractor("I feel happy") assert features == {'last(happy)': False, 'first(I)': True} # using the feature extractor in a classifier by passing it as the second argument of the constructor cl2 = NaiveBayesClassifier(test, feature_extractor=end_word_extractor) blob = TextBlob("I'm excited to try my new classifier.", classifier=cl2) print(blob.classify())
[ "mohammad.nasiruddin@gmail.com" ]
mohammad.nasiruddin@gmail.com
842f5307feedb014785782c8f52774517fee51a9
969be1a2865722374293a286fa0923d6247d54d5
/language/bert/bert_code/fairseq/optim/lr_scheduler/fairseq_lr_scheduler.py
1732e2e495edd65fe93ea60644562119ae0b2e80
[ "MIT" ]
permissive
Relwayg/Differentially-Private-Deep-Learning
36935e882cd37482f690b7f5888835065aa5e127
55eb9a1b75472895588d388de7567f012c9bc9ed
refs/heads/main
2023-09-06T01:23:44.915967
2021-11-23T06:53:29
2021-11-23T06:53:29
null
0
0
null
null
null
null
UTF-8
Python
false
false
1,396
py
# Copyright (c) Facebook, Inc. and its affiliates. # # This source code is licensed under the MIT license found in the # LICENSE file in the root directory of this source tree. from .. import FairseqOptimizer class FairseqLRScheduler(object): def __init__(self, args, optimizer): super().__init__() if not isinstance(optimizer, FairseqOptimizer): raise ValueError('optimizer must be an instance of FairseqOptimizer') self.args = args self.optimizer = optimizer self.best = None @staticmethod def add_args(parser): """Add arguments to the parser for this LR scheduler.""" pass def state_dict(self): """Return the LR scheduler state dict.""" return {'best': self.best} def load_state_dict(self, state_dict): """Load an LR scheduler state dict.""" self.best = state_dict['best'] def step(self, epoch, val_loss=None): """Update the learning rate at the end of the given epoch.""" if val_loss is not None: if self.best is None: self.best = val_loss else: self.best = min(self.best, val_loss) def step_update(self, num_updates): """Update the learning rate after each update.""" return self.optimizer.get_lr() def reinit(self, total_num_update, num_updates): pass
[ "yudakuai18@163.com" ]
yudakuai18@163.com
444b7127cd68eeab661ca9a41c8296181217f564
2a527f6fdee469bf1c3d76c2519149d916fa51d2
/python-client/victorops_client/models/on_call_log.py
870c6428c3097ccdca27122e1b5ff2514712c456
[]
no_license
ArjunSangitrao/python-victorops
e261971800adb337da36b716f2c0fd0fc5c5973c
a9de3c93d141ad19213bdef6aa3790988bf03c66
refs/heads/master
2022-03-26T19:19:06.556731
2017-02-26T12:39:08
2017-02-26T12:39:08
null
0
0
null
null
null
null
UTF-8
Python
false
false
5,018
py
# coding: utf-8 """ VictorOps API This API allows you to interact with the VictorOps platform in various ways. Your account may be limited to a total number of API calls per month. Also, some of these API calls have rate limits. NOTE: In this documentation when creating a sample curl request (clicking the TRY IT OUT! button), in some API viewing interfaces, the '@' in an email address may be encoded. Please note that the REST endpoints will not process the encoded version. Make sure that the encoded character '%40' is changed to its unencoded form before submitting the curl request. OpenAPI spec version: 0.0.1 Generated by: https://github.com/swagger-api/swagger-codegen.git """ from pprint import pformat from six import iteritems import re class OnCallLog(object): """ NOTE: This class is auto generated by the swagger code generator program. Do not edit the class manually. """ def __init__(self, team_slug=None, start=None, end=None, user_logs=None): """ OnCallLog - a model defined in Swagger :param dict swaggerTypes: The key is attribute name and the value is attribute type. :param dict attributeMap: The key is attribute name and the value is json key in definition. """ self.swagger_types = { 'team_slug': 'str', 'start': 'datetime', 'end': 'datetime', 'user_logs': 'list[UserLog]' } self.attribute_map = { 'team_slug': 'teamSlug', 'start': 'start', 'end': 'end', 'user_logs': 'userLogs' } self._team_slug = team_slug self._start = start self._end = end self._user_logs = user_logs @property def team_slug(self): """ Gets the team_slug of this OnCallLog. :return: The team_slug of this OnCallLog. :rtype: str """ return self._team_slug @team_slug.setter def team_slug(self, team_slug): """ Sets the team_slug of this OnCallLog. :param team_slug: The team_slug of this OnCallLog. :type: str """ self._team_slug = team_slug @property def start(self): """ Gets the start of this OnCallLog. :return: The start of this OnCallLog. :rtype: datetime """ return self._start @start.setter def start(self, start): """ Sets the start of this OnCallLog. :param start: The start of this OnCallLog. :type: datetime """ self._start = start @property def end(self): """ Gets the end of this OnCallLog. :return: The end of this OnCallLog. :rtype: datetime """ return self._end @end.setter def end(self, end): """ Sets the end of this OnCallLog. :param end: The end of this OnCallLog. :type: datetime """ self._end = end @property def user_logs(self): """ Gets the user_logs of this OnCallLog. :return: The user_logs of this OnCallLog. :rtype: list[UserLog] """ return self._user_logs @user_logs.setter def user_logs(self, user_logs): """ Sets the user_logs of this OnCallLog. :param user_logs: The user_logs of this OnCallLog. :type: list[UserLog] """ self._user_logs = user_logs def to_dict(self): """ Returns the model properties as a dict """ result = {} for attr, _ in iteritems(self.swagger_types): value = getattr(self, attr) if isinstance(value, list): result[attr] = list(map( lambda x: x.to_dict() if hasattr(x, "to_dict") else x, value )) elif hasattr(value, "to_dict"): result[attr] = value.to_dict() elif isinstance(value, dict): result[attr] = dict(map( lambda item: (item[0], item[1].to_dict()) if hasattr(item[1], "to_dict") else item, value.items() )) else: result[attr] = value return result def to_str(self): """ Returns the string representation of the model """ return pformat(self.to_dict()) def __repr__(self): """ For `print` and `pprint` """ return self.to_str() def __eq__(self, other): """ Returns true if both objects are equal """ if not isinstance(other, OnCallLog): return False return self.__dict__ == other.__dict__ def __ne__(self, other): """ Returns true if both objects are not equal """ return not self == other
[ "vincent.desmet@honestbee.com" ]
vincent.desmet@honestbee.com
eb66983747fd37d5bad2b03c62aa2cb5b9820300
6923f79f1eaaba0ab28b25337ba6cb56be97d32d
/A_Primer_on_Scientific_Programming_with_Python/input/c2f_cml_v3.py
9d13a354db2390ff9b57049d4ab8c67135d83538
[]
no_license
burakbayramli/books
9fe7ba0cabf06e113eb125d62fe16d4946f4a4f0
5e9a0e03aa7ddf5e5ddf89943ccc68d94b539e95
refs/heads/master
2023-08-17T05:31:08.885134
2023-08-14T10:05:37
2023-08-14T10:05:37
72,460,321
223
174
null
2022-10-24T12:15:06
2016-10-31T17:24:00
Jupyter Notebook
UTF-8
Python
false
false
217
py
import sys try: C = float(sys.argv[1]) except: print 'You failed to provide Celsius degrees as input '\ 'on the command line!' sys.exit(1) # abort F = 9.0*C/5 + 32 print '%gC is %.1fF' % (C, F)
[ "bb@b.om" ]
bb@b.om
625ab94e1d6b0ee32774ffd6931f538b4fdb913f
6298ff0f597ec09633696622c42321dd6bbaab6d
/znachkov_oleksandr/03/task3.1.py
061a601d2277c930b0c29708f2bee2eb8ab3078c
[]
no_license
Neckmus/itea_python_basics_3
94db7ea756cbb164b8705445b1ab027bb5ff4d93
9d76fc3d081fa407805e98d12f7f9960c8976366
refs/heads/master
2022-05-28T08:28:00.415625
2020-04-22T11:03:25
2020-04-22T11:03:25
254,841,391
0
0
null
2020-04-11T10:14:57
2020-04-11T10:14:57
null
UTF-8
Python
false
false
455
py
from random import randint ELEMENTS = 5 HIGHLIMIT = 10 exclusion_list = [4,2] def array_diff(source_list): print ("souerce list:", source_list) print ("exclusion list:", exclusion_list) #print (exclusion_list) return [ elem for elem in source_list if not elem in exclusion_list] source_list = lambda listsize, upper : [randint(0, upper) for index in range (listsize)] print ("result:", array_diff(source_list(ELEMENTS, HIGHLIMIT)))
[ "alx.inferno@gmail.com" ]
alx.inferno@gmail.com
2d46fb8bfbd693468cd059acdc41ca93221da9c6
c90b3ac3e5ad11cb93d4e6b76b9b9c4a19d0f512
/.history/copytest_20200502124009.py
413d6f42f34cf44c8b5ee71050cdd80e9dccb60d
[]
no_license
rbafna6507/passwordstorageproject
6465585e36c81075856af8d565fe83e358b4a40a
480c30e358f7902ac0ef5c4e8d9556cb1d6d33f4
refs/heads/master
2022-11-25T12:05:02.625968
2020-07-27T21:33:38
2020-07-27T21:33:38
283,021,426
0
0
null
null
null
null
UTF-8
Python
false
false
954
py
import pickle import cryptography from cryptography.fernet import Fernet def encrypt(message: bytes, key: bytes) -> bytes: return Fernet(key).encrypt(message) def decrypt(token: bytes, key: bytes) -> bytes: return Fernet(key).decrypt(token) infile = open('jeff.pkl','rb') z = pickle.load(infile) key = Fernet.generate_key() e_userpass = z username = input("Username: ") password = input("password: ") website = input("Website: ") e_username = encrypt(username.encode(), key) e_password = encrypt(password.encode(), key) e_list = [b"Username: " + e_username, b"Password: " + e_password] e_userpass["Website: " + website] = e_list outfile = open("jeff.pkl", "wb") pickle.dump(e_userpass, outfile) outfile.close() infile = open('jeff.pkl','rb') z = pickle.load(infile) e_userpass = z j = [e_userpass[k] for k in e_userpass] e = [r.encode() for r in j] q = decrypt(e, key) """for key, value in d_userpass.items(): print(key, ' : ', value)"""
[ "35872545+rbafna6507@users.noreply.github.com" ]
35872545+rbafna6507@users.noreply.github.com
4ff5f832176e95271a6ca6a4df8efc6d8ebb25b5
2dba100f7ebbbb7b6f1850371b8640e0d3c0fc04
/ScientificProgramming/mu.py
1df2355ea7dc71793588406a3bf9cbbd0ecb49f6
[]
no_license
lukasrieger-dev/python-beginner-projects
ea88fca0dc818ed2a63145305e49481dfa5b11af
d7cb2970a4067d150db57985e67e3cb6d3da7009
refs/heads/master
2023-01-04T12:19:34.881748
2020-11-06T18:10:52
2020-11-06T18:10:52
255,638,385
10
1
null
2020-10-03T07:25:23
2020-04-14T14:51:22
Python
UTF-8
Python
false
false
1,059
py
%matplotlib inline import re import matplotlib.pyplot as plt def load_mu_data(path): mu_data = dict() with open(path, 'r') as file: for line in file: if re.match(r'(^ *$|^#.*$)', line): # omit empty lines and comments continue *gas, C, T_0, mu_0 = line.split() gas = ''.join(gas) # for names with more than one word data = {'C':float(C), 'T_0':float(T_0), 'mu_0':float(mu_0)} mu_data[gas] = data return mu_data def mu(T, gas, mu_data): if not gas in mu_data: raise ValueError data = mu_data[gas] T_0 = data['T_0'] C = data['C'] mu_0 = data['mu_0'] mu_T = mu_0 * (T_0-C)/(T+C) * (T/T_0) return mu_T path = 'viscosity_of_gases.dat' mu_data = load_mu_data(path) Ts = list(range(223, 374)) for gas in mu_data: mu_values = [mu(T, gas, mu_data) for T in Ts] plt.plot(Ts, mu_values) plt.legend(list(mu_data)) plt.xlabel('temperatures') plt.ylabel('viscosity') plt.show()
[ "rieger.lks@gmail.com" ]
rieger.lks@gmail.com
68870acb9c8c92e4ebbaf015abbd0b6109358e24
97aa1181a8305fab0cfc635954c92880460ba189
/torch/__init__.py
4442d7e952cc12a7e75937c35b091330e7c69570
[ "BSD-2-Clause" ]
permissive
zhujiang73/pytorch_mingw
64973a4ef29cc10b96e5d3f8d294ad2a721ccacb
b0134a0acc937f875b7c4b5f3cef6529711ad336
refs/heads/master
2022-11-05T12:10:59.045925
2020-08-22T12:10:32
2020-08-22T12:10:32
123,688,924
8
4
NOASSERTION
2022-10-17T12:30:52
2018-03-03T12:15:16
C++
UTF-8
Python
false
false
14,830
py
r""" The torch package contains data structures for multi-dimensional tensors and mathematical operations over these are defined. Additionally, it provides many utilities for efficient serializing of Tensors and arbitrary types, and other useful utilities. It has a CUDA counterpart, that enables you to run your tensor computations on an NVIDIA GPU with compute capability >= 3.0. """ import os import sys import platform import ctypes if sys.version_info < (3,): raise Exception("Python 2 has reached end-of-life and is no longer supported by PyTorch.") from ._utils import _import_dotted_name from ._utils_internal import get_file_path, prepare_multiprocessing_environment, \ USE_RTLD_GLOBAL_WITH_LIBTORCH, USE_GLOBAL_DEPS #from .version import __version__ from ._six import string_classes as _string_classes from typing import Set, Type __all__ = [ 'typename', 'is_tensor', 'is_storage', 'set_default_tensor_type', 'set_rng_state', 'get_rng_state', 'manual_seed', 'initial_seed', 'seed', 'save', 'load', 'set_printoptions', 'chunk', 'split', 'stack', 'matmul', 'no_grad', 'enable_grad', 'rand', 'randn', 'DoubleStorage', 'FloatStorage', 'LongStorage', 'IntStorage', 'ShortStorage', 'CharStorage', 'ByteStorage', 'BoolStorage', 'DoubleTensor', 'FloatTensor', 'LongTensor', 'IntTensor', 'ShortTensor', 'CharTensor', 'ByteTensor', 'BoolTensor', 'Tensor', 'lobpcg', '_set_deterministic', '_is_deterministic' ] ################################################################################ # Load the extension module ################################################################################ # See Note [Global dependencies] def _load_global_deps(): if platform.system() == 'Windows': return lib_name = 'libtorch_global_deps' + ('.dylib' if platform.system() == 'Darwin' else '.so') here = os.path.abspath(__file__) lib_path = os.path.join(os.path.dirname(here), 'lib', lib_name) ctypes.CDLL(lib_path, mode=ctypes.RTLD_GLOBAL) if (USE_RTLD_GLOBAL_WITH_LIBTORCH or os.getenv('TORCH_USE_RTLD_GLOBAL')) and \ platform.system() != 'Windows': # Do it the hard way. You might want to load libtorch with RTLD_GLOBAL in a # few circumstances: # # 1. You're in a build environment (e.g., fbcode) where # libtorch_global_deps is not available, but you still need # to get mkl to link in with RTLD_GLOBAL or it will just # not work. # # 2. You're trying to run PyTorch under UBSAN and you need # to ensure that only one copy of libtorch is loaded, so # vptr checks work properly # # If you're using this setting, you must verify that all the libraries # you load consistently use the same libstdc++, or you may have # mysterious segfaults. # import os as _dl_flags if not hasattr(_dl_flags, 'RTLD_GLOBAL') or not hasattr(_dl_flags, 'RTLD_LAZY'): try: # next try if DLFCN exists import DLFCN as _dl_flags # type: ignore except ImportError: # as a last attempt, use compile-time constants import torch._dl as _dl_flags # type: ignore old_flags = sys.getdlopenflags() sys.setdlopenflags(_dl_flags.RTLD_GLOBAL | _dl_flags.RTLD_LAZY) from torch._C import * sys.setdlopenflags(old_flags) del old_flags del _dl_flags else: # Easy way. You want this most of the time, because it will prevent # C++ symbols from libtorch clobbering C++ symbols from other # libraries, leading to mysterious segfaults. # # If building in an environment where libtorch_global_deps isn't available # like parts of fbsource, but where RTLD_GLOBAL causes segfaults, you will # want USE_RTLD_GLOBAL_WITH_LIBTORCH = False and USE_GLOBAL_DEPS = False # # See Note [Global dependencies] if USE_GLOBAL_DEPS: _load_global_deps() from torch._C import * # Appease the type checker; ordinarily this binding is inserted by the # torch._C module initialization code in C if False: import torch._C as _C __all__ += [name for name in dir(_C) if name[0] != '_' and not name.endswith('Base')] ################################################################################ # Define basic utilities ################################################################################ def typename(o): if isinstance(o, torch.Tensor): return o.type() module = '' class_name = '' if hasattr(o, '__module__') and o.__module__ != 'builtins' \ and o.__module__ != '__builtin__' and o.__module__ is not None: module = o.__module__ + '.' if hasattr(o, '__qualname__'): class_name = o.__qualname__ elif hasattr(o, '__name__'): class_name = o.__name__ else: class_name = o.__class__.__name__ return module + class_name def is_tensor(obj): r"""Returns True if `obj` is a PyTorch tensor. Note that this function is simply doing ``isinstance(obj, Tensor)``. Using that ``isinstance`` check is better for typechecking with mypy, and more explicit - so it's recommended to use that instead of ``is_tensor``. Args: obj (Object): Object to test """ return isinstance(obj, torch.Tensor) def is_storage(obj): r"""Returns True if `obj` is a PyTorch storage object. Args: obj (Object): Object to test """ return type(obj) in _storage_classes def set_default_tensor_type(t): r"""Sets the default ``torch.Tensor`` type to floating point tensor type ``t``. This type will also be used as default floating point type for type inference in :func:`torch.tensor`. The default floating point tensor type is initially ``torch.FloatTensor``. Args: t (type or string): the floating point tensor type or its name Example:: >>> torch.tensor([1.2, 3]).dtype # initial default for floating point is torch.float32 torch.float32 >>> torch.set_default_tensor_type(torch.DoubleTensor) >>> torch.tensor([1.2, 3]).dtype # a new floating point tensor torch.float64 """ if isinstance(t, _string_classes): t = _import_dotted_name(t) _C._set_default_tensor_type(t) def set_default_dtype(d): r"""Sets the default floating point dtype to :attr:`d`. This dtype is: 1. The inferred dtype for python floats in :func:`torch.tensor`. 2. Used to infer dtype for python complex numbers. The default complex dtype is set to ``torch.complex128`` if default floating point dtype is ``torch.float64``, otherwise it's set to ``torch.complex64`` The default floating point dtype is initially ``torch.float32``. Args: d (:class:`torch.dtype`): the floating point dtype to make the default Example:: >>> # initial default for floating point is torch.float32 >>> torch.tensor([1.2, 3]).dtype torch.float32 >>> # initial default for floating point is torch.complex64 >>> torch.tensor([1.2, 3j]).dtype torch.complex64 >>> torch.set_default_dtype(torch.float64) >>> torch.tensor([1.2, 3]).dtype # a new floating point tensor torch.float64 >>> torch.tensor([1.2, 3j]).dtype # a new complex tensor torch.complex128 """ _C._set_default_dtype(d) def _set_deterministic(d): r"""Sets a global flag to force all operations to use a deterministic implementation if available. If an operation that does not have a deterministic implementation is called while this setting is True, the operation will throw a RuntimeError. Note that deterministic operations tend to have worse performance than non-deterministic operations. Args: d (:class:`bool`): If True, force operations to be deterministic. If False, allow non-deterministic operations. .. warning:: This feature is experimental and not complete. The above docstring represents what the future behavior is intended to be. Right now, `_set_deterministic` will only affect `torch.bmm` and convolution operators. """ _C._set_deterministic(d) def _is_deterministic(): r"""Returns True if the global deterministic flag is turned on and operations are being forced to use a deterministic implementation. .. warning:: This feature is experimental and not complete. The above docstring represents what the future behavior is intended to be. Right now, the global deterministic flag will only affect `torch.bmm` and convolution operators. """ return _C._get_deterministic() # If you edit these imports, please update torch/__init__.py.in as well from .random import set_rng_state, get_rng_state, manual_seed, initial_seed, seed from .serialization import save, load from ._tensor_str import set_printoptions ################################################################################ # Define Storage and Tensor classes ################################################################################ from .tensor import Tensor from .storage import _StorageBase class DoubleStorage(_C.DoubleStorageBase, _StorageBase): pass class FloatStorage(_C.FloatStorageBase, _StorageBase): pass class HalfStorage(_C.HalfStorageBase, _StorageBase): pass class LongStorage(_C.LongStorageBase, _StorageBase): pass class IntStorage(_C.IntStorageBase, _StorageBase): pass class ShortStorage(_C.ShortStorageBase, _StorageBase): pass class CharStorage(_C.CharStorageBase, _StorageBase): pass class ByteStorage(_C.ByteStorageBase, _StorageBase): pass class BoolStorage(_C.BoolStorageBase, _StorageBase): pass class BFloat16Storage(_C.BFloat16StorageBase, _StorageBase): pass class ComplexDoubleStorage(_C.ComplexDoubleStorageBase, _StorageBase): pass class ComplexFloatStorage(_C.ComplexFloatStorageBase, _StorageBase): pass class QUInt8Storage(_C.QUInt8StorageBase, _StorageBase): pass class QInt8Storage(_C.QInt8StorageBase, _StorageBase): pass class QInt32Storage(_C.QInt32StorageBase, _StorageBase): pass _storage_classes = { DoubleStorage, FloatStorage, LongStorage, IntStorage, ShortStorage, CharStorage, ByteStorage, HalfStorage, BoolStorage, QUInt8Storage, QInt8Storage, QInt32Storage, BFloat16Storage, ComplexFloatStorage, ComplexDoubleStorage } # The _tensor_classes set is initialized by the call to _C._initialize_tensor_type_bindings() _tensor_classes: Set[Type] = set() ################################################################################ # Initialize extension ################################################################################ def manager_path(): if platform.system() == 'Windows': return b"" path = get_file_path('torch', 'bin', 'torch_shm_manager') prepare_multiprocessing_environment(get_file_path('torch')) if not os.path.exists(path): raise RuntimeError("Unable to find torch_shm_manager at " + path) return path.encode('utf-8') # Shared memory manager needs to know the exact location of manager executable _C._initExtension(manager_path()) del manager_path # Appease the type checker: it can't deal with direct setting of globals(). # Note that we will see "too many" functions when reexporting this way; there # is not a good way to fix this problem. Perhaps, try to redesign VariableFunctions # so that this import is good enough if False: from torch._C._VariableFunctions import * for name in dir(_C._VariableFunctions): if name.startswith('__'): continue globals()[name] = getattr(_C._VariableFunctions, name) __all__.append(name) ################################################################################ # Import interface functions defined in Python ################################################################################ # needs to be after the above ATen bindings so we can overwrite from Python side from .functional import * ################################################################################ # Remove unnecessary members ################################################################################ del DoubleStorageBase del FloatStorageBase del LongStorageBase del IntStorageBase del ShortStorageBase del CharStorageBase del ByteStorageBase del BoolStorageBase del QUInt8StorageBase del BFloat16StorageBase del ComplexDoubleStorageBase del ComplexFloatStorageBase ################################################################################ # Import most common subpackages ################################################################################ import torch.cuda import torch.autograd from torch.autograd import no_grad, enable_grad, set_grad_enabled import torch.futures import torch.nn import torch.nn.intrinsic import torch.nn.quantized import torch.optim import torch.multiprocessing import torch.sparse import torch.utils.backcompat import torch.onnx import torch.jit import torch.hub import torch.random import torch.distributions import torch.testing import torch.backends.cuda import torch.backends.mkl import torch.backends.mkldnn import torch.backends.openmp import torch.backends.quantized import torch.quantization import torch.utils.data import torch.__config__ import torch.__future__ _C._init_names(list(torch._storage_classes)) # attach docstrings to torch and tensor functions from . import _torch_docs, _tensor_docs, _storage_docs del _torch_docs, _tensor_docs, _storage_docs def compiled_with_cxx11_abi(): r"""Returns whether PyTorch was built with _GLIBCXX_USE_CXX11_ABI=1""" return _C._GLIBCXX_USE_CXX11_ABI # Import the ops "namespace" from torch._ops import ops from torch._classes import classes # Import the quasi random sampler import torch.quasirandom # If you are seeing this, it means that this call site was not checked if # the memory format could be preserved, and it was switched to old default # behaviour of contiguous legacy_contiguous_format = contiguous_format # Register fork handler to initialize OpenMP in child processes (see gh-28389) from torch.multiprocessing._atfork import register_after_fork register_after_fork(torch.get_num_threads) del register_after_fork # Import tools that require fully imported torch (for applying # torch.jit.script as a decorator, for instance): from ._lobpcg import lobpcg # These were previously defined in native_functions.yaml and appeared on the # `torch` namespace, but we moved them to c10 dispatch to facilitate custom # class usage. We add these lines here to preserve backward compatbility. quantized_lstm = torch.ops.aten.quantized_lstm quantized_gru = torch.ops.aten.quantized_gru
[ "zhujiangmail@hotmail.com" ]
zhujiangmail@hotmail.com
a19137e2bc295d4d4e9c77c15d61e3a9e4d708f9
ff20661ef00b2db927c78f95a08cd6c40f950ee0
/inputmorethanone.py
81994128fa875ec38b52ef7cf8ec19866fc7810f
[]
no_license
Techsrijan/mppython2021
57ca26e1acdf5adad2afa692dd5ae23336273603
583a991f85e2414c6b8ffe0405f727f3f5d38eee
refs/heads/main
2023-06-18T22:05:44.602220
2021-07-16T00:42:26
2021-07-16T00:42:26
374,290,977
0
9
null
null
null
null
UTF-8
Python
false
false
231
py
'''f=int(input("Enter the first number")) s=int(input("Enter the Second number")) ''' f,s=input("Enter two number").split(',') print("F=",f,"S=",s) j,k=input("Enter two number").split(' ') print("j=",j,"k=",k) print("add=",j+k)
[ "aswanibtech@gmail.com" ]
aswanibtech@gmail.com
9436483bd9dce7f9bd3c1c1c7e5b4682ce2f8b80
e357f337ef80487aa6fafacb3e61cd5b56c78b60
/src/verification/__init__.py
9c7d02b050afd632c953e658bd2ee2c42fdd8642
[]
no_license
kaleissin/django-verification
1537b7e78a8ca07a091bcad5fd014a4544809ba5
33f7d0b7d9cc1494269f16b6f07f6077ca974fff
refs/heads/master
2021-07-15T20:12:43.628199
2021-07-05T11:43:39
2021-07-05T13:08:50
18,205,426
12
2
null
2021-07-05T13:13:16
2014-03-28T08:19:41
Python
UTF-8
Python
false
false
60
py
default_app_config = 'verification.apps.VerificationConfig'
[ "kaleissin@gmail.com" ]
kaleissin@gmail.com
ac8d8c76de9def04f8aad137f4e5827afd1ca93c
1d70ee049c5198b75567e0000c45ef879f6d39be
/JobMatchPonos/server/utils/word2vec/wordEmbeddings.py
7f5247842cca39719da0f18965c455959fc9dc9c
[]
no_license
alexawl/Job-Match-Ponos-Back
95f28185f71c38733973bc6d730947455c2e6c93
c48b4bfddfbf2f4f5aa95409fd2c6ee4f469d9dd
refs/heads/master
2022-09-30T15:07:19.875016
2020-02-01T06:50:34
2020-02-01T06:50:34
237,575,072
0
0
null
2021-05-06T20:07:41
2020-02-01T06:49:35
Python
UTF-8
Python
false
false
4,009
py
from gensim.models import Word2Vec, KeyedVectors from os import listdir from os.path import isfile, join import numpy as np from scipy import spatial from sklearn import decomposition import matplotlib.pyplot as plt from jobmatcher.server.utils.pattern import text # def read_All_CV(filename): # text = textract.process(filename) # return text.decode('utf-8') allText = " Chris Ware 789 E 901 N , Salt Lake City, UT 11111 E: cwse@fastmail.com P: 555-234-2345" \ "Professional Summary" \ "Experienced software engineer with a passion for developing innovative programs that expedite the efficiency and effectiveness of organizational success. Well-versed in technology and writing code to create systems that are reliable and user-friendly. Skilled leader who has the proven ability to motivate, educate, and manage a team of professionals to build software programs and effectively track changes. Confident communicator, strategic thinker, and innovative creator to develop software that is customized to meet a company’s organizational needs, highlight their core competencies, and further their success. " \ "Skills" \ "-Well-versed in software tools including HTML, JavaScript, CSS, BackBone and JQuery, among others. -Skilled at reading and writing code using viable inputs and outputs after accurate assessment of pre- and post-conditions. -Experienced at designing unit tests to measure the effectiveness of software programs, backend services, and user interfaces. -Confident problem-solving abilities to overcome glitches with creative solutions that are strategically designed to last long-term. -Strong communication skills and the ability to listen carefully to user feedback to determine modifications for optimal user-function." \ "Work Experience" \ "Software Engineer-April 2013 – present Rav Industries" \ "Developed and designed three critical software programs for financial tracking and reporting." \ "Optimized user effectiveness by creating a detailed feedback queue for users to discuss functionality, convenience, and effectiveness." \ "Oversee a team of four software developers and lead weekly discussions to brainstorm ideas in software development and to track changes made in existing programs." \ "Software Developer-February 2008 – April 2013 Brac Inc." \ "Participated in creating scalable systems for three primary departments, including human resources, marketing, and supply chain." \ "Ran monthly unit tests to determine software effectiveness and mend broken links or glitches in the system." \ "Gave quarterly reports to executive management regarding current developments, and tracked changes in existing software." \ "Education Internship2010-2011"\ "Estes Corp. Salt Lake City Utah Bachelor of Science 2010 in Computer Engineering 2010" \ "University of Utah Salt Lake City Utah" def preprocess_training_data1(): s = text.parsetree('The cat sat on the mat.', relations=True, lemmata=True) print(s) # result = es.parsetree('The cat sat on the mat.', relations=True, lemmata=True) # # s = en.parse('The cat sat on the mat.', relations=True, lemmata=True) # # # # print(s) # dircvs = [join(dir_cvs, f) for f in listdir(dir_cvs) if isfile(join(dir_cvs, f))] # alltext = ' ' # for cv in dircvs: # yd = read_All_CV(cv) # alltext += yd + " " # alltext = allText.lower() # vector = [] # for sentence in es.parsetree(alltext, tokenize=True, lemmata=True, tags=True): # temp = [] # for chunk in sentence.chunks: # for word in chunk.words: # if word.tag == 'NN' or word.tag == 'VB': # temp.append(word.lemma) # vector.append(temp) # global model # model = Word2Vec(vector, size=200, window=5, min_count=3, workers=4) # # model.save(dir_model_name) # # print("model:") # print(model)
[ "alexawl@bellsouth.net" ]
alexawl@bellsouth.net
0fa20eaa5571bbbc09e2573aff18620152ed29dc
8f8d426ee581545cd3c58f9879305f6536a02426
/extractFramesEverySecond.py
d6cc95318f833a855b945a48b44f163a61dd46f9
[]
no_license
tomtillo/video
bb0c0fabf93f804cf2e9bd7a6b2f781684a840d7
8be7eee7400fc3c5648e8eef90f14c667b7fc9c0
refs/heads/master
2020-06-11T20:32:44.092864
2019-06-27T11:35:43
2019-06-27T11:35:43
194,077,004
0
0
null
null
null
null
UTF-8
Python
false
false
1,176
py
# ---------------------------- Function to get frames every second ( and not all the frames) # Using tqdm to show progressbar # Usage :extract_frames_every_second(videos1,OUT_DIR) # Returns: times taken import cv2 import math import time def extract_frames_every_second(video_file, output_folder): import tqdm t1= time.time() cap = cv2.VideoCapture(video_file) duration_seconds= cap.get(7) / cap.get(5) pbar = tqdm.tqdm(total=duration_seconds) ret = 1 frame_number = 1 frameRate = cap.get(5) # frame rate while (ret): frame_number += 1 frameId = cap.get(1) # current frame number if (ret != True): break ret, frame = cap.read() if (frameId % math.floor(frameRate) == 0): pbar.update(1) out_k2 = cv2.imwrite(output_folder + "im_" + str(frame_number) + ".jpg", frame) cap.release() t2 = time.time() time_taken= t2-t1 return(time_taken) # ----------------------------------- End Function ------------ videos1 = "./videos/video1.mp4" OUT_DIR = "./output/" t1 = extract_frames_every_second(videos1,OUT_DIR) print("Time taken = %s " %(t1))
[ "noreply@github.com" ]
tomtillo.noreply@github.com
ba3177e820dd8a4793158eb218326d48229866ef
bb372428bb90fa80f2e87820b3c8c5ba305dcd4c
/python/bot/eups.py
0e4812155219126d211b7dcd779287ab6d1ce9ec
[]
no_license
TallJimbo/lsst-bot
7eb9b7a71a87a1ed416397c193931c80639bd746
0843afb2fdd5cc9ba62cf424a7dd73672b10e28f
refs/heads/master
2021-01-19T05:43:55.451321
2016-06-04T00:07:25
2016-06-04T00:07:25
60,484,354
0
0
null
null
null
null
UTF-8
Python
false
false
1,534
py
#!/usr/bin/env python from __future__ import absolute_import import eups.table import os import logging __all__ = "get_dependencies" def get_dependencies(config, path, pkg, recursive=False): """Return immediate dependencies from inspecting a table file. NOTE: recursive=True has not been tested. """ e = eups.Eups() t = eups.table.Table(os.path.join(path, "ups", pkg + ".table")) dependencies = t.dependencies(e, recursive=recursive) if recursive: dependencies.sort(key=lambda x: x[2]) for product, optional, depth in dependencies: yield product.name, optional def declare(config, path, pkg, version, tag_only=False): e = eups.Eups() if not tag_only: logging.debug("Declaring {pkg} {version}.".format(pkg=pkg, version=version)) e.declare(productName=pkg, versionName=version, productDir=path) for tmp in config.eups.tags: tag = tmp.format(eups=config.eups) logging.debug("Assigning tag {tag} to {pkg}.".format(pkg=pkg, tag=tag)) e.assignTag(tag, productName=pkg, versionName=version) def undeclare(config, pkg, version): e = eups.Eups() e.undeclare(productName=pkg, versionName=version) def setup(pkg, version, nodepend=False): e = eups.Eups(max_depth=(0 if nodepend else -1)) e.setup(productName=pkg, versionName=version) def tag(pkg, version, tag): e = eups.Eups() logging.debug("Assigning tag {tag} to {pkg}.".format(pkg=pkg, tag=tag)) e.assignTag(tag, productName=pkg, versionName=version)
[ "jbosch@astro.princeton.edu" ]
jbosch@astro.princeton.edu
b3aaed8088ac0f6dc6a87eed88943c27eea6fcb9
c9dd27f95f7a1a25963e0bd107a8fd72033c2168
/src/main.py
562ecf88d07a12c56bdfd8f9c24e342b1632865e
[]
no_license
RamazanDemirci/pythonML
3f807c5ec8e8881fe04197c75c98010b2feb7095
86d7e9286f03f099f64bf39615a9ee8841e3c021
refs/heads/main
2023-02-02T02:55:40.429272
2020-12-27T13:42:00
2020-12-27T13:42:00
319,462,234
0
0
null
null
null
null
UTF-8
Python
false
false
62
py
if __name__ == "__main__": print("Program Entry Point")
[ "radem18@gmail.com" ]
radem18@gmail.com
f5762a0a3bcd4f5ba6d65173ea4d994aece629be
1eb73b6d3801156620cad8707ab9fe70baca3e7c
/library/writer.py
edf9aa2fa31712f34a5cb3209693bb4a227c47d2
[ "MIT" ]
permissive
santiago26/vk4xmpp
732b75ef0a2a4c86d7d704da888e9be76953ec68
369f2550f2b422e6d27f06679ddda8ecd7cae026
refs/heads/master
2021-01-16T19:22:26.318690
2013-09-28T11:37:05
2013-09-28T11:37:05
null
0
0
null
null
null
null
UTF-8
Python
false
false
1,194
py
# /* encoding: utf-8 */ # © simpleApps, 2010 import os, sys, time, logging, traceback logger = logging.getLogger("vk4xmpp") fixme = lambda msg: Print("\n#! fixme: \"%s\"." % msg) lastErrorBody = None def wFile(filename, data, mode = "w"): with open(filename, mode, 0) as file: file.write(data) def rFile(filename): with open(filename, "r") as file: return file.read() def crashLog(name, text = 0, fixMe = True): global lastErrorBody logger.error("writing crashlog %s" % name) if fixMe: fixme(name) try: File = "crash/%s.txt" % name if not os.path.exists("crash"): os.makedirs("crash") exception = wException(True) if exception not in ("None", lastErrorBody): Timestamp = time.strftime("| %d.%m.%Y (%H:%M:%S) |\n") wFile(File, Timestamp + exception + "\n", "a") lastErrorBody = exception except: fixme("crashlog") wException() def Print(text, line = True): try: if line: print text else: sys.stdout.write(text) sys.stdout.flush() except (IOError, OSError): pass def wException(File = False): try: exception = traceback.format_exc().strip() if not File: Print(exception) return exception except (IOError, OSError): pass
[ "mrdoctorwho@gmail.com" ]
mrdoctorwho@gmail.com
924236479976a262f093bcadbe5e2db823bb810b
a6aada9cb93cc23bd17e4213c7a0b5114905a34a
/introduction/lab6.py
fed0db889f02668cb3f0910318040399a0aece3a
[]
no_license
SethKwashie/PythonLabs
90caf6a4fc5159536a892c84245b26bab4e48fbd
0a9578ce2f09fbec5ab0f1accb1e0d10f7792117
refs/heads/main
2023-01-04T12:57:54.497264
2020-11-02T14:46:18
2020-11-02T14:46:18
307,773,931
0
0
null
null
null
null
UTF-8
Python
false
false
268
py
age = int(input("Enter your age >>> ")) total = 0 start = 1 while start <= age: total += start start +=1 month = total * 12 days = total * 365 hours = total * 8760 print(f"{total} years ----> {month} months ----> {days} days ----> {hours} hours")
[ "seth.kwashie@turntabl.io" ]
seth.kwashie@turntabl.io
88a1436030cb0e0d641a755da900e1e5e1086035
d123e83108ffb222a2bceb57e4dcd5d7fd2b5b41
/api/controllers/jobscheduling/build/config.gypi
b4e5770be03aa67be34c86283b682796d7fe765f
[]
no_license
ajaxtream/smart2
0aede425b17f8666750427e4a89e3817995e43b5
7701b98d66f3da201e096eee43a0724c9a494dd1
refs/heads/master
2021-01-10T14:58:08.529403
2015-12-16T04:42:33
2015-12-16T04:42:33
48,087,021
0
0
null
null
null
null
UTF-8
Python
false
false
1,517
gypi
# Do not edit. File was generated by node-gyp's "configure" step { "target_defaults": { "cflags": [], "default_configuration": "Release", "defines": [], "include_dirs": [], "libraries": [] }, "variables": { "clang": 1, "host_arch": "x64", "icu_data_file": "icudt54l.dat", "icu_data_in": "../../deps/icu/source/data/in/icudt54l.dat", "icu_endianness": "l", "icu_gyp_path": "tools/icu/icu-generic.gyp", "icu_locales": "en,root", "icu_path": "./deps/icu", "icu_small": "true", "icu_ver_major": "54", "node_install_npm": "true", "node_prefix": "", "node_shared_cares": "false", "node_shared_http_parser": "false", "node_shared_libuv": "false", "node_shared_openssl": "false", "node_shared_v8": "false", "node_shared_zlib": "false", "node_tag": "", "node_use_dtrace": "true", "node_use_etw": "false", "node_use_mdb": "false", "node_use_openssl": "true", "node_use_perfctr": "false", "openssl_no_asm": 0, "python": "/usr/bin/python", "target_arch": "x64", "uv_library": "static_library", "uv_parent_path": "/deps/uv/", "uv_use_dtrace": "true", "v8_enable_gdbjit": 0, "v8_enable_i18n_support": 1, "v8_no_strict_aliasing": 1, "v8_optimized_debug": 0, "v8_random_seed": 0, "v8_use_snapshot": "false", "want_separate_host_toolset": 0, "nodedir": "/Users/stream/.node-gyp/0.12.7", "copy_dev_lib": "true", "standalone_static_library": 1 } }
[ "noEmail@anonymous.com" ]
noEmail@anonymous.com
389f4af608abd72b7016e5c50c3de3b32f56bf4e
394e0b00b35c61f7f9c2d34b11cb6aa29f6a1eb5
/Used_Others/CRT-Button/g.log.py
0946926cce66bb3bd59dc504aa9555cffe710af9
[]
no_license
congzheng-git/myProject
ecca28eafb129e014981b6fc74d4f8362ea3386b
da4cac92d87520d20036513616609bbffe9d9bef
refs/heads/master
2022-12-18T02:40:21.425117
2020-07-01T07:33:38
2020-07-01T07:33:38
207,499,868
0
0
null
2022-12-08T01:49:32
2019-09-10T08:03:20
Python
UTF-8
Python
false
false
311
py
# $language = "Python" # $interface = "1.0" import signal def main(): crt.Screen.Send('\x03') crt.Sleep(200) crt.Screen.Send('cd ../../../opt/jelly_current/logs/' + '\r') id = crt.Dialog.Prompt("输入要查的ID") crt.Screen.Send('tail -f out.log |grep ' + id + '\r') main()
[ "cczz1226@sina.com" ]
cczz1226@sina.com
31a96cf391d906b0d3d59fcd37437e16f21f474b
fd326562890d4f1987c384fc7c60374938231222
/OOP/ExamPrep/Exam10April21/project/decoration/ornament.py
90d7980c034613da08f4ee857bf726562ac89427
[]
no_license
miro-lp/SoftUni
cc3b0ff742218c9ceaf93f05c319ccfeed5bc8a4
283d9328537919de49f7f6a301e58593bae9ca2a
refs/heads/main
2023-08-23T21:22:07.856226
2021-08-25T15:10:18
2021-08-25T15:10:18
318,134,101
2
1
null
2021-08-10T12:51:54
2020-12-03T09:03:08
Python
UTF-8
Python
false
false
151
py
from project.decoration.base_decoration import BaseDecoration class Ornament(BaseDecoration): def __init__(self): super().__init__(1, 5)
[ "miro_lp@abv.bg" ]
miro_lp@abv.bg
4e6eca2f88c65a14f8f7950765320058fffc7784
6cd24d192fe83e2d4a23b2d7f2fe0c038940a5d9
/trip/models.py
c05e3e881cdb57c13e1a086d6c7b5744615c8a64
[]
no_license
nabeelakhtar20/trip_app_sample
d5370864ae0c872b0bc24bd9c47361c2fcae413c
ae6ab820d9a39fa4072267f09349b2c0d794b979
refs/heads/master
2022-05-18T03:30:03.315671
2019-10-13T12:06:58
2019-10-13T12:06:58
214,809,376
1
0
null
2022-04-22T22:32:11
2019-10-13T11:38:27
Python
UTF-8
Python
false
false
1,491
py
from datetime import datetime from neomodel import StringProperty, StructuredNode, RelationshipTo, \ IntegerProperty, DateTimeProperty, UniqueIdProperty, JSONProperty, DateProperty from auth.models import User class Trip(StructuredNode): uid = UniqueIdProperty() destination = StringProperty() start_date = DateProperty() end_date = DateProperty() adults = IntegerProperty() infants = IntegerProperty() estimated_budget_start = IntegerProperty() estimated_budget_end = IntegerProperty() events = JSONProperty() creation_date = DateTimeProperty(default=datetime.now()) last_updated = DateTimeProperty(default=datetime.now()) user = RelationshipTo(User, 'PLANNED_BY') @property def serialize(self): return { 'node_properties': { "id": self.uid, "destination": self.destination, "start_date": self.start_date, "end_date": self.end_date, "adults": self.adults, "infants": self.infants, "estimated_budget_start": self.estimated_budget_start, "estimated_budget_end": self.estimated_budget_end, "events": self.events, }, } @property def serialize_connections(self): return [ { 'nodes_type': 'User', 'nodes_related': self.serialize_relationships(self.user.all()), }, ]
[ "nabeel_akhtar20@hotmail.com" ]
nabeel_akhtar20@hotmail.com
7dc8d99db40e13c4050036693a90c23d2732f5d1
6af90d24f1f081c1f1046346dc1c7cdb98bcaddb
/tests/django/run_unit_tests.py
9270e1793369a22bae93f3dd891a9bc5c60bb34c
[ "Apache-2.0" ]
permissive
aolabNeuro/brain-python-interface
e4d8fd4682ed63611d44f83e7889f35a71a80b59
bc215d97fe95765c2bf40d1be793d6ffa83586d1
refs/heads/master
2023-08-18T09:53:24.212536
2023-08-17T21:22:01
2023-08-17T21:22:01
250,614,460
3
3
NOASSERTION
2023-09-08T16:44:08
2020-03-27T18:31:13
HTML
UTF-8
Python
false
false
240
py
from django.test.utils import get_runner from django.conf import settings from db.boot_django import boot_django boot_django() TestRunner = get_runner(settings) test_runner = TestRunner() failures = test_runner.run_tests(["tests/django"])
[ "leo.scholl@gmail.com" ]
leo.scholl@gmail.com
70f76d0921495935a8af0e489fc2de27af2fdd67
0fb2607ba5dff6c86e806dd67ba16b158c68b7a3
/bellorest/takeaway/apps.py
8530766b1735e6d11564a7d76bae23c34f085d6f
[]
no_license
shashankiit/Restaraunt-Mangement-System
204259b4b7c0dbd984f4d38dcdbbab39bef2ee02
8a246ff56023a04c996e7fcf0ffb7d9093de1446
refs/heads/main
2023-04-23T06:14:30.348021
2021-05-07T18:14:15
2021-05-07T18:14:15
358,109,274
0
0
null
null
null
null
UTF-8
Python
false
false
148
py
from django.apps import AppConfig class TakeawayConfig(AppConfig): default_auto_field = 'django.db.models.BigAutoField' name = 'takeaway'
[ "ambeloskar@iitb.ac.in" ]
ambeloskar@iitb.ac.in
fad1fce49448f8fa4f4b57f21e6b48648aff90b7
282234bf6467224f769b67c923b111add740c775
/Course_2_Python_Data_Structures/ex_06_05/ex_06_05.py
82c9334e74a4ad3f1c8c105755813523c36f1ccb
[]
no_license
gravitonGB/python_for_everybody
73d0767db1cf0c705144e287fba3fe66bc42ec43
8b3d197247bf52c66551867f2851d3006f7da484
refs/heads/master
2020-03-09T11:45:37.238749
2018-04-09T12:35:03
2018-04-09T12:35:03
128,765,513
0
0
null
null
null
null
UTF-8
Python
false
false
197
py
#Exercise 6.5 #print("Excercise 6.5") text = "X-DSPAM-Confidence: 0.8475"; pos = text.find(':') t2 = text[pos+1:len(text)] last_text = t2.strip() last_value = float(last_text) print(last_value)
[ "gokhanbegce@gmail.com" ]
gokhanbegce@gmail.com
2725285e31284db439a824eacbd0d0ddf6681c04
7a6d30770cd56c2900aa7ef969b3ecfd679078a5
/WebGame/WebGame/game/classes/postgres.py
1dc642d5a3d31b85d596a939756c32fbda900e56
[]
no_license
sherkd/zombiegame
519e1f7166e8e4ca192347682b2e55675757dff5
1438d2267ab2c615e14cf6e5d13525b38f7cb7a1
refs/heads/master
2020-06-14T12:54:40.043993
2017-01-05T12:24:31
2017-01-05T12:24:31
75,178,354
0
0
null
null
null
null
UTF-8
Python
false
false
2,948
py
import psycopg2 import sys class Postgres(object): def __init__(self): conn_string = "host='localhost' dbname='zombiegamers' user='postgres' password='project3'" print("Connecting to database\n ->%s" % (conn_string)) self.connection = psycopg2.connect(conn_string) self.cursor = self.connection.cursor() def getConnection(self): return self.connection def createDatabase(self): self.cursor.execute("CREATE TABLE Account(userid int NOT NULL, password varchar(255), username varchar(255), gender varchar(1), email varchar(255), birthday date, PRIMARY KEY(userid))") self.cursor.execute("CREATE TABLE Player(userid int NOT NULL, health int, attack int , luck int, accuracy int, speed int, skillpoints int, weaponid int, FOREIGN KEY(userid) REFERENCES Account(userid))") self.cursor.execute("CREATE TABLE Weapon(weaponid int NOT NULL, name varchar(255), class varchar(255), description varchar(255), level int, damage int, PRIMARY KEY(weaponid))") self.cursor.execute("CREATE TABLE Player_Weapon(userid int NOT NULL, weaponid int NOT NULL, FOREIGN KEY(userid) REFERENCES Account(userid), FOREIGN KEY(weaponid) REFERENCES Weapon(weaponid))") def getAccount(self): self.cursor.execute("SELECT * FROM Account WHERE userid='12'") return self.cursor.fetchone() def insertAccount(self): try: self.cursor.execute("INSERT INTO Account (userid, password, username, gender, email, birthday) Values(12, 'pass', 'user', 'm', 'user@gmail.com', '2000-10-10')") self.connection.commit() except: self.connection.rollback() def insertTestPlayer(self): try: self.cursor.execute("DELETE FROM Player WHERE userid = '12'") self.connection.commit() except: self.connection.rollback() try: self.cursor.execute("INSERT INTO Player(userid, health, attack, luck, accuracy, speed, skillpoints, weaponid) VALUES(12, 100, 10, 5, 10, 10, 0, 12);") self.connection.commit() except: self.connection.rollback() def getTestPlayer(self): self.cursor.execute("SELECT * FROM Player WHERE userid='12'") return self.cursor.fetchone() def insertWeapon(self, weapon): id = random.randint(0, 1000) try: self.cursor.execute("INSERT INTO Weapon (userid, name, class, description, level, damage) VALUES (" + str(id) + "," + weapon.getName() + "," + weapon.getClass() + "," + weapon.getDescription() + "," + str(weapon.getLevel()) + "," + str(weapon.getDamage()) + ")") self.connection.commit() except: self.connection.rollback() def getWeapon(self): print(self.cursor.execute("SELECT * FROM Weapon WHERE userid='12'")) def getWeapon(self, id): pass def getWeapons(self, username): pass
[ "sivarwerrie@hotmail.nl" ]
sivarwerrie@hotmail.nl
fbcf2fca48d9207fd6d531d83f43f44da2312148
0120813c649236fcb4732723c4b25f6538de04fb
/Image Stitching/Source/main.py
a232db81366c8c7dc347f44a463ca2d9345047ab
[]
no_license
shubh0906/Computer-Vision
47f1e8e55f54138acd070f5f39b722b17a5747b2
e83bd827f1ed9de9218af5e973e69510d826d100
refs/heads/master
2021-01-25T11:15:36.124527
2018-03-06T09:01:35
2018-03-06T09:01:35
123,387,519
0
0
null
null
null
null
UTF-8
Python
false
false
12,840
py
# Instructions: # Do not change the output file names, use the helper functions as you see fit import os import sys import cv2 import numpy as np import matplotlib.pyplot as plt import math def help_message(): print("Usage: [Question_Number] [Input_Options] [Output_Options]") print("[Question Number]") print("1 Perspective warping") print("2 Cylindrical warping") print("3 Bonus perspective warping") print("4 Bonus cylindrical warping") print("[Input_Options]") print("Path to the input images") print("[Output_Options]") print("Output directory") print("Example usages:") print(sys.argv[0] + " 1 " + "[path to input image1] " + "[path to input image2] " + "[path to input image3] " + "[output directory]") ''' Detect, extract and match features between img1 and img2. Using SIFT as the detector/extractor, but this is inconsequential to the user. Returns: (pts1, pts2), where ptsN are points on image N. The lists are "aligned", i.e. point i in pts1 matches with point i in pts2. Usage example: im1 = cv2.imread("image1.jpg", 0) im2 = cv2.imread("image2.jpg", 0) (pts1, pts2) = feature_matching(im1, im2) plt.subplot(121) plt.imshow(im1) plt.scatter(pts1[:,:,0],pts1[:,:,1], 0.5, c='r', marker='x') plt.subplot(122) plt.imshow(im2) plt.scatter(pts1[:,:,0],pts1[:,:,1], 0.5, c='r', marker='x') ''' def feature_matching(img1, img2, savefig=False): # Initiate SIFT detector sift = cv2.xfeatures2d.SIFT_create() # find the keypoints and descriptors with SIFT kp1, des1 = sift.detectAndCompute(img1,None) kp2, des2 = sift.detectAndCompute(img2,None) # FLANN parameters FLANN_INDEX_KDTREE = 1 index_params = dict(algorithm = FLANN_INDEX_KDTREE, trees = 5) search_params = dict(checks=50) # or pass empty dictionary flann = cv2.FlannBasedMatcher(index_params,search_params) matches2to1 = flann.knnMatch(des2,des1,k=2) matchesMask_ratio = [[0,0] for i in xrange(len(matches2to1))] match_dict = {} for i,(m,n) in enumerate(matches2to1): if m.distance < 0.7*n.distance: matchesMask_ratio[i]=[1,0] match_dict[m.trainIdx] = m.queryIdx good = [] recip_matches = flann.knnMatch(des1,des2,k=2) matchesMask_ratio_recip = [[0,0] for i in xrange(len(recip_matches))] for i,(m,n) in enumerate(recip_matches): if m.distance < 0.7*n.distance: # ratio if m.queryIdx in match_dict and match_dict[m.queryIdx] == m.trainIdx: #reciprocal good.append(m) matchesMask_ratio_recip[i]=[1,0] if savefig: draw_params = dict(matchColor = (0,255,0), singlePointColor = (255,0,0), matchesMask = matchesMask_ratio_recip, flags = 0) img3 = cv2.drawMatchesKnn(img1,kp1,img2,kp2,recip_matches,None,**draw_params) plt.figure(),plt.xticks([]),plt.yticks([]) plt.imshow(img3,) plt.savefig("feature_matching.png",bbox_inches='tight') return ([ kp1[m.queryIdx].pt for m in good ],[ kp2[m.trainIdx].pt for m in good ]) ''' Warp an image from cartesian coordinates (x, y) into cylindrical coordinates (theta, h) Returns: (image, mask) Mask is [0,255], and has 255s wherever the cylindrical images has a valid value. Masks are useful for stitching Usage example: im = cv2.imread("myimage.jpg",0) #grayscale h,w = im.shape f = 700 K = np.array([[f, 0, w/2], [0, f, h/2], [0, 0, 1]]) # mock calibration matrix imcyl = cylindricalWarpImage(im, K) ''' def cylindricalWarpImage(img1, K, savefig=False): f = K[0,0] im_h,im_w = img1.shape # go inverse from cylindrical coord to the image # (this way there are no gaps) cyl = np.zeros_like(img1) cyl_mask = np.zeros_like(img1) cyl_h,cyl_w = cyl.shape x_c = float(cyl_w) / 2.0 y_c = float(cyl_h) / 2.0 for x_cyl in np.arange(0,cyl_w): for y_cyl in np.arange(0,cyl_h): theta = (x_cyl - x_c) / f h = (y_cyl - y_c) / f X = np.array([math.sin(theta), h, math.cos(theta)]) X = np.dot(K,X) x_im = X[0] / X[2] if x_im < 0 or x_im >= im_w: continue y_im = X[1] / X[2] if y_im < 0 or y_im >= im_h: continue cyl[int(y_cyl),int(x_cyl)] = img1[int(y_im),int(x_im)] cyl_mask[int(y_cyl),int(x_cyl)] = 255 if savefig: plt.imshow(cyl, cmap='gray') plt.savefig("cyl.png",bbox_inches='tight') return (cyl,cyl_mask) ''' Calculate the geometric transform (only affine or homography) between two images, based on feature matching and alignment with a robust estimator (RANSAC). Returns: (M, pts1, pts2, mask) Where: M is the 3x3 transform matrix pts1 are the matched feature points in image 1 pts2 are the matched feature points in image 2 mask is a binary mask over the lists of points that selects the transformation inliers Usage example: im1 = cv2.imread("image1.jpg", 0) im2 = cv2.imread("image2.jpg", 0) (M, pts1, pts2, mask) = getTransform(im1, im2) # for example: transform im1 to im2's plane # first, make some room around im2 im2 = cv2.copyMakeBorder(im2,200,200,500,500, cv2.BORDER_CONSTANT) # then transform im1 with the 3x3 transformation matrix out = cv2.warpPerspective(im1, M, (im1.shape[1],im2.shape[0]), dst=im2.copy(), borderMode=cv2.BORDER_TRANSPARENT) plt.imshow(out, cmap='gray') plt.show() ''' def getTransform(src, dst, method='affine'): pts1,pts2 = feature_matching(src,dst) src_pts = np.float32(pts1).reshape(-1,1,2) dst_pts = np.float32(pts2).reshape(-1,1,2) if method == 'affine': M, mask = cv2.estimateAffine2D(src_pts, dst_pts, cv2.RANSAC, ransacReprojThreshold=5.0) #M = np.append(M, [[0,0,1]], axis=0) if method == 'homography': M, mask = cv2.findHomography(src_pts, dst_pts, cv2.RANSAC, 5.0) matchesMask = mask.ravel().tolist() return (M, pts1, pts2, mask) # =================================================== # ================ Perspective Warping ============== # =================================================== def Perspective_warping(im1, im2, im3): im1 = cv2.copyMakeBorder(im1,200,200,500,500, cv2.BORDER_CONSTANT) (M, pts1, pts2, mask) = getTransform(im2, im1, 'homography') out = cv2.warpPerspective(im2, M, (im1.shape[1],im1.shape[0]), dst=im1.copy(), borderMode=cv2.BORDER_TRANSPARENT) (M, pts1, pts2, mask) = getTransform(im3, out, 'homography') out = cv2.warpPerspective(im3, M, (out.shape[1],out.shape[0]), dst=out.copy(), borderMode=cv2.BORDER_TRANSPARENT) output_image = out # This is dummy output, change it to your output # Write out the result output_name = sys.argv[5] + "output_homography.png" cv2.imwrite(output_name, output_image) imM = cv2.imread('example_output1.png', 0); #print '---q1---', RMSD(1, out, imM); return True def Bonus_perspective_warping(img1, img2, img3): # Write your codes here output_image = img1 # This is dummy output, change it to your output # Write out the result output_name = sys.argv[5] + "output_homography_lpb.png" cv2.imwrite(output_name, output_image) return True # =================================================== # =============== Cynlindrical Warping ============== # =================================================== def Cylindrical_warping(im1, im2, im3): f = 500 h,w = im1.shape K = np.array([[f, 0, w/2], [0, f, h/2], [0, 0, 1]]) # mock calibration matrix im1, k1 = cylindricalWarpImage(im1, K) h,w = im2.shape K = np.array([[f, 0, w/2], [0, f, h/2], [0, 0, 1]]) # mock calibration matrix im2, k2 = cylindricalWarpImage(im2, K) h,w = im3.shape K = np.array([[f, 0, w/2], [0, f, h/2], [0, 0, 1]]) # mock calibration matrix im3, k3 = cylindricalWarpImage(im3, K) im1 = cv2.copyMakeBorder(im1, 50, 50, 300, 300, cv2.BORDER_CONSTANT) (M1, pts1, pts2, mask) = getTransform(im2, im1) (M2, pts1, pts2, mask) = getTransform(im3, im1) out1 = cv2.warpAffine(im2, M1, (im1.shape[1],im1.shape[0])) outM1 = cv2.warpAffine(k2, M1, (im1.shape[1],im1.shape[0])) r,c = im1.shape for i in xrange(r): for j in xrange(c): if outM1[i,j] == 255: im1[i, j] = out1[i, j] (M, pts1, pts2, mask) = getTransform(im3, im1) out = cv2.warpAffine(im3, M2, (out1.shape[1],out1.shape[0])) outM = cv2.warpAffine(k3, M2, (out1.shape[1],out1.shape[0])) r,c = im1.shape for i in xrange(r): for j in xrange(c): if outM[i,j] == 255: im1[i, j] = out[i, j] output_image = im1 # This is dummy output, change it to your output # Write out the result output_name = sys.argv[5] + "output_cylindrical.png" cv2.imwrite(output_name, output_image) imM = cv2.imread('example_output2.png', 0); #print RMSD(2, im1, imM); return True '''# Write your codes here output_image = img1 # This is dummy output, change it to your output # Write out the result output_name = sys.argv[5] + "output_cylindrical.png" cv2.imwrite(output_name, output_image) return True''' def Bonus_cylindrical_warping(img1, img2, img3): # Write your codes here output_image = img1 # This is dummy output, change it to your output # Write out the result output_name = sys.argv[5] + "output_cylindrical_lpb.png" cv2.imwrite(output_name, output_image) return True ''' This exact function will be used to evaluate your results for HW2 Compare your result with master image and get the difference, the grading criteria is posted on Piazza ''' '''def RMSD(target, master): # Get width, height, and number of channels of the master image master_height, master_width = master.shape[:2] master_channel = len(master.shape) # Get width, height, and number of channels of the target image target_height, target_width = target.shape[:2] target_channel = len(target.shape) # Validate the height, width and channels of the input image if (master_height != target_height or master_width != target_width or master_channel != target_channel): return -1 else: total_diff = 0.0; master_channels = cv2.split(master); target_channels = cv2.split(target); for i in range(0, len(master_channels), 1): dst = cv2.absdiff(master_channels[i], target_channels[i]) dst = cv2.pow(dst, 2) mean = cv2.mean(dst) total_diff = total_diff + mean[0]**(1/2.0) return total_diff;''' def RMSD(questionID, target, master): # Get width, height, and number of channels of the master image master_height, master_width = master.shape[:2] master_channel = len(master.shape) # Get width, height, and number of channels of the target image target_height, target_width = target.shape[:2] target_channel = len(target.shape) # Validate the height, width and channels of the input image if (master_height != target_height or master_width != target_width or master_channel != target_channel): return -1 else: nonZero_target = cv2.countNonZero(target) nonZero_master = cv2.countNonZero(master) if (questionID == 1): if (nonZero_target < 1200000): return -1 elif(questionID == 2): if (nonZero_target < 700000): return -1 else: return -1 total_diff = 0.0; master_channels = cv2.split(master); target_channels = cv2.split(target); for i in range(0, len(master_channels), 1): dst = cv2.absdiff(master_channels[i], target_channels[i]) dst = cv2.pow(dst, 2) mean = cv2.mean(dst) total_diff = total_diff + mean[0]**(1/2.0) return total_diff; if __name__ == '__main__': question_number = -1 # Validate the input arguments if (len(sys.argv) != 6): help_message() sys.exit() else: question_number = int(sys.argv[1]) if (question_number > 4 or question_number < 1): print("Input parameters out of bound ...") sys.exit() input_image1 = cv2.imread(sys.argv[2], 0) input_image2 = cv2.imread(sys.argv[3], 0) input_image3 = cv2.imread(sys.argv[4], 0) function_launch = { 1 : Perspective_warping(input_image1, input_image2, input_image3), 2 : Cylindrical_warping(input_image1, input_image2, input_image3), 3 : Bonus_perspective_warping(input_image1, input_image2, input_image3), 4 : Bonus_cylindrical_warping(input_image1, input_image2, input_image3), } # Call the function function_launch[question_number]()
[ "noreply@github.com" ]
shubh0906.noreply@github.com
efc58d0b9b812e8507cfe77bd976ff7d1d82c672
342b0835ad165d8fcb647792e13440ed8d0579c0
/posts/views.py
e7bd842707ebd9559bd68fbdda4e6b4983a21463
[]
no_license
Str8jckn/mb
a2d1df745840a7654bef179e4088b2964f0608bc
d64fd91397eb979ca85e7cabefe7c290c04434e2
refs/heads/master
2023-02-23T11:37:24.684042
2021-01-25T05:10:35
2021-01-25T05:10:35
332,100,152
0
0
null
null
null
null
UTF-8
Python
false
false
421
py
from django.shortcuts import render from django.views.generic import ListView from .models import Post from django.contrib import admin from django.urls import path, include class HomePageView(ListView): model = Post template_name = 'home.html' context_object_name = 'all_posts_list' urlpatterns = [ path('admin/', admin.site.urls), path('', include('posts.urls')), # new ] # Create your views here.
[ "jaebrownjr@gmail.com" ]
jaebrownjr@gmail.com
fb935ebf7929bf8547c110f220afd9dd747ddc54
a90792aec007ab37cdd7d21dfb5340a88b87132f
/shorten/views.py
41aca3353713eacb4762171447c1d714b33da29e
[]
no_license
chetangargnitd/url_shortener
fee4b837c79b118bf1cfea3f80582c05326a0c3d
b94c1f83ac56e623fedb3e2ca211a9f2f5b35ff2
refs/heads/master
2021-06-24T04:20:27.291801
2019-06-17T14:22:18
2019-06-17T14:22:18
191,707,083
0
0
null
null
null
null
UTF-8
Python
false
false
1,451
py
from __future__ import unicode_literals from django.shortcuts import render from django.shortcuts import redirect from django.http import HttpResponse from django.http import Http404 from django.shortcuts import get_object_or_404 from .models import URLs import urllib import hashlib def home(request): context = URLs.objects.order_by('-created')[:5] return render(request, 'shorten/index.html', {'context' : context}) def shrink(request): url = (request.GET["url"]) c_url = (request.GET["c_url"]) print(type(c_url)) encoded_url = url.encode('utf-8') hashObject = hashlib.md5(encoded_url) shrinked_url = hashObject.hexdigest()[:8] context = URLs.objects.order_by('-created')[:5] if(c_url == ""): try: check = URLs.objects.get(shrinked_url = shrinked_url) except URLs.DoesNotExist: entry = URLs(shrinked_url = shrinked_url, original_url = url) entry.save() return render(request, 'shorten/index.html', {'shrinked_url' : shrinked_url, 'context' : context}) else: try: check = URLs.objects.get(shrinked_url = c_url) except URLs.DoesNotExist: entry = URLs(shrinked_url = c_url, original_url = url) entry.save() return render(request, 'shorten/index.html', {'shrinked_url' : c_url, 'context' : context}) def retrieve(request, id): target = get_object_or_404(URLs, shrinked_url = id) targetURL = target.original_url if(targetURL[:4] != 'http'): targetURL = 'http://'+targetURL return redirect(targetURL)
[ "chetangarg1102@gmail.com" ]
chetangarg1102@gmail.com
aee5135a64857c7423190a714cc62b15b207d49f
9612da0b14b7e9f883a2ae50f00af87571bccabc
/Analyzer.py
2f1b4c07946a7828146daa3d16607d8aad4e94ea
[]
no_license
andrewghaly/HandwrittenEstimation
e56bb5489b5833a515dff529e0f9172ed7c380db
97424d64a4b3ae784499d0bcf660797f056fc026
refs/heads/master
2021-01-23T03:27:17.421327
2017-03-24T14:45:44
2017-03-24T14:45:44
86,078,278
1
0
null
null
null
null
UTF-8
Python
false
false
1,685
py
import cv2 import numpy as np drawing_averages = [] for imageNumber in range(1,51): img = cv2.imread('C:\\Users\\ghalya\\Pictures\\Hands\\Saunders_hand\\al_r_' + str(imageNumber) + '.png', 0) if img is None: print "Error loading image" exit() newImg = cv2.cvtColor(img, cv2.COLOR_GRAY2RGB) #cv2.imshow('Image', img) # define the list of boundaries boundaries = [ ([151], [255]), ] # loop over the boundaries for (lower, upper) in boundaries: # create NumPy arrays from the boundaries lower = np.array(lower, dtype="uint8") upper = np.array(upper, dtype="uint8") #apply the mask mask = cv2.inRange(img, lower, upper) output = cv2.bitwise_and(img, img, mask=mask) # show the images #cv2.imshow("images", np.hstack([img, output])) #cv2.imwrite('C:\Users\ghalya\Pictures\lol_testlulz.png', np.hstack([img, output])) height, width = img.shape points = 0 xSum = 0 ySum = 0 for i in range(0, width): for j in range(0, height): if img[j][i] <= 150: points += 1 xSum += i ySum += j xAvg = xSum/points yAvg = ySum/points drawing_averages.append([xAvg, yAvg]) #print("xAvg: ", xAvg, " yAvg: ", yAvg) cv2.circle(newImg, (xAvg, yAvg), 5, (0, 0, 255), -1) #cv2.imshow("image", newImg) #cv2.imwrite("C:\Users\ghalya\Pictures\genLEL.png", newImg) cv2.waitKey(0) print drawing_averages count = xTotal = yTotal = 0 for i, j in drawing_averages: xTotal += i yTotal += j count += 1 print "average:", xTotal/count, yTotal/count,
[ "ghalya@wit.edu" ]
ghalya@wit.edu
588948095f2db1f4d431c649e77a76b72ecf54b8
68f57fd1dd274be72af6d85762b67bbf8d2ef6d6
/tests/test_cosine.py
3ac719652f889a7529befb8bcbf87a328c003cfa
[]
no_license
afcarl/simplecosine
287cbf40ef8aa2251ea538b7b3c2d28c5b6f2488
1ba869198ab3211dd4b0412e80e670308007f687
refs/heads/master
2020-03-17T23:56:28.854494
2017-06-15T21:33:36
2017-06-15T21:33:36
134,069,251
1
0
null
2018-05-19T14:29:05
2018-05-19T14:29:05
null
UTF-8
Python
false
false
2,909
py
import unittest from simplecosine.cosine import CosineSetSimilarity, CosineTextSimilarity import numpy import pickle class TestSetCosineClass(unittest.TestCase): def setUp(self): self.ilist = [('a', 'b', 'c'), ['b', 'c', 'd k'], ('d k', 'e', 'f') ] def test_cosine(self): cosine = CosineSetSimilarity(self.ilist) s1 = self.ilist[0] s2 = self.ilist[1] cosine_sim = cosine(s1, s2) self.assertAlmostEqual(cosine_sim, 0.378, places=3) cosine_sim = cosine(('g', 'h', 'd k', 'd k'), s2) self.assertAlmostEqual(cosine_sim, 0.267, places=3) def test_cosine_na(self): cosine = CosineSetSimilarity(self.ilist) cosine_sim = cosine(self.ilist[0], ()) assert numpy.isnan(cosine_sim) def test_cosine_identical(self): cosine = CosineSetSimilarity(self.ilist) cosine_sim = cosine(self.ilist[0], self.ilist[0]) self.assertAlmostEqual(cosine_sim, 1, places=5) def test_cosine_cache(self): cosine = CosineSetSimilarity(self.ilist) s1 = self.ilist[0] s2 = self.ilist[1] cosine_sim = cosine(s1, s2) self.assertAlmostEqual(cosine_sim, 0.378, places=3) cosine_sim = cosine(s1, s2) self.assertAlmostEqual(cosine_sim, 0.378, places=3) def test_cosine_no_corpus(self): cosine = CosineSetSimilarity([]) s1 = self.ilist[0] s2 = self.ilist[1] cosine_sim = cosine(s1, s2) self.assertAlmostEqual(cosine_sim, 0.667, places=3) cosine_sim = cosine(('g', 'h', 'd k'), s2) self.assertAlmostEqual(cosine_sim, 0.333, places=3) def test_cosine_pickle(self) : cosine = CosineSetSimilarity(self.ilist) s1 = self.ilist[0] s2 = self.ilist[1] cosine_sim = cosine(s1, s2) pickle.dumps(cosine) cosine = CosineSetSimilarity([]) s1 = self.ilist[0] s2 = self.ilist[1] cosine_sim = cosine(s1, s2) pickle.dumps(cosine) class TestTextCosineClass(unittest.TestCase): def setUp(self): self.ilist = ['a b c', 'b c d', 'd e f'] def test_cosine(self): cosine = CosineTextSimilarity(self.ilist) s1 = self.ilist[0] s2 = self.ilist[1] cosine_sim = cosine(s1, s2) self.assertAlmostEqual(cosine_sim, 0.378, places=3) def test_cosine_na(self): cosine = CosineTextSimilarity(self.ilist) cosine_sim = cosine(self.ilist[0], '') assert numpy.isnan(cosine_sim) def test_cosine_identical(self): cosine = CosineTextSimilarity(self.ilist) cosine_sim = cosine(self.ilist[0], self.ilist[0]) self.assertAlmostEqual(cosine_sim, 1, places=5) if __name__ == '__main__': unittest.main()
[ "fgregg@uchicago.edu" ]
fgregg@uchicago.edu
362e1c5d6c3bd62bae910a57c6d2f267d85924f0
6cf9b92459bb6bc5d9941728f9b0d697bd5db29f
/task.py
121bdbcc4864c0ce62df6fde1bfd892db04b2cf0
[]
no_license
mehar13hamza13/competetive-learning-using-python
6eb2da2ab6ccdc17bc47e432a27e5582ddd5186c
f62f079c1facd7536c51fb4dad3e8edf867482c3
refs/heads/master
2023-06-24T22:22:41.173546
2021-07-27T15:55:52
2021-07-27T15:55:52
390,043,235
0
0
null
null
null
null
UTF-8
Python
false
false
3,303
py
import numpy as np class CompetetiveLearning(): def __init__(self, input_nodes, output_nodes): self.input_nodes = input_nodes self.output_nodes = output_nodes #initializing random weights self.weights = np.random.uniform(low=0.0, high=1.0, size=(output_nodes, input_nodes)).round(1) def train(self, train_x): print("----training for "+str(len(train_x))+" samples------") clustering = {"A": [], "B": [], "C": []} count = 1 for i in train_x: print("Iteration "+str(count)) x = i.reshape((6, 1)) # reshaping the ith input value so matrix multiplication can be applied result = np.matmul(self.weights, x) #multiplying wieghts with input nodes (w11X1 + w21X2 + ....) winning_unit = result.argmax() # index with maximum value will be the winning unit (only row with these weights will be updated) print("Output Values for Iteration "+str(count)+": ") print(result) print("Winning Unit: "+str(winning_unit+1)) print("Adjusting the weight for only row "+str(winning_unit+1)) self.adjust_weights(0.5, winning_unit, x) clustering[list(clustering.keys())[winning_unit]].append("R"+str(count)) count+=1 self.print_final_weights() print("\nFinal Cluster Results: ") print(clustering) def print_final_weights(self): print("\nFinal Weights for Output P: ") print(self.weights[0]) print("Final Weights for Output Q: ") print(self.weights[1]) print("Final Weights for Output R: ") print(self.weights[2]) def adjust_weights(self, learning_rate, row_no, inputs): for i in range(len(self.weights[row_no])): #adjusting the weights self.weights[row_no][i] = self.weights[row_no][i] + learning_rate*inputs[i] #normalizing the weights self.weights[row_no][i]/=2 def test(self, test_x): print() print("----testing for " + str(len(test_x)) + " samples------") print() count = 1 classes = ["Class A", "Class B", "Class C"] for i in test_x: print("Iteration " + str(count)) x = i.reshape((6, 1)) # reshaping the ith input value so matrix multiplication can be applied result = np.matmul(self.weights, x) # multiplying wieghts with input nodes (w11X1 + w21X2 + ....) winning_unit = result.argmax() # index with maximum value will be the winning unit (only row with these weights will be updated) print("Output Values for t" + str(count) + ": ") print(result) print("Winning Unit: " + str(winning_unit + 1)) print("t"+str(count)+" belongs to "+classes[winning_unit]) count += 1 cl = CompetetiveLearning(6, 3) train_x = np.array([ [1, 0, 0, 0, 0, 0], [1, 0, 0, 1, 0, 0], [1, 0, 0, 0, 1, 0], [0, 0, 1, 1, 0, 1], [0, 0, 1, 1, 0, 1], [0, 0, 1, 0, 0, 1], [0, 1, 0, 0, 0, 1], [0, 1, 0, 0, 0, 1], [0, 1, 0, 0, 1, 1], [1, 0, 0, 0, 0, 0] ]) test_x = np.array([ [0, 0, 1, 1, 1, 1], [1, 0, 0, 0, 1, 1], [0, 1, 0, 1, 0, 1] ]) cl.train(train_x) cl.test(test_x)
[ "hassaanfarooq71@gmail.com" ]
hassaanfarooq71@gmail.com
71099406ad6fc49075c14031a16a80c20794895f
a5315e8edc48c5fabcf6aaaa56de737d9594cddf
/deepc/modules/resnet.py
6e109cf6e94138f8dae8fb9e4efd2080ec7f1fd4
[]
no_license
elirshabat/deepc-pytorch
efead7622d09c0c14f2b2c941cc7202e0278bbc8
c946ad55c5a8fedff1055f34b92b1a8e9cb3df2a
refs/heads/master
2020-03-23T01:41:05.212437
2018-11-06T07:31:50
2018-11-06T07:31:50
140,931,467
0
0
null
null
null
null
UTF-8
Python
false
false
4,955
py
import torch from torchvision import models import torch.nn.init as init import numpy as np def initialize_weights(method='kaiming', *models): for model in models: for module in model.modules(): if isinstance(module, torch.nn.Conv2d) or isinstance(module, torch.nn.ConvTranspose2d) or isinstance(module, torch.nn.Linear): if method == 'kaiming': init.kaiming_normal(module.weight.data, np.sqrt(2.0)) elif method == 'xavier': init.xavier_normal(module.weight.data, np.sqrt(2.0)) elif method == 'orthogonal': init.orthogonal(module.weight.data, np.sqrt(2.0)) elif method == 'normal': init.normal(module.weight.data,mean=0, std=0.02) if module.bias is not None: init.constant_(module.bias.data,0) class GlobalConvolutionBlock(torch.nn.Module): def __init__(self, in_channels, out_channels, k): super().__init__() super(GlobalConvolutionBlock, self).__init__() self.left = torch.nn.Sequential(torch.nn.Conv2d(in_channels, out_channels, kernel_size=(k[0],1), padding=(k[0]//2,0)), torch.nn.Conv2d(out_channels, out_channels, kernel_size=(1,k[1]), padding=(0,k[1]//2))) self.right = torch.nn.Sequential(torch.nn.Conv2d(in_channels, out_channels, kernel_size=(1,k[1]), padding=(0,k[1]//2)), torch.nn.Conv2d(out_channels, out_channels, kernel_size=(k[0],1), padding=(k[0]//2,0))) def forward(self, x): left = self.left(x) right = self.right(x) return left + right class BoundaryRefine(torch.nn.Module): def __init__(self, in_channels): super(BoundaryRefine, self).__init__() self.layer = torch.nn.Sequential(torch.nn.Conv2d(in_channels, in_channels, kernel_size=3, padding=1), torch.nn.BatchNorm2d(in_channels), torch.nn.ReLU(inplace=True), torch.nn.Conv2d(in_channels, in_channels, kernel_size=3, padding=1), torch.nn.BatchNorm2d(in_channels)) def forward(self, x): convs = self.layer(x) return x.expand_as(convs)+convs class ResnetMIS(torch.nn.Module): def __init__(self, pretrained_resnet=True, out_channels=3): super().__init__() resent = models.resnet101(pretrained=pretrained_resnet) self.layer0 = torch.nn.Sequential(resent.conv1, resent.bn1, resent.relu, resent.maxpool) self.layer1 = resent.layer1 self.layer2 = resent.layer2 self.layer3 = resent.layer3 self.layer4 = resent.layer4 # Assuming input of size 240x320 self.gcn256 = GlobalConvolutionBlock(256, out_channels, (59, 79)) self.br256 = BoundaryRefine(out_channels) self.gcn512 = GlobalConvolutionBlock(512, out_channels, (29, 39)) self.br512 = BoundaryRefine(out_channels) self.gcn1024 = GlobalConvolutionBlock(1024, out_channels, (13, 19)) self.br1024 = BoundaryRefine(out_channels) self.gcn2048 = GlobalConvolutionBlock(2048, out_channels, (7, 9)) self.br2048 = BoundaryRefine(out_channels) self.br1 = BoundaryRefine(out_channels) self.br2 = BoundaryRefine(out_channels) self.br3 = BoundaryRefine(out_channels) self.br4 = BoundaryRefine(out_channels) self.br5 = BoundaryRefine(out_channels) self.activation = torch.nn.Sigmoid() self.deconv1 = torch.nn.ConvTranspose2d(out_channels, out_channels, 2, stride=2) self.deconv2 = torch.nn.ConvTranspose2d(out_channels, out_channels, 2, stride=2) initialize_weights(self.gcn256, self.gcn512, self.gcn1024, self.gcn2048, self.br5, self.br4, self.br3, self.br2, self.br1, self.br256, self.br512, self.br1024, self.br2048, self.deconv1, self.deconv2) def forward(self, x): x = self.layer0(x) layer1 = self.layer1(x) layer2 = self.layer2(layer1) layer3 = self.layer3(layer2) layer4 = self.layer4(layer3) enc1 = self.br256(self.gcn256(layer1)) enc2 = self.br512(self.gcn512(layer2)) enc3 = self.br1024(self.gcn1024(layer3)) enc4 = self.br2048(self.gcn2048(layer4)) dec1 = self.br1(torch.nn.functional.interpolate(enc4, size=enc3.size()[2:], mode='bilinear') + enc3) dec2 = self.br2(torch.nn.functional.interpolate(dec1, size=enc2.size()[2:], mode='bilinear') + enc2) dec3 = self.br3(torch.nn.functional.interpolate(dec2, size=enc1.size()[2:], mode='bilinear') + enc1) dec4 = self.br4(self.deconv1(dec3)) score_map = self.br5(self.deconv2(dec4)) return self.activation(score_map)
[ "shabat.eliran@gmail.com" ]
shabat.eliran@gmail.com
40c87a8686a9baae4fa8aeedbe59625b342a5fff
485384a2e1faa338d3372a8c000ccab6f6891cfa
/CGAT/Mali.py
6337aeb47d18672ec25b144ac94a69699dfb9d3a
[ "BSD-2-Clause" ]
permissive
wangzhennan14/cgat
30c8ec4f8d1e14413c5d8193352348137d60a611
38ff24edc2a1f8b0da5716aea4fbf6a9328342b0
refs/heads/master
2021-01-18T15:43:02.544097
2017-03-16T15:27:11
2017-03-16T15:27:11
null
0
0
null
null
null
null
UTF-8
Python
false
false
44,924
py
''' Mali.py - Tools for multiple alignments ======================================= :Author: Andreas Heger :Release: $Id$ :Date: |today| :Tags: Python Code ---- ''' import re import string import math import copy import random class AlignedString: mGapChars = ("-", ".") mGapChar = "-" def __init__(self, identifier, fr, to, s, is_mangled=False): self.mId = identifier self.mFrom = fr self.mTo = to self.mString = s self.mIsMangled = is_mangled def __len__(self): return len(self.mString) def maskColumn(self, column, mask_char="x"): self.maskColumns(self, [column, ], mask_char=mask_char) def maskColumns(self, columns, mask_char="x"): s = list(self.mString) for c in columns: if s[c] not in self.mGapChars: s[c] = mask_char self.mString = "".join(s) def mapColumns(self, columns, map_function): s = list(self.mString) for c in columns: if s[c] not in self.mGapChars: s[c] = map_function(s[c]) self.mString = "".join(s) def takeColumns(self, columns): """take columns. not implemented yet: updating of residue numbers. """ s = [] for c in columns: s.append(self.mString[c]) self.mString = "".join(s) def truncate(self, first, last): """truncate aligned string. not implemented yet: updating of residue numbers. """ self.mString = self.mString[first:last] def insertColumns(self, position, num_columns, char="-"): """insert num_columns columns at position.""" self.mString = self.mString[:position] + \ char * num_columns + self.mString[position:] def getIdentifier(self): '''return demangled identifier.''' if self.mIsMangled: return self.mId.rsplit("_", 1)[0] else: return self.mId def getResidueNumber(self, position): x = self.mFrom for c in self.mString[:position]: if c not in self.mGapChars: x += 1 return x def getSequence(self): """return sequence without gaps.""" return re.sub("[%s]" % "".join(self.mGapChars), "", self.mString) def getSegments(self, transitions, gap_chars="-."): """return segments of alignment according to transitions. Transitions is a list of residue numbers. """ segments = [] pos = self.mFrom first_x = 0 for x in range(len(self.mString)): char = self.mString[x] if char not in gap_chars: if pos in transitions: if first_x != x: segments.append((first_x, x)) first_x = x pos += 1 # add last segment, unless it has already been added # because it was part of the transitions if first_x != x: segments.append((first_x, len(self.mString))) elif pos in transitions: segments.append((first_x, len(self.mString))) return segments def threadSequence(self, new_sequence, map_old2new): """thread a new sequence onto this sequence replacing characters in this sequence with those in new sequence. ``map_old2new`` lists the positions that should be replaced. It maps non-gap positions in old to those in new. Non-gap positions than are not mapped are set to :attr:`gapchar`. """ s = list(self.mString) c = 0 for x in range(len(s)): m = map_old2new.mapRowToCol(c) if m >= 0: s[c] = new_sequence[m] else: s[c] = self.mGapChar if s[c] not in self.mGapChars: c += 1 self.mString = "".join(s) class MaliData: def __init__(self, line, gap_chars="-.", mask_chars="Nn"): self.mNGaps = len(re.sub("[^%s]" % gap_chars, "", line)) self.mNMasked = len(re.sub("[^%s]" % mask_chars, "", line)) self.mNAll = len(line) self.mNChars = self.mNAll - self.mNMasked - self.mNGaps def __str__(self): return "%i\t%i\t%i\t%i" % (self.mNAll, self.mNChars, self.mNGaps, self.mNMasked) class Mali: mGapChars = ("-", ".") mGapPattern = re.compile("[.-]") # character used for gaps mGapChar = "-" # admissable gap characters mGapChars = ("-", ".") mMaskChar = "X" def __init__(self): self.mIdentifiers = [] self.mMali = {} self.mLength = 0 self.mAnnotations = {} self.mName = None # set to false, if ranges shall not be output. self.mWriteRanges = True def __contains__(self, key): return key in self.mMali def __getitem__(self, key): return self.mMali[key].mString def __len__(self): return len(self.mIdentifiers) def __delitem__(self, key): del self.mMali[key] self.mIdentifiers.remove(key) def getClone(self): return copy.deepcopy(self) def iteritems(self): return iter(self.mMali.items()) def items(self): return list(self.mMali.items()) def values(self): return list(self.mMali.values()) def keys(self): return list(self.mMali.keys()) def getName(self): return self.mName def setName(self, name): self.mName = name def getIdentifiers(self): return self.mIdentifiers def getLength(self): """deprecated.""" return self.getNumSequences() def getNumSequences(self): return self.__len__() def getNumColumns(self): if self.mIdentifiers: return len(self.mMali[self.mIdentifiers[0]]) else: return 0 def getWidth(self): """deprecated.""" return self.getNumColumns() def rename(self, old_name, new_name): """rename an entry.""" if old_name not in self.mMali: raise KeyError("%s not in mali" % old_name) self.mIdentifiers[self.mIdentifiers.index(old_name)] = new_name self.mMali[new_name] = self.mMali[old_name] del self.mMali[old_name] def isEmpty(self): return len(self.mMali) == 0 def getSequence(self, key): return self.mMali[key] def getResidueNumber(self, key, position): """return residue number in sequence key at position position.""" return self.mMali[key].getResidueNumber(position) def setSequence(self, key, sequence): # TODO: updating of residue numbers self.mMali[key].mString = sequence def getEntry(self, key): return self.mMali[key] def addEntry(self, s): """add an aligned string object.""" if s.mId in list(self.mMali.keys()): raise KeyError("id %s already in mali" % s.mId) self.mIdentifiers.append(s.mId) self.mMali[s.mId] = s def addSequence(self, id, fr, to, sequence): if to < 0: to = self.countCharacters(sequence) s = AlignedString(id, fr, to, sequence) self.addEntry(s) def countCharacters(self, row): return len(row) - len(self.mGapPattern.findall(row)) def deleteEntry(self, identifier): if identifier not in self.mMali: raise KeyError("identifier %s not in mali." % identifier) del self.mMali[identifier] self.mIdentifiers.remove(identifier) def getColumns(self): """return mali in column orientation.""" args = [self.mMali[x].mString for x in self.mIdentifiers] return ["".join(x) for x in zip(*args)] def getConsensus(self, mark_with_gaps=False): """return consensus string. The consensus string returns the most frequent character per column that is not a gap. If mark_with_gaps is set to True, positions with any gap characater are set to gaps. """ columns = self.getColumns() seq = [] for x in range(len(columns)): s = columns[x] counts = [(a, s.count(a)) for a in set(list(s)).difference(set(self.mGapChar))] if mark_with_gaps and self.mGapChar in columns[x]: seq.append(self.mGapChar) else: counts.sort(key=lambda x: -x[1]) seq.append(counts[0][0]) return "".join(seq) def readFromFile(self, infile, format="fasta"): """read multiple alignment from file in various format.""" self.mMali = {} self.mIdentifiers = [] pattern_parse_ranges = re.compile("(\S+)/(\d+)-(\d+)") # read profiles - a profile possibly consists of several # entries per file so treat it differently if format.lower() == "profile": while 1: line = infile.readline() if not line: return False if line[0] != "#": break if line[0] != ">": raise "expected '>' at as first character in line %s" % line try: self.mName, length, width = re.match( ">profile=(\S+) length=(\d+) width=(\d+)", line).groups() except AttributeError: raise "could not parse header line %s" % line width = int(width) for x in range(0, width): id = "seq%i" % x self.mIdentifiers.append(id) line = infile.readline() if not line: raise "expected %i sequences, only got %i" % (width, x) self.mMali[id] = AlignedString( id, 0, self.countCharacters(line[:-1]), line[:-1]) return True if isinstance(infile, list) or \ isinstance(infile, tuple): lines = infile else: lines = infile.readlines() if format not in ("stockholm"): # save comments self.mComments = [x for x in lines if x[0] == "#"] lines = [x for x in lines if x[0] != "#"] else: self.mComments = [] # remove empty lines lines = [x for x in lines if x.strip()] if not lines: raise AttributeError("empty alignment") def getId(id, s): x = pattern_parse_ranges.match(id) if x: id, fr, to = x.groups() fr, to = int(fr) - 1, int(to) else: fr, to = 0, self.countCharacters(s) self.mWriteRanges = False return id, fr, to ####################################################################### if format.lower() == "plain": for line in lines: if not line.strip(): continue data = line[:-1].split("\t") id = data[3] xid = id x = 0 while xid in self.mMali: xid = id + "-" + str(x) x += 1 self.addEntry( AlignedString(xid, int(data[0]) - 1, int(data[2]), data[1])) ####################################################################### elif format.lower() == "fasta": pattern_identifier = "\S+" id = None fragments = [] for line in lines: if line[0] == ">": if id: s = re.sub("\s", "", string.join(fragments, "")) id, fr, to = getId(id, s) self.addEntry(AlignedString(id, fr, to, s)) id = re.search( "^(%s)" % pattern_identifier, line[1:-1]).group(0) fragments = [] continue fragments.append(line[:-1]) s = re.sub("\s", "", string.join(fragments, "")) id, fr, to = getId(id, s) self.addEntry(AlignedString(id, fr, to, s)) ####################################################################### elif format.lower() == "phylip": nsequences, nchars = re.split("\s+", lines[0][:-1].strip()) nsequences = int(nsequences) for line in lines[1:]: l = line[:-1].strip() if not l: continue id, sequence = re.match("(\S+)\s+(.*)", l).groups() sequence = re.sub("\s", "", sequence) if id not in self.mMali: self.mIdentifiers.append(id) self.mMali[id] = [] self.mMali[id].append(sequence) for id, frags in list(self.mMali.items()): s = "".join(frags) fr, to = 0, self.countCharacters(s) self.mMali[id] = AlignedString(id, fr, to, s) ####################################################################### elif format.lower() == "clustal": # skip header line del lines[0] fragments = {} # prune lines lines = [x.strip() for x in lines] # remove empty lines lines = [x for x in lines if len(x[:-1]) > 0] for line in lines: # remove consensus lines if line[0] in ("*", ":"): continue data = re.split("\s+", line) if len(data) != 2: raise ValueError("parsing error in line %s" % line) id, fragment = data if id not in fragments: fragments[id] = [] self.mIdentifiers.append(id) fragments[id].append(fragment) for id, f in list(fragments.items()): s = re.sub("\s", "", string.join(f, "")) self.mMali[id] = AlignedString( id, 0, self.countCharacters(s), s) elif format.lower() == "stockholm": # skip header line assert lines[0].startswith( "# STOCKHOLM"), "file is not in stockholm format" del lines[0] fragments = {} annotations = {} # prune lines lines = [x.strip() for x in lines] # remove empty lines lines = [x for x in lines if len(x[:-1]) > 0] for line in lines: data = re.split("\s+", line) if data[0] == "//": break if line[0] == '#': if data[0] == "#=GC": id, fragment = data[1:3] else: self.mComments.append(line) continue if id not in annotations: annotations[id] = [] annotations[id].append(fragment) else: if len(data) > 2: raise ValueError("parsing error in line %s" % line) elif len(data) == 1: # treat empty alignments/lines id = data[0] fragment = "" else: id, fragment = data if id not in fragments: fragments[id] = [] self.mIdentifiers.append(id) fragments[id].append(fragment) n = [] for id in self.mIdentifiers: f = fragments[id] s = re.sub("\s", "", string.join(f, "")) x = pattern_parse_ranges.match(id) if x: id, fr, to = x.groups() fr, to = int(fr) - 1, int(to) else: fr, to = 0, self.countCharacters(s) n.append(id) self.mMali[id] = AlignedString(id, fr, to, s) self.mIdentifiers = n for id, f in list(annotations.items()): s = re.sub("\s", "", string.join(f, "")) annotations[id] = s self.mAnnotations = annotations else: raise "unknown alignment format %s" % format if len(self.mMali) == 0: self.mLength = 0 else: self.mLength = min( [len(x.mString) for x in list(self.mMali.values())]) def writeToFile(self, outfile, write_ranges=True, format="plain", options=None): """write alignment to file. If options is given, these lines are output into the multiple alignment. """ if format == "plain-fasta": format = "fasta" write_ranges = False write_ranges = write_ranges and self.mWriteRanges if format == "plain": for identifier in self.mIdentifiers: m = self.mMali[identifier] outfile.write("%i\t%s\t%i\t%s\n" % ( m.mFrom + 1, m.mString, m.mTo, m.getIdentifier())) elif format == "fasta": for identifier in self.mIdentifiers: m = self.mMali[identifier] if write_ranges: outfile.write( ">%s/%i-%i\n%s\n" % (m.getIdentifier(), m.mFrom + 1, m.mTo, m.mString)) else: outfile.write(">%s\n%s\n" % (identifier, m.mString)) elif format == "stockholm": outfile.write("# STOCKHOLM 1.0\n") if options: for o in options: outfile.write("%s\n" % o) # calculate offset: max_l = 0 for identifier in self.mIdentifiers: m = self.mMali[identifier] id = m.getIdentifier() # tab does not work as separator if m.mTo and write_ranges: x = "%s/%i-%i" % (id, m.mFrom + 1, m.mTo) else: x = "%s" % (id) max_l = max(max_l, len(x)) for identifier in list(self.mAnnotations.keys()): x = "#=GC %s" % identifier max_l = max(max_l, len(x)) format = "%-" + str(max_l) + "s %s\n" for identifier in self.mIdentifiers: m = self.mMali[identifier] id = m.getIdentifier() # tab does not work as separator if m.mTo and write_ranges: x = "%s/%i-%i" % (id, m.mFrom + 1, m.mTo) else: x = "%s" % (id) outfile.write(format % (x, m.mString)) for identifier, value in list(self.mAnnotations.items()): x = "#=GC %s" % identifier outfile.write(format % (x, value)) outfile.write("//\n") elif format == "phylip": outfile.write("%i %i\n" % (self.getLength(), self.getWidth())) for identifier in self.mIdentifiers: outfile.write("%s %s\n" % (identifier, self.mMali[identifier].mString)) elif format.lower() == "profile": if self.mName: name = self.mName else: name = ",".join(self.mIdentifiers) outfile.write(">profile=%s length=%i width=%i\n" % (name, self.getWidth(), self.getLength())) for identifier in self.mIdentifiers: outfile.write("%s\n" % (self.mMali[identifier].mString)) elif format == "nexus": # nexus formatted output - MrBayes conformant. outfile.write("#NEXUS\n") outfile.write("begin data;\n") outfile.write(" dimensions ntax=%i nchar=%i;\n" % (self.getLength(), self.getWidth())) outfile.write( " format datatype=dna interleave=no gap=%s;\n" % (self.mGapChar)) outfile.write(" matrix\n") max_len = max([len(x) for x in self.mIdentifiers]) format = " %-" + str(max_len) + "s %s\n" for identifier in self.mIdentifiers: outfile.write( format % (identifier, self.mMali[identifier].mString)) outfile.write(" ;\n") outfile.write("end;\n") else: raise "unknown alignment format %s" % format def removeUnalignedEnds(self): """remove unaligned ends in the multiple alignment. unaligned ends correspond to lower-case characters. """ pattern_start = re.compile("^([- .a-z]+)") pattern_unaligned = re.compile("[a-z]") for s in list(self.mMali.values()): first = pattern_start.match(s.mString) if first: first = first.groups()[0] nchars = len(pattern_unaligned.findall(first)) s.mFrom += nchars s.mString = self.mGapChar * len(first) + s.mString[len(first):] # search from the back end by reversing. This is much faster than # using $ from the back. last = pattern_start.match(s.mString[::-1]) if last: last = last.groups()[0] nchars = len(pattern_unaligned.findall(last)) s.mTo -= nchars l = len(s) - len(last) s.mString = s.mString[:l] + self.mGapChar * l def upperCase(self): """set all characters to upper case.""" for k, s in list(self.mMali.items()): s.mString = s.mString.upper() def lowerCase(self): """set all characters to lower case.""" for k, s in list(self.mMali.items()): s.mString = s.mString.lower() def removeEndGaps(self): """remove end gaps. end gaps do not include any characters and thus the alignment coordinates won't change. """ pattern_start_gaps = re.compile("^([- ]+)") min_from = self.mLength max_to = 0 for s in list(self.mMali.values()): first = pattern_start_gaps.match(s.mString) if first: first = first.groups()[0] min_from = min(min_from, len(first)) # search from the back end by reversing. This is much faster than # using $ from the back. last = pattern_start_gaps.search(s.mString[::-1]) if last: last = last.groups()[0] max_to = max(max_to, len(s) - len(last)) for s in list(self.mMali.values()): s.mString = s.mString[min_from:max_to] self.mLength = min([x.mString for x in list(self.mMali.values())]) def insertColumns(self, position, num_gaps, keep_fixed=None, char="-"): """insert gaps at position into multiple alignment. if keep_constant is a list of identifiers, those are kept constant, instead, gaps are added to the end. """ last_pos = min(self.getWidth(), position + num_gaps) for id, seq in list(self.mMali.items()): if keep_fixed and id in keep_fixed: seq.insertColumns(last_pos, num_gaps, char) else: seq.insertColumns(position, num_gaps, char) def removeGaps(self, allowed_gaps=0, minimum_gaps=1, frame=1): """remove gappy columns. allowed_gaps: number of gaps allowed for column to be kept minimum_gaps: number of gaps for column to be removed set minimum_gaps to the number of sequences to remove columns with all gaps. If frame is > 1 (3 most likely), then a whole codon is removed as soon as there is one column to be removed. """ self.removePattern( match_function=lambda x: x in self.mGapChars, allowed_matches=allowed_gaps, minimum_matches=minimum_gaps, delete_frame=frame) def removePattern(self, match_function, allowed_matches=0, minimum_matches=1, delete_frame=1, search_frame=1): """remove columns (or group of columns), that match a certain pattern. allowed_matches: number of matches allowed so that column is still kept minimum_matches: number of matches required for column to be removed set minimum_matches to the number of sequences to remove columns with all gaps. Patterns are matches in search_frame. For example, if frame is 3, whole codons are supplied to match_function. delete_frame specifies the frame for deletion. If it is set to 3, codons are removed if already one column matches. Example: remove all columns that contain at least one stop-codon: removePattern( lambda x: x.upper() in ("TAG", "TAA", "TGA"), allowed_matches = 0, minimum_matches = 1, search_frame = 3, delete_frame = 3) """ nmatches = [0] * self.getWidth() for s in [x.mString for x in list(self.mMali.values())]: for x in range(0, len(s), search_frame): segment = s[x:x + search_frame] if match_function(segment): nmatches[x] += 1 columns = [] delete_columns = [] for x in range(len(nmatches)): if nmatches[x] >= allowed_matches and nmatches[x] < minimum_matches: columns.append(x) else: delete_columns.append(x) if delete_frame != 1: s = set(columns) for x in delete_columns: start = int(math.floor(float(x) / delete_frame) * delete_frame) end = start + delete_frame for c in range(start, end): if c in s: s.remove(c) columns = list(s) columns.sort() self.takeColumns(columns) def removeEmptySequences(self): """remove sequences that are completely empty. """ new_ids = [] for id in self.mIdentifiers: if self.countCharacters(self.mMali[id].mString) == 0: del self.mMali[id] continue new_ids.append(id) self.mIdentifiers = new_ids def upper(self): """convert all characters in mali to uppercase.""" for s in list(self.mMali.values()): s.mString = s.mString.upper() def lower(self): """convert all characters in mali to lowercase.""" for s in list(self.mMali.values()): s.mString = s.mString.lower() def shiftAlignment(self, map_id2offset): """shift alignment by offset.""" for identifier, m in list(self.mMali.items()): if identifier in map_id2offset: o = map_id2offset[identifier] m.mFrom += o m.mTo += o def markCodons(self, mode="case"): """mark codons. """ for identifier, m in list(self.mMali.items()): s = m.mString if len(s) % 3 != 0: raise "sequence %s not divisible by 3" % (m.mId) is_upper = True sequence = [] for x in range(0, len(s), 3): if is_upper: sequence.append(s[x:x + 3].upper()) is_upper = False else: sequence.append(s[x:x + 3].lower()) is_upper = True m.mString = "".join(sequence) def markTransitions(self, map_id2transitions, mode="case"): """mark transitions in the multiple alignment. if mode == case, then upper/lower case is used for the transitions Otherwise, a character given by mode is inserted. """ if mode in ("case", "keep-odd", "keep-even"): # check, if the whole alignment needs to be masked/marked: if "mali" in map_id2transitions: transitions = map_id2transitions["mali"] for identifier, s in list(self.mMali.items()): new_chars = [] is_upper = True is_first = False for c in range(len(s)): if c in transitions: is_first = True if is_upper: is_upper = False else: is_upper = True x = s.mString[c] if mode == "case": if x in string.lowercase: x = self.mMaskChar if is_upper: x = string.upper(x) else: x = string.lower(x) elif mode == "keep-even": if is_upper: x = self.mGapChar elif mode == "keep-odd": if not is_upper: x = self.mGapChar new_chars.append(x) s.mString = "".join(new_chars) # now do individual sequences for identifier, s in list(self.mMali.items()): if identifier not in map_id2transitions: continue new_chars = [] c = s.mFrom is_upper = True transitions = map_id2transitions[identifier] for x in s.mString: if x in self.mGapChars: pass else: if c in map_id2transitions[identifier]: if is_upper: is_upper = False else: is_upper = True c += 1 if x in string.lowercase: x = self.mMaskChar if mode == "case": if is_upper: x = string.upper(x) else: x = string.lower(x) elif mode == "keep-even": if is_upper: x = self.mGapChar elif mode == "keep-odd": if not is_upper: x = self.mGapChar new_chars.append(x) s.mString = "".join(new_chars) else: raise ValueError("character insertion not implemented yet") def buildColumnMap(self, other, join_field=None): """build map of columns in other to this.""" if not join_field: join_field = other.mIdentifiers[0] if join_field not in other.mMali or \ join_field not in self.mMali: raise "line %s not in both alignments." % (join_field) this_seq = self.mMali[join_field] other_seq = other.mMali[join_field] if this_seq.mFrom != other_seq.mFrom or \ this_seq.mTo != other_seq.mTo: raise ValueError("residue ranges for sequence %s doe not correspond." % ( join_field)) map_this2other = [] this_seq = this_seq.mString.upper() other_seq = other_seq.mString.upper() # position in other o = 0 for c in this_seq: if c in self.mGapChars: map_this2other.append(None) else: while other_seq[o] != c: o += 1 map_this2other.append(o) o += 1 return map_this2other def shuffle(self, frame=1): """shuffle multiple alignment. The frame determines the block size for shuffling. Use 3 for codons in a multiple alignment without frame-shifts. """ columns = list(range(self.getNumColumns() // frame)) random.shuffle(columns) if frame > 1: cc = [] for x in columns: cc += [x * frame + y for y in range(frame)] columns = cc self.takeColumns(columns) def propagateMasks(self, min_chars=1, mask_char="x"): """propagate masked characters to all rows of a multiple alignment within a column. If there is at least min_chars in a mali column, that are masks, propagate the masks to all other rows. """ masks_per_column = {} for identifier, s in list(self.mMali.items()): r = s.mString.lower() for x in range(len(r)): if r[x] == mask_char: if x not in masks_per_column: masks_per_column[x] = 0 masks_per_column[x] += 1 columns_to_mask = [] for c, n in list(masks_per_column.items()): if n >= min_chars: columns_to_mask.append(c) columns_to_mask.sort() self.maskColumns(columns_to_mask, mask_char=mask_char) def propagateTransitions(self, min_chars=1): """propagate lower case in a column to all residues. """ columns_to_change = set() for identifier, s in list(self.mMali.items()): r = s.mString for x in range(len(r)): if r[x] in string.lowercase: columns_to_change.add(x) columns_to_change = list(columns_to_change) columns_to_change.sort() self.mapColumns(columns_to_change, string.lower) def takeColumns(self, columns): """restrict alignments to certain columns.""" for identifier, s in list(self.mMali.items()): s.takeColumns(columns) for key, anno in list(self.mAnnotations.items()): self.mAnnotations[key] = "".join([anno[c] for c in columns]) def maskColumns(self, columns, mask_char="x"): """mask columns in a multiple alignment.""" for identifier, s in list(self.mMali.items()): s.maskColumns(columns, mask_char=mask_char) def mapColumns(self, columns, map_function): """apply map_function to all residues in columns.""" for identifier, s in list(self.mMali.items()): s.mapColumns(columns, map_function) def recount(self, reset_first=False): """recount residue in alignments.""" for id, seq in list(self.mMali.items()): if reset_first: seq.mFrom = 0 seq.mTo = seq.mFrom + self.countCharacters(seq.mString) def maskColumn(self, column, mask_char="x"): """mask a column.""" for identifier, s in list(self.mMali.items()): s.maskColumn(column, mask_char) def copyAnnotations(self, other): """copy annotations from annother mali.""" map_this2other = self.buildColumnMap(other) ncols = self.getWidth() for key, annotation in list(other.mAnnotations.items()): a = [] for x in range(ncols): m = map_this2other[x] if m is not None: a.append(annotation[m]) else: a.append(self.mGapChar) self.mAnnotations[key] = "".join(a) def getAnnotation(self, key): """return annotation associated with key.""" return self.mAnnotations[key] def setAnnotation(self, key, value): """set annotation associated with key to value.""" self.mAnnotations[key] = value def addAnnotation(self, key, annotation): """add annotation.""" xkey = key n = 0 while xkey in self.mAnnotations: xkey = "%s_%i" % (key, n) self.mAnnotations[xkey] = re.sub("\s", "", annotation) def truncate(self, first, last): """truncate alignment within range.""" for key, value in list(self.mMali.items()): value.truncate(first, last) def mapIdentifiers(self, map_old2new=None, pattern_identifier="ID%06i"): """map identifiers in multiple aligment. if map_old2new is not given, a new map is created (map_new2old) """ new_identifiers = [] if map_old2new is not None: for id in self.mIdentifiers: new_id = map_old2new[id] new_identifiers.append(new_id) entry = self.mMali[id] del self.mMali[id] # entry.mId = new_id self.mMali[new_id] = entry else: map_old2new = {} n = 1 for id in self.mIdentifiers: new_id = pattern_identifier % n new_identifiers.append(new_id) entry = self.mMali[id] del self.mMali[id] # entry.mId = new_id self.mMali[new_id] = entry map_old2new[new_id] = id n += 1 self.mIdentifiers = new_identifiers return map_old2new def getAlphabet(self): """get alphabet from the multiple alignment. Alphabet is "na", if more than 90% of characaters are "actgxn", otherwise it is "aa". """ s = "".join([x.mString for x in list(self.values())]).lower() s = re.sub("[%s]" % "".join(self.mGapChars), "", s) ss = re.sub("[acgtxn]", "", s) if float(len(ss)) < (len(s) * 0.1): return "na" else: return "aa" def checkLength(self): """check lengths of aligned strings. Return false if they are inconsistent.""" l = None for s in list(self.mMali.values()): if l is None: l = len(s.mString) else: if l != len(s.mString): return False return True def clipByAnnotation(self, key, chars=""): """restrict alignment to positions where annotation identified by key in chars. if chars is empty, nothing is clipped. """ if key not in self.mAnnotations: return if chars == "": return if self.getNumColumns() == 0: return anno = self.mAnnotations[key] columns = [] for x in range(len(anno)): if anno[x] in chars: columns.append(x) self.takeColumns(columns) def apply(self, f): """apply function f to every row in the multiple alignment.""" for s in list(self.mMali.values()): f(s) def filter(self, f): '''filter multiple alignment using function *f*. The function *f* should return True for entries that should be kept and False for those that should be removed. ''' ids, vals = [], [] for id in self.mIdentifiers: s = self.mMali[id] if f(s): vals.append(s) ids.append(id) self.mIdentifiers = ids self.mMali = vals class SequenceCollection(Mali): """reads in a sequence collection, but permits several entries per id. Note that this might cause problems with interleaved formats like phylips or clustal. This mapping is achieved by appending a numeric suffix. The suffix is retained for the life-time of the object, but not output to a file. """ def __init__(self): Mali.__init__(self) def addEntry(self, s): """add an aligned string object.""" id = s.mId if id in list(self.mMali.keys()): x = 1 while "%s_%i" % (id, x) in list(self.mMali.keys()): x += 1 id = "%s_%i" % (id, x) self.mIdentifiers.append(id) s.mIsMangled = True self.mMali[id] = s def readFromFile(self, infile, format="fasta"): """read multiple alignment from file in various format.""" self.mMali = {} self.mIdentifiers = [] pattern_parse_ranges = re.compile("(\S+)/(\d+)-(\d+)") if isinstance(infile, list) or \ isinstance(infile, tuple): lines = infile else: lines = infile.readlines() if format not in ("stockholm"): # save comments self.mComments = [x for x in lines if x[0] == "#"] lines = [x for x in lines if x[0] != "#"] else: self.mComments = [] def getId(id, s): x = pattern_parse_ranges.match(id) if x: id, fr, to = x.groups() fr, to = int(fr) - 1, int(to) else: fr, to = 0, self.countCharacters(s) self.mWriteRanges = False return id, fr, to ####################################################################### if format.lower() == "plain": for line in lines: if not line.strip(): continue data = line[:-1].split("\t") id = data[3] xid = id x = 0 while xid in self.mMali: xid = id + "-" + str(x) x += 1 self.addEntry( AlignedString(xid, int(data[0]) - 1, int(data[2]), data[1])) ####################################################################### elif format.lower() == "fasta": pattern_identifier = "\S+" id = None fragments = [] for line in lines: if line[0] == ">": if id: s = re.sub("\s", "", string.join(fragments, "")) id, fr, to = getId(id, s) self.addEntry(AlignedString(id, fr, to, s)) id = re.search( "^(%s)" % pattern_identifier, line[1:-1]).group(0) fragments = [] continue fragments.append(line[:-1]) s = re.sub("\s", "", string.join(fragments, "")) id, fr, to = getId(id, s) self.addEntry(AlignedString(id, fr, to, s)) ####################################################################### elif format.lower() == "phylip": raise ValueError("phylip not implemented") ####################################################################### elif format.lower() == "clustal": raise ValueError("clustal not implemented") elif format.lower() == "stockholm": raise ValueError("stockholm not implemented") if len(self.mMali) == 0: self.mLength = 0 else: self.mLength = min( [len(x.mString) for x in list(self.mMali.values())]) def convertMali2Alignlib(mali): '''convert a multiple alignment of type :class:`Mali` into an alignlib_lite.py_multiple alignment object. ''' import alignlib_lite m = alignlib_lite.py_makeMultipleAlignment() for identifier in mali.getIdentifiers(): a = alignlib_lite.py_makeAlignatum(mali[identifier]) m.add(a) return m def convertAlignlib2Mali(mali, identifiers=None, seqs=None): """convert a multiple alignment into an alignlib_lite.py_multiple alignment object.""" m = Mali() if not identifiers: identifiers = ["%i" % x for x in range(mali.getNumSequences())] if seqs is None: # old style MultipleAlignment for x in range(mali.getNumSequences()): a = mali.getRow(x) m.addSequence( identifiers[x], a.getFrom(), a.getTo(), a.getString()) else: import alignlib_lite output = alignlib_lite.py_MultAlignmentFormatPlain(mali, seqs) for x in range(mali.getNumSequences()): a = output.mData[x] m.addSequence( identifiers[x], a.getFrom(), a.getTo(), a.getString()) return m
[ "andreas.heger@gmail.com" ]
andreas.heger@gmail.com
9e8faea111d72f8b2e5ec2a6e79cd03f082cad11
8f24e443e42315a81028b648e753c50967c51c78
/python/ray/train/examples/horovod/horovod_pytorch_example.py
f4d15ae0515b7fad3547711c35c685f4407f2ced
[ "MIT", "BSD-3-Clause", "Apache-2.0" ]
permissive
simon-mo/ray
d07efdada8d05c6e10417f96e8dfc35f9ad33397
1e42e6cd15e2fb96c217cba8484e59ed0ef4b0c8
refs/heads/master
2023-03-06T00:09:35.758834
2022-12-23T18:46:48
2022-12-23T18:46:48
122,156,396
4
2
Apache-2.0
2023-03-04T08:56:56
2018-02-20T04:47:06
Python
UTF-8
Python
false
false
8,402
py
import argparse from filelock import FileLock import horovod.torch as hvd import os import torch.nn as nn import torch.nn.functional as F import torch.optim as optim import torch.utils.data.distributed from torchvision import datasets, transforms from ray.air import session from ray.air.config import ScalingConfig from ray.train.horovod import HorovodTrainer from ray.train.torch.torch_checkpoint import TorchCheckpoint import ray.train.torch def metric_average(val, name): tensor = torch.tensor(val) avg_tensor = hvd.allreduce(tensor, name=name) return avg_tensor.item() class Net(nn.Module): def __init__(self): super(Net, self).__init__() self.conv1 = nn.Conv2d(1, 10, kernel_size=5) self.conv2 = nn.Conv2d(10, 20, kernel_size=5) self.conv2_drop = nn.Dropout2d() self.fc1 = nn.Linear(320, 50) self.fc2 = nn.Linear(50, 10) def forward(self, x): x = F.relu(F.max_pool2d(self.conv1(x), 2)) x = F.relu(F.max_pool2d(self.conv2_drop(self.conv2(x)), 2)) x = x.view(-1, 320) x = F.relu(self.fc1(x)) x = F.dropout(x, training=self.training) x = self.fc2(x) return F.log_softmax(x) def setup(config): data_dir = config.get("data_dir", None) seed = config.get("seed", 42) batch_size = config.get("batch_size", 64) use_adasum = config.get("use_adasum", False) lr = config.get("lr", 0.01) momentum = config.get("momentum", 0.5) use_cuda = config.get("use_cuda", False) # Horovod: initialize library. hvd.init() torch.manual_seed(seed) if use_cuda: # Horovod: pin GPU to local rank. torch.cuda.set_device(hvd.local_rank()) torch.cuda.manual_seed(seed) # Horovod: limit # of CPU threads to be used per worker. torch.set_num_threads(1) kwargs = {"pin_memory": True} if use_cuda else {} data_dir = data_dir or "~/data" with FileLock(os.path.expanduser("~/.horovod_lock")): train_dataset = datasets.MNIST( data_dir, train=True, download=True, transform=transforms.Compose( [transforms.ToTensor(), transforms.Normalize((0.1307,), (0.3081,))] ), ) # Horovod: use DistributedSampler to partition the training data. train_sampler = torch.utils.data.distributed.DistributedSampler( train_dataset, num_replicas=hvd.size(), rank=hvd.rank() ) # Note, don't set `num_workers` in DataLoader (not even 1), # as that will separately start multiple processes (each corresponding to 1 worker) # to load the data. This is known to cause issues with Ray. train_loader = torch.utils.data.DataLoader( train_dataset, batch_size=batch_size, sampler=train_sampler, **kwargs ) model = Net() # By default, Adasum doesn't need scaling up learning rate. lr_scaler = hvd.size() if not use_adasum else 1 if use_cuda: # Move model to GPU. model.cuda() # If using GPU Adasum allreduce, scale learning rate by local_size. if use_adasum and hvd.nccl_built(): lr_scaler = hvd.local_size() # Horovod: scale learning rate by lr_scaler. optimizer = optim.SGD(model.parameters(), lr=lr * lr_scaler, momentum=momentum) # Horovod: wrap optimizer with DistributedOptimizer. optimizer = hvd.DistributedOptimizer( optimizer, named_parameters=model.named_parameters(), op=hvd.Adasum if use_adasum else hvd.Average, ) return model, optimizer, train_loader, train_sampler def train_epoch( model, optimizer, train_sampler, train_loader, epoch, log_interval, use_cuda ): loss = None model.train() # Horovod: set epoch to sampler for shuffling. train_sampler.set_epoch(epoch) for batch_idx, (data, target) in enumerate(train_loader): if use_cuda: data, target = data.cuda(), target.cuda() optimizer.zero_grad() output = model(data) loss = F.nll_loss(output, target) loss.backward() optimizer.step() if batch_idx % log_interval == 0: # Horovod: use train_sampler to determine the number of # examples in this worker's partition. print( "Train Epoch: {} [{}/{} ({:.0f}%)]\tLoss: {:.6f}".format( epoch, batch_idx * len(data), len(train_sampler), 100.0 * batch_idx / len(train_loader), loss.item(), ) ) return loss.item() if loss else None def train_func(config): num_epochs = config.get("num_epochs", 10) log_interval = config.get("log_interval", 10) use_cuda = config.get("use_cuda", False) save_model_as_dict = config.get("save_model_as_dict", False) model, optimizer, train_loader, train_sampler = setup(config) results = [] for epoch in range(num_epochs): loss = train_epoch( model, optimizer, train_sampler, train_loader, epoch, log_interval, use_cuda ) if save_model_as_dict: checkpoint = TorchCheckpoint.from_state_dict(model.state_dict()) else: checkpoint = TorchCheckpoint.from_model(model) results.append(loss) session.report(dict(loss=loss), checkpoint=checkpoint) # Only used for testing. return results def main(num_workers, use_gpu, kwargs): trainer = HorovodTrainer( train_loop_per_worker=train_func, train_loop_config={ "num_epochs": kwargs["num_epochs"], "log_interval": kwargs["log_interval"], "use_cuda": kwargs["use_cuda"], }, scaling_config=ScalingConfig(num_workers=num_workers, use_gpu=use_gpu), ) result = trainer.fit() print(result) if __name__ == "__main__": # Training settings parser = argparse.ArgumentParser( description="PyTorch MNIST Example", formatter_class=argparse.ArgumentDefaultsHelpFormatter, ) parser.add_argument( "--batch-size", type=int, default=64, metavar="N", help="input batch size for training (default: 64)", ) parser.add_argument( "--num-epochs", type=int, default=5, metavar="N", help="number of epochs to train (default: 10)", ) parser.add_argument( "--lr", type=float, default=0.01, metavar="LR", help="learning rate (default: 0.01)", ) parser.add_argument( "--momentum", type=float, default=0.5, metavar="M", help="SGD momentum (default: 0.5)", ) parser.add_argument( "--use-gpu", action="store_true", default=False, help="enables CUDA training" ) parser.add_argument( "--seed", type=int, default=42, metavar="S", help="random seed (default: 42)" ) parser.add_argument( "--log-interval", type=int, default=10, metavar="N", help="how many batches to wait before logging training status", ) parser.add_argument( "--use-adasum", action="store_true", default=False, help="use adasum algorithm to do reduction", ) parser.add_argument( "--num-workers", type=int, default=2, help="Number of Ray workers to use for training.", ) parser.add_argument( "--data-dir", help="location of the training dataset in the local filesystem (" "will be downloaded if needed)", ) parser.add_argument( "--address", required=False, type=str, default=None, help="Address of Ray cluster.", ) args = parser.parse_args() if args.address: ray.init(args.address) else: ray.init() use_cuda = args.use_gpu if args.use_gpu is not None else False kwargs = { "data_dir": args.data_dir, "seed": args.seed, "use_cuda": use_cuda, "batch_size": args.batch_size, "use_adasum": args.use_adasum if args.use_adasum else False, "lr": args.lr, "momentum": args.momentum, "num_epochs": args.num_epochs, "log_interval": args.log_interval, } main(num_workers=args.num_workers, use_gpu=use_cuda, kwargs=kwargs)
[ "noreply@github.com" ]
simon-mo.noreply@github.com
188c2472291601922fddbb95d9c6cdbe3ca24173
d12e13bab06ba7083a41aba7e7f74fa40926f0cc
/seq2seq-affect-attention/model/Decoder.py
4205fde03a3c06a698c12fa5da91565c6a50e86d
[]
no_license
HengYangDS/seq2seq-affect
39ca15998f7824f29932832c880cba416a478682
e5c40651540fea258a2d683be0fe532763168853
refs/heads/master
2020-09-25T16:09:53.663929
2019-12-05T02:10:13
2019-12-05T02:10:13
226,040,899
1
0
null
2019-12-05T07:21:28
2019-12-05T07:21:27
null
UTF-8
Python
false
false
1,031
py
import torch.nn as nn # 解码器 class Decoder(nn.Module): def __init__(self, cell_type, # rnn类型 input_size, # 输入维度 output_size, # 输出维度 num_layer, # rnn层数 dropout=0): # dropout super(Decoder, self).__init__() assert cell_type in ['GRU', 'LSTM'] # 限定rnn类型 self.cell_type = cell_type self.rnncell = getattr(nn, cell_type)( # rnncell input_size=input_size, hidden_size=output_size, num_layers=num_layer, dropout=dropout) def forward(self, input, # 输入 [seq, batch, dim] 或者单步输入 [1, batch, dim] state): # 初始状态 [layers*directions, batch, dim] # output = [seq, batch, dim*directions] 每个时间步的输出 # final_state = [layers*directions, batch, dim] # 每一层的最终状态 output, final_state = self.rnncell(input, state) return output, final_state
[ "1759567121@qq.com" ]
1759567121@qq.com
9cee7ee01989eaac94944331ddf1326936e6d652
b45df36da7d09b3170c503753dabf62ac28db8d8
/book/exercise/Key-valueOperations/Python/TweetService.py
da23adede52957c795737d066a9b02e35525edcb
[ "MIT" ]
permissive
skatsuta/aerospike-training
42b9a6e9372bd396d3489ec73778df7c00f5829e
d08d1805e19277b08327df5670e4c335ee63b853
refs/heads/master
2023-08-16T13:03:09.930256
2014-12-12T17:18:02
2014-12-12T17:18:02
27,905,857
0
0
MIT
2023-08-03T19:46:21
2014-12-12T04:52:47
C#
UTF-8
Python
false
false
8,093
py
#!/usr/bin/env python # # * Copyright 2012-2014 by Aerospike. # * # * Permission is hereby granted, free of charge, to any person obtaining a copy # * of this software and associated documentation files (the "Software"), to # * deal in the Software without restriction, including without limitation the # * rights to use, copy, modify, merge, publish, distribute, sublicense, and/or # * sell copies of the Software, and to permit persons to whom the Software is # * furnished to do so, subject to the following conditions: # * # * The above copyright notice and this permission notice shall be included in # * all copies or substantial portions of the Software. # * # * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR # * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, # * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE # * AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER # * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING # * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS # * IN THE SOFTWARE. # from __future__ import print_function import aerospike import sys import time from aerospike import predicates as p import random AS_POLICY_W_EXISTS = "exists" AS_POLICY_EXISTS_UNDEF = 0 # Use default value AS_POLICY_EXISTS_IGNORE= 1 # Write the record, regardless of existence. AS_POLICY_EXISTS_CREATE= 2 # Create a record, ONLY if it doesn't exist. AS_POLICY_EXISTS_UPDATE= 3 # Update a record, ONLY if it exist (NOT YET IMPL). class TweetService(object): def __init__(self, client): self.client = client def createTweet(self): print("\n********** Create Tweet **********\n") # /*********************/// # /*****Data Model*****/// # Namespace: test # Set: tweets # Key: <username:<counter>> # Bins: # tweet - string # ts - int (Stores epoch timestamp of the tweet) # username - string # Sample Key: dash:1 # Sample Record: # { tweet: 'Put. A. Bird. On. It.', # ts: 1408574221, # username: 'dash' # } # /*********************/// userRecord = None userKey = None tweetKey = None # Get username username = str() username = raw_input("Enter username: ") if len(username) > 0: # Check if username exists # Exercise 2 print("\nTODO: Check if username exists") meta = None policy = None record = {} if userRecord: # Set Tweet Count if 'tweetcount' in userRecord: nextTweetCount = int(userRecord['tweetcount']) + 1 else: nextTweetCount = 1 # Get tweet record['tweet'] = raw_input("Enter tweet for " + username + ":") # Create timestamp to store along with the tweet so we can # query, index and report on it ts= self.getTimeStamp() # TODO: Create WritePolicy instance # Exercise 2 print("\nTODO: Create WritePolicy instance"); #TODO: Create Key and Bin instances for the tweet record. HINT: tweet key should be in username:nextTweetCount format # Exercise 2 print("\nTODO: Create Key and Bin instances for the tweet record"); # TODO: Write tweet record # Exercise 2 print("\nTODO: Write tweet record"); # TODO: Update tweet count and last tweeted timestamp in the user # Exercise 2 print("\nINFO: Tweet record created!\n",record,tweetKey) # Update tweet count and last tweet'd timestamp in the user record else: print("ERROR: User record not found!\n") def scanAllTweetsForAllUsers(self): # Initiate scan operation that invokes callback for outputting tweets on the console # Exercise 4 print("\nTODO: Initiate scan operation that invokes callback for outputting tweets to the console") def updateUser(self, client, userKey, policy, ts, tweetCount): # TODO: Update tweet count and last tweeted timestamp in the user record # Exercise 2 print("\nTODO: Update tweet count and last tweeted timestamp in the user record") def updateUserUsingOperate(self, client, userKey, policy, ts): """ operate not supported in Python Client """ print("\nINFO: The tweet count now is: ") def queryTweetsByUsername(self): print("\n********** Query Tweets By Username **********\n") def queryUsersByTweetCount(self): print("\n********** Query Users By Tweet Count Range **********\n") def getTimeStamp(self): return int(round(time.time() * 1000)) def createTweets(self): randomTweets = ["For just $1 you get a half price download of half of the song and listen to it just once.", "People tell me my body looks like a melted candle", "Come on movie! Make it start!", "Byaaaayy", "Please, please, win! Meow, meow, meow!", "Put. A. Bird. On. It.", "A weekend wasted is a weekend well spent", "Would you like to super spike your meal?", "We have a mean no-no-bring-bag up here on aisle two.", "SEEK: See, Every, EVERY, Kind... of spot", "We can order that for you. It will take a year to get there.", "If you are pregnant, have a soda.", "Hear that snap? Hear that clap?", "Follow me and I may follow you", "Which is the best cafe in Portland? Discuss...", "Portland Coffee is for closers!", "Lets get this party started!", "How about them portland blazers!", "You got school'd, yo", "I love animals", "I love my dog", "What's up Portland", "Which is the best cafe in Portland? Discuss...", "I dont always tweet, but when I do it is on Tweetaspike"] totalUsers = 10000 maxTweets = 20 username = str() ts = 0 wr_policy = { AS_POLICY_W_EXISTS: AS_POLICY_EXISTS_IGNORE } print("\nCreate up to " , maxTweets , " tweets each for " , totalUsers , " users. Press any key to continue...\n") raw_input("..") j = 0 while j < totalUsers: username = "user" + str(random.randint(1,totalUsers)) userKey = ("test", "users", username) meta = None policy = None ts = None k = 0 (key, metadata,userRecord) = self.client.get(userKey,policy) if userRecord: totalTweets = random.randint(1,maxTweets) while k <= totalTweets: record = {} ts = self.getTimeStamp() tweetKey = ("test", "tweets", username + ":" + str(k)) record["tweet"] = random.choice(randomTweets) record["ts"] = ts record["username"]= username self.client.put(tweetKey,record, meta, wr_policy) k += 1 # Create timestamp to store along with the tweet so we can # query, index and report on it print("\nWrote " , totalTweets , " tweets for " , username , "!") if totalTweets > 0: # Update tweet count and last tweet'd timestamp in the user # record self.updateUser(self.client, userKey, wr_policy, ts, totalTweets) j += 1 # Check if user record exists # create up to maxTweets random tweets for this user # Create timestamp to store along with the tweet so we can # query, index and report on it # Update tweet count and last tweet'd timestamp in the user # record print("\n\nDone creating up to " , maxTweets , " tweets each for " , totalUsers , " users!\n")
[ "katsuta_soshi@cyberagent.co.jp" ]
katsuta_soshi@cyberagent.co.jp
73f8ddeef35504d0ce12496657651f487254073c
9edff345ec24494058d5ce01b4fb46012dc8bf29
/中级算法/回溯算法/单词搜索.py
180c8e196a24a8a9cd990344a4e486c18e38aac9
[]
no_license
gump1368/leetcode-python
65909bb9931e3840d29743222d92af3a77902f33
2d4f47e21cc105700feccb0d525408bdc54c663c
refs/heads/master
2023-01-24T03:50:19.447898
2023-01-12T07:03:12
2023-01-12T07:03:12
235,063,257
0
0
null
null
null
null
UTF-8
Python
false
false
3,073
py
#! -*- coding: utf-8 -*- """ @Author: gump @Create Time: 20220714 @Info: 单词搜索 """ from typing import List # def exist(board: List[List[str]], word: str) -> bool: # """ # 输入:board = [["A","B","C","E"],["S","F","C","S"],["A","D","E","E"]], word = "ABCB" # 输出:false # # :param board: # :param word: # :return: # """ # row_length = len(board) # col_length = len(board[0]) # size = row_length * col_length # # def trace_back(value, k, last_position): # if k >= len(word) or value < 0: # return k # # while value < size: # row = value // col_length # col = value % col_length # if board[row][col] == word[k]: # left = value - 1 # right = value + 1 # up = value - row_length # down = value + row_length # k += 1 # k = trace_back(left, k, 'left') if last_position != 'right' else k # k = trace_back(right, k, 'right') if last_position != 'left' else k # k = trace_back(up, k, 'up') if last_position != 'down' else k # k = trace_back(down, k, 'down') if last_position != 'up' else k # else: # value += 1 # # if k >= len(word): # break # # return k # # pos = trace_back(0, 0, 'None') # return True if pos == len(word) else False def exist(board: List[List[str]], word: str) -> bool: """ 输入:board = [["A","B","C","E"],["S","F","C","S"],["A","D","E","E"]], word = "ABCB" 输出:false :param board: :param word: :return: """ row_length = len(board) col_length = len(board[0]) def trace_back(i, j, k, temp_path): if (i, j) in temp_path: return k - 1 if k >= len(word): return k temp_path.append((i, j)) depth = k if j-1 >= 0 and board[i][j-1] == word[k]: temp_1 = trace_back(i, j-1, k+1, temp_path) depth = temp_1 if temp_1 > depth else depth if j+1 < col_length and board[i][j+1] == word[k]: temp_1 = trace_back(i, j+1, k + 1, temp_path) depth = temp_1 if temp_1 > depth else depth if i-1 >= 0 and board[i-1][j] == word[k]: temp_1 = trace_back(i-1, j, k+1, temp_path) depth = temp_1 if temp_1 > depth else depth if i+1 < row_length and board[i+1][j] == word[k]: temp_1 = trace_back(i+1, j, k+1, temp_path) depth = temp_1 if temp_1 > depth else depth temp_path.pop(-1) return depth for row in range(row_length): for col in range(col_length): if board[row][col] == word[0]: pos = trace_back(row, col, 1, []) if pos >= len(word): return True return False if __name__ == '__main__': board = [["A","B","C","E"],["S","F","C","S"],["A","D","E","E"]] word = "ABCCED" print(exist(board, word))
[ "guan_dongpu@gowild.cn" ]
guan_dongpu@gowild.cn
5db923fb9cbe244877c30b08690665cc6022a117
2b98c555e146487a08b0f9c2e94f2564977623dd
/main.py
14aeb66d5b362c719da26bf8648a29d9ed64f39f
[]
no_license
RossT18/Binary-Tree
400eb9efc79f3c3072a46d6e5c7ae2e37be4940b
5f22c8d76208cfe6283f0bfef573b481002c209e
refs/heads/main
2023-03-27T00:25:06.151084
2021-03-28T17:57:01
2021-03-28T17:57:01
331,387,644
1
0
null
null
null
null
UTF-8
Python
false
false
346
py
import BinaryTree as bt # https://upload.wikimedia.org/wikipedia/commons/d/da/Binary_search_tree.svg root = bt.Node(8) tree = bt.BinaryTree(root) tree.add(3) tree.add(1) tree.add(6) tree.add(4) tree.add(7) tree.add(10) tree.add(14) tree.add(13) # Expected output order: # 1,3,4,6,7,8,10,13,14 print(tree) print("Length: " + str(len(tree)))
[ "rtinsley1199@gmail.com" ]
rtinsley1199@gmail.com
a06a1754d3ec0bbd190006a04d477e5c50c00b20
4e3f83be95d2225fd18047fc6eaf09992d4491ea
/src/utils/Datetimes.py
b01224589683651301f55b46dc3ec0fa6b2561c3
[ "Apache-2.0" ]
permissive
Gabvaztor/TFBoost
7f1d56e44498a23f2e9d884fcd66942cb9e6d622
a37b906f5cb47becc3275def8282ff395d06ef45
refs/heads/master
2021-06-06T07:56:05.995559
2019-10-14T15:27:29
2019-10-14T15:36:43
83,965,500
9
3
null
null
null
null
UTF-8
Python
false
false
119
py
from datetime import datetime def date_from_format(date, format="%Y-%m-%d %H:%M:%S"): return date.strftime(format)
[ "gabvaztor@gmail.com" ]
gabvaztor@gmail.com
7adb46a971ce547c265474004b96ae65283904dc
a0cfde6971fbe3b2c1de726a7bfc1c60fba3a137
/userbot/plugins/git.py
8d3857c9917303b94e1f9cebb82f052502dce43e
[ "MIT" ]
permissive
No-OnE-Kn0wS-Me/FridayUserbot
6efb25dfd2eb06674b99bc158a6bbddcd128012c
9ef60066b72fa085300408855010ea05b9026f82
refs/heads/master
2022-11-19T23:10:49.342679
2020-06-22T13:49:56
2020-06-22T13:49:56
274,447,417
2
1
MIT
2020-06-23T15:52:27
2020-06-23T15:52:26
null
UTF-8
Python
false
false
648
py
from telethon import events import asyncio @borg.on(events.NewMessage(pattern=r"\.(.*)", outgoing=True)) async def _(event): if event.fwd_from: return animation_interval = 0.1 animation_ttl = range(0, 101) input_str = event.pattern_match.group(1) if input_str == "githubs": await event.edit(input_str) animation_chars = [ "https://github.com/midhunkm1294-bit/friday", "https://github.com/midhunkm1294-bit/friday" ] for i in animation_ttl: await asyncio.sleep(animation_interval) await event.edit(animation_chars[i % 2])
[ "66051049+StarkGang@users.noreply.github.com" ]
66051049+StarkGang@users.noreply.github.com
48a149e36e40844f00d5d10b8710b2422b11ab35
4bbb4fca1829ec8b146f9c73d2358b897e28d4ae
/main/views.py
4070691290ea2909d585902532f0a46b40ad40be
[]
no_license
richardwalkerdev/handofcards
e54e1bc4bf868b2a9e02c63a063706daf1573f98
312e77884695e5e1442bff40660b6be08623604b
refs/heads/master
2022-12-16T00:35:24.096302
2020-02-04T17:12:38
2020-02-04T17:12:38
234,924,220
0
0
null
2022-12-08T03:27:57
2020-01-19T15:49:12
Python
UTF-8
Python
false
false
959
py
from django.shortcuts import render import requests import os from django.core.exceptions import ImproperlyConfigured from .models import Hand import datetime def hand(request): def get_env_value(env_variable): try: return os.environ[env_variable] except KeyError: error_msg = 'Set the {} environment variable'.format(env_variable) raise ImproperlyConfigured(error_msg) # Get environmant variable DECKOFCARDS_URL DECKOFCARDS_URL = get_env_value('DECKOFCARDS_URL') response = requests.get(DECKOFCARDS_URL) deckdata = response.json() hand = [] total = 0 for n in range(2): hand.append(deckdata.pop(0)) for i in hand: total = total + i.get("value") handTotal = Hand() handTotal.total = total handTotal.created = datetime.datetime.now() handTotal.save() return render(request, 'main/index.html', {'array': hand, 'total': total})
[ "rwalker@rwalker.remote.csb" ]
rwalker@rwalker.remote.csb
93e2cb9162dfaedfe3a58c9892ccb9936f9405c9
9e7d7b4d029554eed0f760a027cd94558b919ae2
/CHAPTER15/overlaying.py
e320bf396d4410f1a0cc189810fc886ac93deca0
[]
no_license
pooja1506/AutomateTheBoringStuff_2e
8247b68a195d5e1976c6474f0e97d947906ffd35
5bab9ccdcdb22ee10fe1272c91042be40fd67c17
refs/heads/master
2022-04-10T19:21:44.402829
2020-04-05T12:10:32
2020-04-05T12:10:32
249,620,282
1
0
null
null
null
null
UTF-8
Python
false
false
589
py
import PyPDF2 minutesFile = open('meetingminutes.pdf', 'rb') pdfReader = PyPDF2.PdfFileReader(minutesFile) minutesFirstPage = pdfReader.getPage(0) pdfWatermarkReader = PyPDF2.PdfFileReader(open('watermark.pdf', 'rb')) minutesFirstPage.mergePage(pdfWatermarkReader.getPage(0)) pdfWriter = PyPDF2.PdfFileWriter() pdfWriter.addPage(minutesFirstPage) for pageNum in range(1, pdfReader.numPages): pageObj = pdfReader.getPage(pageNum) pdfWriter.addPage(pageObj) resultPdfFile = open('watermarkedCover.pdf', 'wb') pdfWriter.write(resultPdfFile) minutesFile.close() resultPdfFile.close()
[ "pooja.dmehta15@gmail.com" ]
pooja.dmehta15@gmail.com
b3d7305f178eead8c1316698c2989f3d44540d31
71b3715408330a42c62c7176a0f8fb1901d3ba6c
/src/day01/HelloWorld.py
3767e64837fc2b4f8dab04a81c60ccf1cac9530c
[]
no_license
hubin9218/LearnPython
41fac40b4d883fde62c4a0f0405da76da0bda5df
a9fd0e55bddc5f1a2a07212a7f80c603ea2dc735
refs/heads/master
2021-02-17T06:13:48.945425
2020-03-23T09:49:39
2020-03-23T09:49:39
245,076,608
0
0
null
null
null
null
UTF-8
Python
false
false
3,489
py
""" 绘制小猪佩奇... """ from turtle import * def nose(x,y): """画鼻子""" penup() # 将海龟移动到指定的坐标 goto(x,y) pendown() # 设置海龟的方向(0-东、90-北、180-西、270-南) setheading(-30) begin_fill() a = 0.4 for i in range(120): if 0 <= i < 30 or 60 <= i <90: a = a + 0.08 # 向左转3度 left(3) # 向前走 forward(a) else: a = a - 0.08 left(3) forward(a) end_fill() penup() setheading(90) forward(25) setheading(0) forward(10) pendown() # 设置画笔的颜色(红, 绿, 蓝) pencolor(255, 155, 192) setheading(10) begin_fill() circle(5) color(160, 82, 45) end_fill() penup() setheading(0) forward(20) pendown() pencolor(255, 155, 192) setheading(10) begin_fill() circle(5) color(160, 82, 45) end_fill() def head(x, y): """画头""" color((255, 155, 192), "pink") penup() goto(x,y) setheading(0) pendown() begin_fill() setheading(180) circle(300, -30) circle(100, -60) circle(80, -100) circle(150, -20) circle(60, -95) setheading(161) circle(-300, 15) penup() goto(-100, 100) pendown() setheading(-30) a = 0.4 for i in range(60): if 0<= i < 30 or 60 <= i < 90: a = a + 0.08 lt(3) #向左转3度 fd(a) #向前走a的步长 else: a = a - 0.08 lt(3) fd(a) end_fill() def ears(x,y): """画耳朵""" color((255, 155, 192), "pink") penup() goto(x, y) pendown() begin_fill() setheading(100) circle(-50, 50) circle(-10, 120) circle(-50, 54) end_fill() penup() setheading(90) forward(-12) setheading(0) forward(30) pendown() begin_fill() setheading(100) circle(-50, 50) circle(-10, 120) circle(-50, 56) end_fill() def eyes(x,y): """画眼睛""" color((255, 155, 192), "white") penup() setheading(90) forward(-20) setheading(0) forward(-95) pendown() begin_fill() circle(15) end_fill() color("black") penup() setheading(90) forward(12) setheading(0) forward(-3) pendown() begin_fill() circle(3) end_fill() color((255, 155, 192), "white") penup() seth(90) forward(-25) seth(0) forward(40) pendown() begin_fill() circle(15) end_fill() color("black") penup() setheading(90) forward(12) setheading(0) forward(-3) pendown() begin_fill() circle(3) end_fill() def cheek(x,y): """画脸颊""" color((255, 155, 192)) penup() goto(x,y) pendown() setheading(0) begin_fill() circle(30) end_fill() def mouth(x,y): """画嘴巴""" color(239, 69, 19) penup() goto(x, y) pendown() setheading(-80) circle(30, 40) circle(40, 80) def setting(): """设置参数""" pensize(4) # 隐藏海龟 hideturtle() colormode(255) color((255, 155, 192), "pink") setup(840, 500) speed(10) def main(): """主函数""" setting() nose(-100, 100) head(-69, 167) ears(0, 160) eyes(0, 140) cheek(80, 10) mouth(-20, 30) done() if __name__ == '__main__': main()
[ "hubin9218@163.com" ]
hubin9218@163.com
5a4ce30950656174d6f9d7daeedd4c6f6fb567c7
7e83e6b4e1f6c461e8b77eb620a9a1c851b611e8
/hello_django/hello_django/urls.py
f5dc4fe5f9d8a8d42a862b0098131600a0887911
[]
no_license
guess3233qa/opendata
fa64e6331b1c1f08972d44a339966444dc45ff96
b3d95ce69ce6cfd23947a2f32b7ab2f058eb7ed2
refs/heads/master
2021-01-17T05:17:38.871841
2017-06-28T01:06:27
2017-06-28T01:06:27
95,610,418
0
0
null
null
null
null
UTF-8
Python
false
false
818
py
"""hello_django URL Configuration The `urlpatterns` list routes URLs to views. For more information please see: https://docs.djangoproject.com/en/1.11/topics/http/urls/ Examples: Function views 1. Add an import: from my_app import views 2. Add a URL to urlpatterns: url(r'^$', views.home, name='home') Class-based views 1. Add an import: from other_app.views import Home 2. Add a URL to urlpatterns: url(r'^$', Home.as_view(), name='home') Including another URLconf 1. Import the include() function: from django.conf.urls import url, include 2. Add a URL to urlpatterns: url(r'^blog/', include('blog.urls')) """ from django.conf.urls import url , include from django.contrib import admin urlpatterns = [ url(r'^admin/', admin.site.urls), url(r'' , include('hello.urls')), ]
[ "noreply@github.com" ]
guess3233qa.noreply@github.com
cce85590e54ae1deb81edc2cb58d4281f38508e8
2d27f9735aa104cffb0701683d70021d61729f57
/template.py
8d2e3a024ba2f3223f9cdd13efe292b2c6176d09
[]
no_license
indexample/AOC2020
2b84f07fd479c4fc7df0b1b3c2b65176fcf6a7de
300363a693e933173f3a5edfc3bc1797428f0346
refs/heads/main
2023-02-06T03:19:43.667864
2020-12-22T13:04:18
2020-12-22T13:04:18
318,496,610
0
0
null
null
null
null
UTF-8
Python
false
false
1,054
py
import os import timeit #import numpy as np #from itertools import count # ideas # part1: # part2: def log(*args): if LOGGING: for i in args: print( str(i) ) def timefunc(iter, function, *args): def wrap(): function(*args) t = timeit.Timer(wrap) return t.timeit(iter) / iter #average def solve1(d): pass LOGGING = 1 f_loc = 'D:/GIT/AOC2020-1/day14/input.txt' #set = {}, list = [], generator = () #data = [int(x) for line in open(f_loc, 'r').read().rstrip().split("\n") for x in line.split(',') if x != 'x' ] #or read().splitlines() #data = [x for line in open(f_loc, 'r').read().rstrip().split("\n") for x in line.split(',') ] data = [line for line in open(f_loc, 'r').read().rstrip().split("\n") ] #data = list(map(char_to_int, open(f_loc, 'r').readlines())) #i = dict(enumerate(data)) print('\n---- part 1 ----') print(f': {solve1(data)}') print('\n---- part 2 ----') #print(f': {solve2(data)}') # timeit #print(f'timeit: {timefunc(10, solve1, data)}' ) # part 1: # part 2:
[ "3052927+devale@users.noreply.github.com" ]
3052927+devale@users.noreply.github.com
03295f3478f5d7e47e2204ea64855e9de3926e54
121ab18bfbdb690a3365e84b7da6d1b143fb8768
/tabTools/add_tag.py
aacad3786e6de8bf2b6c20cec752c44ac4347af1
[]
no_license
zy041225/python_bin
1661ef120ce22d63f7f8929dfe637bf12e7b31b4
4b0219b4a99394b48c3dad19727cbfb2038d2e3e
refs/heads/master
2020-09-13T15:34:02.145545
2019-11-20T02:07:34
2019-11-20T02:07:34
222,830,326
0
0
null
null
null
null
UTF-8
Python
false
false
552
py
#!/usr/bin/env python3 import sys import getopt if len(sys.argv) != 6 or sys.argv[5] == '-': sys.exit("python3 %s <tab> <tab_column> <tag_lst> <tag_column> <tag_name>\n<tag_name> should NOT be '-'"% (sys.argv[0])) dict = {} tag_lst = open(sys.argv[3]) for line in tag_lst: line = line.rstrip() tmp = line.split('\t') dict[tmp[int(sys.argv[4])-1]] = 1 tab = open(sys.argv[1]) for line in tab: line = line.rstrip() tmp = line.split('\t') tag = '-' if tmp[int(sys.argv[2])-1] in dict: tag = sys.argv[5] line += '\t%s' % (tag) print(line)
[ "zy041225@gmail.com" ]
zy041225@gmail.com
d8db374595ea9b2c375ef52a5364a9fa9f258336
626a08cf643775368dff313e1049ba55d559e986
/libs/metrics/type_composite_index.py
01973da142755b0460a52ac3eddbd4907af9cfba
[]
no_license
NashLea/SecuritiesAnalysisTools
3fd995a4d4d714cff81cd60cb6f885880c175d19
3fd5ae12714f56efd5dc395ae7a1e5acc7778aba
refs/heads/master
2022-11-06T06:43:09.399530
2020-06-20T03:27:22
2020-06-20T03:27:22
null
0
0
null
null
null
null
UTF-8
Python
false
false
9,056
py
import os import json import pandas as pd import numpy as np from libs.tools import cluster_oscs from libs.tools import windowed_moving_avg from libs.utils import download_data_indexes from libs.utils import dual_plotting, generic_plotting from libs.utils import ProgressBar, index_appender from libs.utils import STANDARD_COLORS ERROR_COLOR = STANDARD_COLORS["error"] WARNING = STANDARD_COLORS["warning"] NORMAL = STANDARD_COLORS["normal"] def type_composite_index(**kwargs) -> list: """Type Composite Index (MCI) Similar to MCI, TCI compares broader market types (sensitive, cyclical, and defensive) Optional Args: config {dict} -- controlling config dictionary (default: {None}) plot_output {bool} -- True to render plot in realtime (default: {True}) period {str / list} -- time period for data (e.g. '2y') (default: {None}) clock {float} -- time for prog_bar (default: {None}) data {pd.DataFrame} -- fund datasets (default: {None}) sectors {list} -- list of sectors (default: {None}) returns: list -- dict contains all tci information, data, sectors """ config = kwargs.get('config') period = kwargs.get('period') plot_output = kwargs.get('plot_output', True) clock = kwargs.get('clock') data = kwargs.get('data') sectors = kwargs.get('sectors') if config is not None: period = config['period'] properties = config['properties'] elif period is None: print( f"{ERROR_COLOR}ERROR: config and period both provided {period} " + f"for type_composite_index{NORMAL}") return {} else: # Support for release 1 versions period = period properties = dict() properties['Indexes'] = {} properties['Indexes']['Type Sector'] = True # Validate each index key is set to True in the --core file if properties is not None: if 'Indexes' in properties.keys(): props = properties['Indexes'] if 'Type Sector' in props.keys(): if props['Type Sector'] == True: m_data = get_metrics_content() if data is None or sectors is None: data, sectors = metrics_initializer( m_data, period='2y') if data: p = ProgressBar( 19, name='Type Composite Index', offset=clock) p.start() tci = dict() composite = {} for sect in sectors: cluster = cluster_oscs( data[sect], plot_output=False, function='market', wma=False, progress_bar=p ) graph = cluster['tabular'] composite[sect] = graph defensive = type_composites( composite, m_data, type_type='Defensive') p.uptick() sensitive = type_composites( composite, m_data, type_type='Sensitive') p.uptick() cyclical = type_composites( composite, m_data, type_type='Cyclical') p.uptick() d_val = weighted_signals( data, m_data, type_type='Defensive') p.uptick() s_val = weighted_signals( data, m_data, type_type='Sensitive') p.uptick() c_val = weighted_signals( data, m_data, type_type='Cyclical') p.uptick() d_val = windowed_moving_avg(d_val, 3, data_type='list') c_val = windowed_moving_avg(c_val, 3, data_type='list') s_val = windowed_moving_avg(s_val, 3, data_type='list') p.uptick() tci['defensive'] = { "tabular": d_val, "clusters": defensive } tci['sensitive'] = { "tabular": s_val, "clusters": sensitive } tci['cyclical'] = { "tabular": c_val, "clusters": cyclical } dates = data['VGT'].index if plot_output: dual_plotting(y1=d_val, y2=defensive, y1_label='Defensive Index', y2_label='Clustered Osc', title='Defensive Index', x=dates) dual_plotting(y1=s_val, y2=sensitive, y1_label='Sensitive Index', y2_label='Clustered Osc', title='Sensitive Index', x=dates) dual_plotting(y1=c_val, y2=cyclical, y1_label='Cyclical Index', y2_label='Clustered Osc', title='Cyclical Index', x=dates) generic_plotting([d_val, s_val, c_val], legend=[ 'Defensive', 'Sensitive', 'Cyclical'], title='Type Indexes', x=dates) else: generic_plotting( [d_val, s_val, c_val], legend=['Defensive', 'Sensitive', 'Cyclical'], title='Type Indexes', x=dates, saveFig=True, ylabel='Normalized "Price"', filename='tci.png' ) p.end() return tci, data, sectors return {}, None, None def metrics_initializer(m_data: dict, period='2y'): """Metrics Initializer Keyword Arguments: period {str} -- (default: {'2y'}) Returns: list -- downloaded_data, sector_list, index, metrics_file data """ sectors = m_data['Components'] tickers = " ".join(sectors) tickers = index_appender(tickers) all_tickers = tickers.split(' ') if isinstance(period, (list)): period = period[0] # tickers = index_appender(tickers) print(" ") print(f'Fetching Type Composite Index funds for {period}...') data, _ = download_data_indexes( indexes=sectors, tickers=all_tickers, period=period, interval='1d') print(" ") return data, sectors def get_metrics_content() -> dict: """Get Metrics Content Returns: dict -- metrics file data """ metrics_file = os.path.join("resources", "sectors.json") if not os.path.exists(metrics_file): print( f"{WARNING}WARNING: '{metrics_file}' not found for " + f"'metrics_initializer'. Failed.{NORMAL}") return None, [], None with open(metrics_file) as m_file: m_data = json.load(m_file) m_file.close() m_data = m_data.get("Type_Composite") return m_data def type_composites(composite: dict, m_data: dict, type_type='Defensive') -> list: """Type Composites Create the summed clustered composites Arguments: composite {dict} -- composite dictionary m_data {dict} -- data from sectors.json Keyword Arguments: type_type {str} -- key for each m_data (default: {'Defensive'}) Returns: list -- summed list of composites """ sector_data = m_data[type_type] start_key = list(sector_data.keys())[0] new_composite = [] for i in range(len(composite[start_key])): value = 0.0 for fund in sector_data: value += float(composite[fund][i]) * sector_data[fund] new_composite.append(value) return new_composite def weighted_signals(data: dict, m_data: dict, type_type='Defensive') -> list: """Weighted Signals Arguments: data {dict} -- tci data object m_data {dict} -- sectors.json content Keyword Arguments: type_type {str} -- (default: {'Defensive'}) Returns: list -- weighted signal """ sector_data = m_data[type_type] start_key = list(sector_data.keys())[0] new_composite = [25.0] for i in range(1, len(data[start_key]['Close'])): value = 0.0 for fund in sector_data: value += (data[fund]['Close'][i] - data[fund]['Close'][i-1]) /\ data[fund]['Close'][i-1] * sector_data[fund] value = new_composite[-1] * (1.0 + value) new_composite.append(value) return new_composite
[ "ngamell@mmm.com" ]
ngamell@mmm.com
da60ebdb0c9f8856df41a22fbe7a5925c0a77927
a4d250ce393012dc251cb955096cb8f284c57439
/gunion/data/battle.py
2e86c5bc5fb1e0344791a01e941d7330c554ed7e
[]
no_license
daxingyou/test-2
b02af312784d06a46e29acd42e756e92afee6ed9
a16c872ba781855a8c891eff41e8e651cd565ebf
refs/heads/master
2023-03-16T04:21:23.704482
2018-09-28T00:51:23
2018-09-28T00:51:23
null
0
0
null
null
null
null
UTF-8
Python
false
false
11,188
py
#coding:utf8 """ Created on 2016-07-28 @Author: jiangtaoran(jiangtaoran@ice-time.cn) @Brief : 联盟战争 """ import base64 from utils import logger from utils import utils from datalib.data_loader import data_loader class UnionBattleInfo(object): """一场联盟战争 """ BATTLE_STAGE_INVALID = 0 #非法 BATTLE_STAGE_IDLE = 1 #无战争 BATTLE_STAGE_PREPARE = 2 #备战阶段 BATTLE_STAGE_FIGHT = 3 #战斗阶段 BATTLE_STAGE_CLOSE = 4 #结束 __slots__ = [ "id", "union_id", "index", "stage", "rival_union_id", "rival_battle_id", "is_initiator", #是否战争发起方 "launch_time", #战争发起时间 "fight_time", #战争开战时间 "close_time", #战争结束时间 "finish_time", #战争生命周期终止时间,可以开始下一场战争 "is_deployed", #是否已经完成防御部署 "battle_count", #战斗的数量 "score", #胜场积分 "individuals_score", #成员战功之和 "drum", "attack_level", "attack_win_count_this_level", #本轮战斗中攻击胜利次数 "attack_lose_count_this_level", #本轮战斗中攻击失败次数 "defend_nodes_level", #防守方的节点level "record_index", "accepted_members", #大宝箱领取的记录 "accepted_names", "accepted_icons", "reward_items", "reward_nums", "accept_times" ] def __init__(self): self.id = 0 self.union_id = 0 self.index = 0 self.stage = UnionBattleInfo.BATTLE_STAGE_INVALID self.rival_union_id = 0 self.rival_battle_id = 0 self.is_initiator = False self.launch_time = 0 self.fight_time = 0 self.close_time = 0 self.finish_time = 0 self.is_deployed = False self.battle_count = 0 self.score = 0 self.individuals_score = 0 self.drum = 0 self.attack_level = 1 self.attack_win_count_this_level = 0 self.attack_lose_count_this_level = 0 self.defend_nodes_level = "" self.record_index = 0 @staticmethod def generate_id(union_id, index): id = union_id << 32 | index return id @staticmethod def create(union_id, index, invalid): battle = UnionBattleInfo() battle.id = UnionBattleInfo.generate_id(union_id, index) battle.union_id = union_id battle.index = index if invalid: battle.stage = UnionBattleInfo.BATTLE_STAGE_INVALID #无法发起战争 else: battle.stage = UnionBattleInfo.BATTLE_STAGE_IDLE battle.rival_union_id = 0 battle.rival_battle_id = 0 battle.is_initiator = False battle.launch_time = 0 battle.fight_time = 0 battle.close_time = 0 battle.finish_time = 0 battle.is_deployed = False battle.battle_count = 0 battle.score = 0 battle.individuals_score = 0 battle.drum = 0 battle.attack_level = 1 battle.attack_win_count_this_level = 0 battle.attack_lose_count_this_level = 0 battle.defend_nodes_level = "" battle.record_index = 0 battle.accepted_members = "" #奖励箱领取 battle.accepted_names = "" battle.accepted_icons = "" battle.reward_items = "" battle.reward_nums = "" battle.accept_times = "" return battle def force_update_fight_time(self, time): """强制改变开战时间(内部接口) """ assert self.stage == self.BATTLE_STAGE_PREPARE self.fight_time = time def force_update_close_time(self, time): """强制改变结束战斗时间(内部接口) """ assert self.stage == self.BATTLE_STAGE_FIGHT self.close_time = time def force_update_finish_time(self, time): """强制改变结束时间(内部接口) """ assert self.stage == self.BATTLE_STAGE_CLOSE self.finish_time = time def is_able_to_join(self): """是否可以参战 """ return self.stage == self.BATTLE_STAGE_IDLE def is_able_to_deploy(self): """是否可以部署防御 """ return self.stage == self.BATTLE_STAGE_PREPARE def is_able_to_drum(self): """是否可以擂鼓 """ return self.stage == self.BATTLE_STAGE_FIGHT def is_at_war(self): """是否在交战中 """ return (self.stage == self.BATTLE_STAGE_PREPARE or self.stage == self.BATTLE_STAGE_FIGHT or self.stage == self.BATTLE_STAGE_CLOSE) def launch(self, now, rival_union_id, rival_battle_id, initiative = True): """发起战争 """ assert self.stage == self.BATTLE_STAGE_IDLE self.stage = self.BATTLE_STAGE_PREPARE self.rival_union_id = rival_union_id self.rival_battle_id = rival_battle_id self.is_initiator = initiative self.launch_time = now #self.fight_time = utils.get_spec_second( # self.launch_time, "22:30") + utils.SECONDS_OF_DAY #launch time 次日22:30 #self.close_time = self.fight_time + utils.SECONDS_OF_DAY #fight time 次日22:30 #self.finish_time = utils.get_spec_second( # self.close_time, "05:00" ) + utils.SECONDS_OF_DAY #close time 次日05:00 self.fight_time = utils.get_spec_second(self.launch_time, "21:00") #fight time 当日21:00 self.close_time = utils.get_spec_second(self.launch_time, "23:00") #close time 当日23:00 self.finish_time = utils.get_spec_second(self.launch_time, "05:00" ) + utils.SECONDS_OF_DAY #finish time 次日05:00 self.is_deployed = False self.accepted_members = "" self.accepted_names = "" self.accepted_icons = "" self.reward_items = "" self.reward_nums = "" self.accept_times = "" def start_fight(self, now): """进入开战阶段 """ assert self.stage == self.BATTLE_STAGE_PREPARE self.stage = self.BATTLE_STAGE_FIGHT self.is_deployed = True self.attack_level = 1 def is_fight_closed(self, now): """战斗结算是否结束 """ return self.launch_time != 0 and now >= self.close_time def close_fight(self): """战争结束 """ #assert self.stage == self.BATTLE_STAGE_FIGHT self.stage = self.BATTLE_STAGE_CLOSE def is_finished(self, now): """战争是否结束 """ return self.launch_time != 0 and now >= self.finish_time and now >= self.close_time def is_able_to_start(self): """是否可以开战 """ return self.stage == self.BATTLE_STAGE_FIGHT def beat_drum(self, value = 1): """擂鼓 """ assert value >= 0 self.drum += value def get_attack_buff_count(self): """获取当前攻击 buff 加成 """ drum_ratio = int(float( data_loader.UnionConfInfo_dict["attack_buff_count_per_drum"].value)) lose_ratio = int(float( data_loader.UnionConfInfo_dict["attack_buff_count_per_lose"].value)) return self.drum * drum_ratio + self.attack_lose_count_this_level * lose_ratio def get_attack_buff_temporary_count(self): """获取当前轮次临时攻击 buff 加成 """ lose_ratio = int(float( data_loader.UnionConfInfo_dict["attack_buff_count_per_lose"].value)) return self.attack_lose_count_this_level * lose_ratio def mark_attack_result(self, win): """记录攻击结果 """ if win: self.attack_win_count_this_level += 1 #else: # self.attack_lose_count_this_level += 1 #攻击进入下一轮 count = int(float(data_loader.UnionConfInfo_dict["battle_map_node_count"].value)) if self.attack_win_count_this_level >= count: self.attack_level += 1 self.attack_win_count_this_level = 0 self.attack_lose_count_this_level = 0 def gain_union_score(self, value = 1): """增加联盟胜场积分 """ assert value >= 0 self.score += value def gain_individuals_score(self, value): """增加成员战功点数 """ assert value >= 0 self.individuals_score += value def get_next_record_index(self): """获取下一个战斗记录 index """ self.record_index += 1 return self.record_index def is_able_to_accept_box(self): """是否可以领取大宝箱/ """ if self.stage != self.BATTLE_STAGE_CLOSE: return False level = int(float(data_loader.UnionConfInfo_dict["battle_map_total_level"].value)) count = int(float(data_loader.UnionConfInfo_dict["battle_map_node_count"].value)) if level * count > self.score: return False return True def get_accepted_members(self): """获取领取过奖励的成员""" return utils.split_to_int(self.accepted_members) def get_reward_record(self): """获取奖励领取记录""" members = utils.split_to_int(self.accepted_members) names = utils.split_to_string(self.accepted_names) icons = utils.split_to_int(self.accepted_icons) items_id = utils.split_to_int(self.reward_items) items_num = utils.split_to_int(self.reward_nums) times = utils.split_to_int(self.accept_times) names = [base64.b64decode(name) for name in names] return map(None, members, names, icons, items_id, items_num, times) def add_reward_record(self, user_id, user_name, icon_id, item_id, item_num, now): """添加领奖记录""" members = utils.split_to_int(self.accepted_members) names = utils.split_to_string(self.accepted_names) icons = utils.split_to_int(self.accepted_icons) items_id = utils.split_to_int(self.reward_items) items_num = utils.split_to_int(self.reward_nums) times = utils.split_to_int(self.accept_times) members.append(user_id) names.append(user_name) icons.append(icon_id) items_id.append(item_id) items_num.append(item_num) times.append(now) self.accepted_members = utils.join_to_string(members) self.accepted_names = utils.join_to_string(names) self.accepted_icons = utils.join_to_string(icons) self.reward_items = utils.join_to_string(items_id) self.reward_nums = utils.join_to_string(items_num) self.accept_times = utils.join_to_string(times)
[ "luhongwei1@ice-time.cn" ]
luhongwei1@ice-time.cn
64c56cb8b06a7ea97fa9eaa40bc7a4d99d330d48
11b0c124262ac40de87d756389082462d8452e4d
/keras/keras44_3_cancer_conv1d.py
7aaef711d66a0e504dbd364159570a7de778fe5d
[]
no_license
Hyunwoo29/keras01
164b519f9bb70d55f7bfa91c66529cee4d012b24
494611a94420f8e268c37085ccf0b0de4aa71048
refs/heads/main
2023-08-06T07:19:13.288634
2021-10-05T14:03:56
2021-10-05T14:03:56
383,744,859
0
0
null
null
null
null
UTF-8
Python
false
false
2,188
py
from sklearn.preprocessing import MaxAbsScaler, RobustScaler, QuantileTransformer, PowerTransformer, StandardScaler from tensorflow.keras.callbacks import EarlyStopping import numpy as np from sklearn.datasets import load_breast_cancer from icecream import ic from tensorflow.keras.models import Sequential from tensorflow.keras.layers import Dense, Dropout, LSTM, Conv1D, Flatten from tensorflow.keras.utils import to_categorical from sklearn.model_selection import train_test_split import time from sklearn.metrics import r2_score #1. 데이터 datasets = load_breast_cancer() x = datasets.data y = datasets.target ic(x, y) # ic| x.shape: (569, 30), y.shape: (569,) x_train, x_test, y_train, y_test = train_test_split(x,y, train_size=0.7, random_state=60) # train 309, test 133 # scaler = QuantileTransformer() scaler = StandardScaler() # scaler = PowerTransformer() # scaler = RobustScaler() x_train = scaler.fit_transform(x_train) x_test = scaler.transform(x_test) x_train = x_train.reshape(x_train.shape[0], x_train.shape[1], 1) x_test = x_test.reshape(x_test.shape[0], x_test.shape[1], 1) # ic(x_train.shape, x_test.shape) # ic(np.unique(y)) model = Sequential() model.add(LSTM(40, activation='relu', input_shape=(30,1), return_sequences=True)) model.add(Conv1D(128, 2)) model.add(Dense(128, activation='relu')) model.add(Dropout(0.2)) model.add(Dense(64, activation='relu')) model.add(Dense(32, activation='relu')) model.add(Dropout(0.2)) model.add(Dense(16, activation='relu')) model.add(Flatten()) model.add(Dense(1,activation='sigmoid')) model.summary() #3. 컴파일, 훈련 model.compile(loss='binary_crossentropy', optimizer='adam', metrics=['acc']) es = EarlyStopping(monitor='acc', patience=20, mode='auto', verbose=1) start = time.time() model.fit(x_train, y_train, epochs=100, verbose=1, validation_split=0.2, batch_size=32, shuffle=True, callbacks=[es]) 걸린시간 = round((time.time() - start) /60,1) #4. 평가, 예측 y_predict = model.predict(x_test) loss = model.evaluate(x_test, y_test) ic(loss[0]) ic(loss[1]) ic(f'{걸린시간}분') # ic| loss[0]: 0.028651483356952667 # ic| loss[1]: 0.988304078578949 # ic| f'{걸린시간}분': '0.8분'
[ "nbaksa3@gamil.com" ]
nbaksa3@gamil.com
da4d9970097abb9879bdaf10f8d859c5287053b0
5b8fcb1bf82a7c1ef5b6c2a939b1d1597bc7a24b
/create_json_for_airtable_operator.py
e00238b2f39eae43c6d55eae4974dcf2d194d262
[]
no_license
katerinekhh/airflow_custom_stuff
2420c3ee95dab01e5eeeb8248500e253126e5b48
43ba78d96770a575ba7ab11a691b101e6d6604af
refs/heads/master
2022-10-12T13:55:01.916266
2020-06-12T13:17:06
2020-06-12T13:17:06
271,645,308
0
0
null
null
null
null
UTF-8
Python
false
false
2,245
py
from datetime import datetime import json from airflow.utils.decorators import apply_defaults from airflow.models.baseoperator import BaseOperator from airflow.hooks.http_hook import HttpHook class CreateJsonForAirtableOperator(BaseOperator): @apply_defaults def __init__( self, endpoint: str, http_conn_id: str, message_id_filepath: str, update_filepath: str, method='GET', request_params=None, *args, **kwargs, ): super().__init__(*args, **kwargs) self.http_conn_id = http_conn_id self.request_params = request_params or {} self.endpoint = endpoint self.message_id_filepath = message_id_filepath self.update_filepath = update_filepath self.hook = HttpHook( method=method, http_conn_id=http_conn_id) def execute(self, context): response = self.hook.run(self.endpoint, data=self.request_params) with open(self.message_id_filepath, 'r') as id_file: message_id = id_file.read() json_response = json.loads(response.text) airtable_updates_data = {} airtable_updates_data['records'] = [] for update in json_response['result']: update_data_fields = {} update_data = {} if update['callback_query']['message']['message_id'] == int(message_id): chat_id = update['callback_query']['message']['chat']['id'] username = update['callback_query']['from']['username'] triggered_at = datetime.fromtimestamp( update['callback_query']['message']['date']).isoformat()[:-3] + "Z" update_data['chat_id'] = chat_id update_data['username'] = username update_data['triggered_at'] = triggered_at update_data['event_type'] = 'push_go_button' update_data['reporter_name'] = 'khkaterina' update_data_fields['fields'] = update_data airtable_updates_data['records'].append(update_data_fields) with open(self.update_filepath, 'w') as file: json.dump(airtable_updates_data, file)
[ "you@example.com" ]
you@example.com
cf4cd861229bcd84028e5670c4292b18a9ce0692
e0cf5219ff9ad4eab2000516739ee651d7aa4c8f
/models/nets.py
ca62b8e0f15aec8c1b12dff278ba3f28a6d8c6c6
[ "MIT" ]
permissive
Perfec-Yu/Lifelong-ED
cbf32f6e2d3ccf393eec08e5dbfb29e5e3c1b28b
f1af49129dd6ed4ff545f84e680565cccdb5b55a
refs/heads/main
2023-07-19T23:52:41.932196
2021-09-02T15:26:36
2021-09-02T15:26:36
402,468,387
13
2
null
null
null
null
UTF-8
Python
false
false
32,092
py
import numpy import torch import torch.nn as nn import torch.nn.functional as F from torch import autograd import math from typing import Any, Dict, Tuple, List, Union, Set import warnings from collections import OrderedDict from torch.nn.modules.linear import Linear from torchmeta.modules import MetaLinear, MetaSequential, MetaModule from tqdm import tqdm class LInEx(MetaModule): def __init__(self,input_dim:int,hidden_dim:int,max_slots:int,init_slots:int,device:Union[torch.device, None]=None,**kwargs)->None: super().__init__() if input_dim != hidden_dim: self.input_map = MetaSequential(OrderedDict({ "linear_0": MetaLinear(input_dim, hidden_dim), "relu_0": nn.ReLU(), "dropout_0": nn.Dropout(0.2), "linear_1": MetaLinear(hidden_dim, hidden_dim), "relu_1": nn.ReLU() })) else: self.input_map = lambda x:x self.classes = MetaLinear(hidden_dim, max_slots, bias=False) _mask = torch.zeros(1, max_slots, dtype=torch.float, device=device) _mask[:, init_slots:] = float("-inf") self.register_buffer(name="_mask", tensor=_mask) self.crit = nn.CrossEntropyLoss() self.device = device self.to(device=device) self.nslots = init_slots self.max_slots = max_slots self.maml = True self.outputs = {} self.history = None self.exemplar_features = None self.exemplar_labels = None self.dev_exemplar_features = None self.dev_exemplar_labels = None @property def mask(self,): self._mask[:, :self.nslots] = 0 self._mask[:, self.nslots:] = float("-inf") return self._mask def idx_mask(self, idx:Union[torch.LongTensor, int, List[int], None]=None, max_idx:Union[torch.LongTensor, int, None]=None): assert (idx is not None) or (max_idx is not None) assert (idx is None) or (max_idx is None) mask = torch.zeros_like(self._mask) + float("-inf") if idx is not None: mask[:, idx] = 0 if max_idx is not None: if isinstance(max_idx, torch.LongTensor): max_idx = max_idx.item() mask[:, :max_idx] = 0 return mask @property def features(self): return self.classes.weight[:self.nslots] def forward(self, batch, nslots:int=-1, exemplar:bool=False, exemplar_distill:bool=False, feature_distill:bool=False, mul_distill=False, distill:bool=False, return_loss:bool=True, return_feature:bool=False, tau:float=1.0, log_outputs:bool=True, params=None): if isinstance(batch, (tuple, list)) and len(batch) == 2: features, labels = batch else: features, labels = batch.features, batch.labels inputs = self.input_map(features, params=self.get_subdict(params, "input_map")) scores = self.classes(inputs, params=self.get_subdict(params, "classes")) if torch.any(torch.isnan(scores)): print(scores[0]) input('a') if nslots == -1: scores += self.mask if torch.any(torch.isnan(scores)): print(scores[0]) input() nslots = self.nslots else: scores += self.idx_mask(max_idx=nslots) scores[:, 0] = 0 if scores.size(0) != labels.size(0): assert scores.size(0) % labels.size(0) == 0 labels = labels.repeat_interleave(scores.size(0) // labels.size(0), dim=0) else: labels = labels if log_outputs: pred = torch.argmax(scores, dim=1) acc = torch.mean((pred == labels).float()) self.outputs["accuracy"] = acc.item() self.outputs["prediction"] = pred.detach().cpu() self.outputs["label"] = labels.detach().cpu() self.outputs["input_features"] = features.detach().cpu() self.outputs["encoded_features"] = inputs.detach().cpu() if return_loss: labels.masked_fill_(labels >= nslots, 0) valid = labels < nslots nvalid = torch.sum(valid.float()) if nvalid == 0: loss = 0 else: loss = self.crit(scores[valid], labels[valid]) if torch.isnan(loss): print(labels, nslots, scores[:, :nslots]) input() if distill and self.history is not None: old_scores, old_inputs = self.forward(batch, nslots=self.history["nslots"], return_loss=False, log_outputs=False, return_feature=True, params=self.history["params"]) old_scores = old_scores.detach() old_inputs = old_inputs.detach() new_scores = scores[:, :self.history["nslots"]] if mul_distill: loss_distill = - torch.sum(torch.softmax(old_scores*tau, dim=1) * torch.log_softmax(new_scores*tau, dim=1), dim=1).mean() old_dist = torch.softmax(old_scores/tau, dim=1) old_valid = (old_dist[:, 0] < 0.9) old_num = torch.sum(old_valid.float()) if old_num > 0: # print(old_dist[old_valid].topk(5, dim=1), batch.labels[old_valid]) # input() loss_mul_distill = - torch.sum(old_dist[old_valid] * torch.log_softmax(new_scores[old_valid], dim=1), dim=1).sum() loss_distill = (loss_distill * old_dist.size(0) + loss_mul_distill) / (old_dist.size(0) + old_num) else: loss_distill = - torch.sum(torch.softmax(old_scores*tau, dim=1) * torch.log_softmax(new_scores*tau, dim=1), dim=1).mean() if feature_distill: loss_f_distill = (1 - (old_inputs / old_inputs.norm(dim=-1, keepdim=True) * inputs / inputs.norm(dim=-1, keepdim=True)).sum(dim=-1)).mean(dim=0) loss_distill += loss_f_distill d_weight = self.history["nslots"] c_weight = (self.nslots - self.history["nslots"]) loss = ( d_weight * loss_distill+ c_weight* loss) / (d_weight+c_weight) if torch.isnan(loss): print(old_scores, new_scores) input() if exemplar and self.exemplar_features is not None: if self.exemplar_features.size(0) < 128: exemplar_inputs = self.input_map(self.exemplar_features.to(self.device), params=self.get_subdict(params, "input_map")) exemplar_scores = self.classes(exemplar_inputs, params=self.get_subdict(params, "classes")) else: exemplar_scores = [] exemplar_inputs = [] for _beg in range(0, self.exemplar_features.size(0), 128): _features = self.exemplar_features[_beg:_beg+128, :] _inputs = self.input_map(_features.to(self.device), params=self.get_subdict(params, "input_map")) exemplar_inputs.append(_inputs) exemplar_scores.append(self.classes(_inputs, params=self.get_subdict(params, "classes"))) exemplar_inputs = torch.cat(exemplar_inputs, dim=0) exemplar_scores = torch.cat(exemplar_scores, dim=0) exemplar_scores[:, 0] = 0. loss_exemplar = self.crit(exemplar_scores+self.mask, self.exemplar_labels.to(self.device)) if torch.isnan(loss_exemplar): print(self.exemplar_labels, nslots) input() if exemplar_distill: if self.exemplar_features.size(0) < 128: exemplar_old_inputs = self.input_map(self.exemplar_features.to(self.device), params=self.get_subdict(self.history["params"], "input_map")) exemplar_old_scores = self.classes(exemplar_old_inputs, params=self.get_subdict(self.history["params"], "classes")) else: exemplar_old_scores = [] exemplar_old_inputs = [] for _beg in range(0, self.exemplar_features.size(0), 128): _features = self.exemplar_features[_beg:_beg+128, :] _inputs = self.input_map(_features.to(self.device), params=self.get_subdict(self.history["params"], "input_map")) exemplar_old_inputs.append(_inputs) exemplar_old_scores.append(self.classes(_inputs, params=self.get_subdict(self.history["params"], "classes"))) exemplar_old_inputs = torch.cat(exemplar_old_inputs, dim=0) exemplar_old_scores = torch.cat(exemplar_old_scores, dim=0) exemplar_old_scores[:, 0] = 0. exemplar_old_scores = exemplar_old_scores[:self.history["nslots"]] loss_exemplar_distill = - torch.sum(torch.softmax(exemplar_old_scores[:self.history["nslots"]]*tau, dim=1) * torch.log_softmax(exemplar_scores[:self.history["nslots"]], dim=1), dim=1).mean() if feature_distill: loss_exemplar_feat_distill = (1 - (exemplar_old_inputs / exemplar_old_inputs.norm(dim=-1, keepdim=True) * exemplar_inputs / exemplar_inputs.norm(dim=-1, keepdim=True)).sum(dim=-1)).mean(dim=0) loss_exemplar_distill += loss_exemplar_feat_distill d_weight = self.history["nslots"] c_weight = (self.nslots - self.history["nslots"]) loss_exemplar = (d_weight * loss_exemplar_distill+ c_weight* loss_exemplar) / (d_weight+c_weight) e_weight = self.exemplar_features.size(0) loss = (nvalid * loss + e_weight * loss_exemplar) / (nvalid + e_weight) if torch.isnan(loss): print(loss, loss_exemplar) return loss else: if return_feature: return scores[:, :nslots], inputs else: return scores[:, :nslots] def score(self, *args, **kwargs): return self.forward(*args, **kwargs) def clone_params(self,): return OrderedDict({k:v.clone().detach() for k,v in self.meta_named_parameters()}) def set_history(self,): self.history = {"params": self.clone_params(), "nslots": self.nslots} def set_exemplar(self, dataloader, q:int=20, params=None, label_sets:Union[List, Set, None]=None, collect_none:bool=False, use_input:bool=False, output_only:bool=False, output:Union[str, None]=None): self.eval() with torch.no_grad(): ifeat = []; ofeat = []; label = [] num_batches = len(dataloader) for batch in tqdm(dataloader, "collecting exemplar", ncols=128): batch = batch.to(self.device) loss = self.forward(batch, params=params) ifeat.append(self.outputs["input_features"]) if use_input: ofeat.append(self.outputs["input_features"]) else: ofeat.append(self.outputs["encoded_features"]) label.append(self.outputs["label"]) ifeat = torch.cat(ifeat, dim=0) ofeat = torch.cat(ofeat, dim=0) label = torch.cat(label, dim=0) nslots = max(self.nslots, torch.max(label).item()+1) exemplar = {} if label_sets is None: if collect_none: label_sets = range(nslots) else: label_sets = range(1, nslots) else: if collect_none: if 0 not in label_sets: label_sets = sorted([0] + list(label_sets)) else: label_sets = sorted(list(label_sets)) else: label_sets = sorted([t for t in label_sets if t != 0]) for i in label_sets: idx = (label == i) if i == 0: # random sample for none type nidx = torch.nonzero(idx, as_tuple=True)[0].tolist() exemplar[i] = numpy.random.choice(nidx, q, replace=False).tolist() continue if torch.any(idx): exemplar[i] = [] nidx = torch.nonzero(idx, as_tuple=True)[0].tolist() mfeat = torch.mean(ofeat[idx], dim=0, keepdims=True) if len(nidx) < q: exemplar[i].extend(nidx * (q // len(nidx)) + nidx[:(q % len(nidx))]) else: for j in range(q): if j == 0: dfeat = torch.sum((ofeat[nidx] - mfeat)**2, dim=1) else: cfeat = ofeat[exemplar[i]].sum(dim=0, keepdims=True) cnum = len(exemplar[i]) dfeat = torch.sum((mfeat * (cnum + 1) - ofeat[nidx] - cfeat)**2, ) tfeat = torch.argmin(dfeat) exemplar[i].append(nidx[tfeat]) nidx.pop(tfeat.item()) exemplar = {i: ifeat[v] for i,v in exemplar.items()} exemplar_features = [] exemplar_labels = [] for label, features in exemplar.items(): exemplar_features.append(features) exemplar_labels.extend([label]*features.size(0)) exemplar_features = torch.cat(exemplar_features, dim=0).cpu() exemplar_labels = torch.LongTensor(exemplar_labels).cpu() if not output_only or output is not None: if output == "train" or output is None: if self.exemplar_features is None: self.exemplar_features = exemplar_features self.exemplar_labels = exemplar_labels else: self.exemplar_features = torch.cat((self.exemplar_features, exemplar_features), dim=0) self.exemplar_labels = torch.cat((self.exemplar_labels, exemplar_labels), dim=0) elif output == "dev": if self.dev_exemplar_features is None: self.dev_exemplar_features = exemplar_features self.dev_exemplar_labels = exemplar_labels else: self.dev_exemplar_features = torch.cat((self.dev_exemplar_features, exemplar_features), dim=0) self.dev_exemplar_labels = torch.cat((self.dev_exemplar_labels, exemplar_labels), dim=0) return {i: v.cpu() for i,v in exemplar.items()} def initialize(self, exemplar, ninstances:Dict[int, int], gamma:float=1.0, tau:float=1.0, alpha:float=0.5, params=None): self.eval() with torch.no_grad(): weight_norm = torch.norm(self.classes.weight[1:self.nslots], dim=1).mean(dim=0) label_inits = [] label_kt = {} for label, feats in exemplar.items(): exemplar_inputs = self.input_map(feats.to(self.device), params=self.get_subdict(params, "input_map")) exemplar_scores = self.classes(exemplar_inputs, params=self.get_subdict(params, "classes")) exemplar_scores = exemplar_scores + self.mask exemplar_scores[:, 0] = 0 exemplar_weights = torch.softmax(exemplar_scores * tau, dim=1) normalized_inputs = exemplar_inputs / torch.norm(exemplar_inputs, dim=1, keepdim=True) * weight_norm proto = (exemplar_weights[:, :1] * normalized_inputs).mean(dim=0) knowledge = torch.matmul(exemplar_weights[:, 1:self.nslots], self.classes.weight[1:self.nslots]).mean(dim=0) gate = alpha * math.exp(- ninstances[label] * gamma) # gate = 1 / (1 + ninstances[label] * gamma) rnd = torch.randn_like(proto) * weight_norm / math.sqrt(self.classes.weight.size(1)) initvec = proto * gate + knowledge * gate + (1 - gate) * rnd label_inits.append((label, initvec.cpu())) label_kt[label] = exemplar_weights.mean(dim=0).cpu() label_inits.sort(key=lambda t:t[0]) inits = [] for i, (label, init) in enumerate(label_inits): assert label == self.nslots + i inits.append(init) inits = torch.stack(inits, dim=0) self.outputs["new2old"] = label_kt return inits.detach() def initialize2(self, exemplar, ninstances:Dict[int, int], gamma:float=1.0, tau:float=1.0, alpha:float=0.5, delta:float=0.5, params=None): self.eval() def top_p(probs, p=0.9): _val, _idx = torch.sort(probs, descending=True, dim=1) top_mask = torch.zeros_like(probs).float() - float("inf") for _type in range(probs.size(0)): accumulated = 0 _n = 0 while accumulated < p or _n <= 1: top_mask[_type, _idx[_type, _n]] = 0 accumulated += _val[_type, _n] _n += 1 return top_mask with torch.no_grad(): weight_norm = torch.norm(self.classes.weight[1:self.nslots], dim=1).mean(dim=0) label_inits = [] label_kt = {} for label, feats in exemplar.items(): exemplar_inputs = self.input_map(feats.to(self.device), params=self.get_subdict(params, "input_map")) exemplar_scores = self.classes(exemplar_inputs, params=self.get_subdict(params, "classes")) exemplar_scores = exemplar_scores + self.mask exemplar_scores[:, 0] = 0 top_mask = top_p(torch.softmax(exemplar_scores, dim=1)) exemplar_scores = exemplar_scores + top_mask exemplar_scores[:, 0] = 0 exemplar_weights = torch.softmax(exemplar_scores * tau, dim=1) normalized_inputs = exemplar_inputs / torch.norm(exemplar_inputs, dim=1, keepdim=True) * weight_norm proto = delta * (exemplar_weights[:, :1] * normalized_inputs).mean(dim=0) kweight = (1 - exemplar_weights[:, :1]) knowledge = torch.matmul((1-delta*exemplar_weights[:, :1]) * (exemplar_weights[:, 1:self.nslots] + 1e-8) / torch.clamp(1 - exemplar_weights[:, :1], 1e-8), self.classes.weight[1:self.nslots]).mean(dim=0) gate = alpha * math.exp(- ninstances[label] * gamma) rnd = torch.randn_like(proto) * weight_norm / math.sqrt(self.classes.weight.size(1)) initvec = proto * gate + knowledge * gate + (1 - gate) * rnd if torch.any(torch.isnan(initvec)): print(proto, knowledge, rnd, gate, exemplar_weights[:, :1], exemplar_scores[-1, :self.nslots]) input() label_inits.append((label, initvec.cpu())) label_kt[label] = exemplar_weights.mean(dim=0).cpu() label_inits.sort(key=lambda t:t[0]) inits = [] for i, (label, init) in enumerate(label_inits): assert label == self.nslots + i inits.append(init) inits = torch.stack(inits, dim=0) self.outputs["new2old"] = label_kt return inits.detach() def set(self, features:torch.tensor, ids:Union[int, torch.Tensor, List, None]=None, max_id:int=-1): with torch.no_grad(): if isinstance(ids, (torch.Tensor, list)): if torch.any(ids > self.nslots): warnings.warn("Setting features to new classes. Using 'extend' or 'append' is preferred for new classes") self.classes.weight[ids] = features elif isinstance(ids, int): self.classes.weight[ids] = features else: if max_id == -1: raise ValueError(f"Need input for either ids or max_id") self.classes.weight[:max_id] = features def append(self, feature): with torch.no_grad(): self.classes.weight[self.nslots] = feature self.nslots += 1 def extend(self, features): with torch.no_grad(): features = features.to(self.device) if len(features.size()) == 1: warnings.warn("Extending 1-dim feature vector. Using 'append' instead is preferred.") self.append(features) else: nclasses = features.size(0) self.classes.weight[self.nslots:self.nslots+nclasses] = features self.nslots += nclasses class BIC(LInEx): def __init__(self,input_dim:int,hidden_dim:int,max_slots:int,init_slots:int,device:Union[torch.device, None]=None, **kwargs)->None: super().__init__(input_dim,hidden_dim,max_slots,init_slots,device,**kwargs) self.correction_weight = nn.Parameter(torch.ones(1, dtype=torch.float, device=self.device, requires_grad=True)) self.correction_bias = nn.Parameter(torch.zeros(1, dtype=torch.float, device=self.device, requires_grad=True)) self.correction_stream = [init_slots] def add_stream(self, num_classes): self.correction_stream.append(self.correction_stream[-1]+num_classes) def forward(self, batch, nslots:int=-1, bias_correction:str="none", exemplar:bool=False, exemplar_distill:bool=False, distill:bool=False, return_loss:bool=True, tau:float=1.0, log_outputs:bool=True, params=None): assert bias_correction in ["none", "last", "current"] if distill: assert bias_correction != "current" if isinstance(batch, (tuple, list)) and len(batch) == 2: features, labels = batch else: features, labels = batch.features, batch.labels inputs = self.input_map(features, params=self.get_subdict(params, "input_map")) scores = self.classes(inputs, params=self.get_subdict(params, "classes")) if nslots == -1: scores += self.mask nslots = self.nslots else: scores += self.idx_mask(max_idx=nslots) scores[:, 0] = 0 if bias_correction == "current": assert len(self.correction_stream) >= 2 scores[:, self.correction_stream[-2]:self.correction_stream[-1]] *= self.correction_weight scores[:, self.correction_stream[-2]:self.correction_stream[-1]] += self.correction_bias if scores.size(0) != labels.size(0): assert scores.size(0) % labels.size(0) == 0 labels = labels.repeat_interleave(scores.size(0) // labels.size(0), dim=0) else: labels = labels if log_outputs: pred = torch.argmax(scores, dim=1) acc = torch.mean((pred == labels).float()) self.outputs["accuracy"] = acc.item() self.outputs["prediction"] = pred.detach().cpu() self.outputs["label"] = labels.detach().cpu() self.outputs["input_features"] = features.detach().cpu() self.outputs["encoded_features"] = inputs.detach().cpu() if return_loss: labels.masked_fill_(labels >= nslots, 0) valid = labels < nslots nvalid = torch.sum(valid.float()) if nvalid == 0: loss = 0 else: loss = self.crit(scores[valid], labels[valid]) if distill and self.history is not None: old_scores = self.forward(batch, nslots=self.history["nslots"], return_loss=False, log_outputs=False, params=self.history["params"]).detach() if bias_correction == "last": old_scores[:, self.correction_stream[-2]:self.correction_stream[-1]] *= self.history['correction_weight'] old_scores[:, self.correction_stream[-2]:self.correction_stream[-1]] += self.history['correction_bias'] new_scores = scores[:, :self.history["nslots"]] loss_distill = - torch.sum(torch.softmax(old_scores*tau, dim=1) * torch.log_softmax(new_scores*tau, dim=1), dim=1).mean() d_weight = self.history["nslots"] c_weight = (self.nslots - self.history["nslots"]) loss = ( d_weight * loss_distill+ c_weight* loss) / (d_weight+c_weight) if exemplar and self.exemplar_features is not None: if self.exemplar_features.size(0) < 128: exemplar_inputs = self.input_map(self.exemplar_features.to(self.device), params=self.get_subdict(params, "input_map")) exemplar_scores = self.classes(exemplar_inputs, params=self.get_subdict(params, "classes")) else: exemplar_scores = [] for _beg in range(0, self.exemplar_features.size(0), 128): _features = self.exemplar_features[_beg:_beg+128, :] _inputs = self.input_map(_features.to(self.device), params=self.get_subdict(params, "input_map")) exemplar_scores.append(self.classes(_inputs, params=self.get_subdict(params, "classes"))) exemplar_scores = torch.cat(exemplar_scores, dim=0) exemplar_scores[:, 0] = 0. loss_exemplar = self.crit(exemplar_scores+self.mask, self.exemplar_labels.to(self.device)) if exemplar_distill: if self.exemplar_features.size(0) < 128: exemplar_old_inputs = self.input_map(self.exemplar_features.to(self.device), params=self.get_subdict(self.history["params"], "input_map")) exemplar_old_scores = self.classes(exemplar_old_inputs, params=self.get_subdict(self.history["params"], "classes")) else: exemplar_old_scores = [] for _beg in range(0, self.exemplar_features.size(0), 128): _features = self.exemplar_features[_beg:_beg+128, :] _inputs = self.input_map(_features.to(self.device), params=self.get_subdict(self.history["params"], "input_map")) exemplar_old_scores.append(self.classes(_inputs, params=self.get_subdict(self.history["params"], "classes"))) exemplar_old_scores = torch.cat(exemplar_old_scores, dim=0) exemplar_old_scores[:, 0] = 0. if bias_correction == "last": exemplar_old_scores[:, self.correction_stream[-2]:self.correction_stream[-1]] *= self.history['correction_weight'] exemplar_old_scores[:, self.correction_stream[-2]:self.correction_stream[-1]] += self.history['correction_bias'] exemplar_old_scores = exemplar_old_scores[:self.history["nslots"]] loss_exemplar_distill = - torch.sum(torch.softmax(exemplar_old_scores[:self.history["nslots"]]*tau, dim=1) * torch.log_softmax(exemplar_scores[:self.history["nslots"]], dim=1), dim=1).mean() d_weight = self.history["nslots"] c_weight = (self.nslots - self.history["nslots"]) loss_exemplar = (d_weight * loss_exemplar_distill+ c_weight* loss_exemplar) / (d_weight+c_weight) e_weight = self.exemplar_features.size(0) loss = (nvalid * loss + e_weight * loss_exemplar) / (nvalid + e_weight) if torch.isnan(loss): print(loss, loss_exemplar) return loss else: return scores[:, :nslots] def forward_correction(self, *args, **kwargs): ''' training: entropy: normal distill: old, last Fold, Fold * correction_weight + correction_bias, ''' if len(args) >= 3: args[2] = "current" else: kwargs["bias_correction"] = "current" return self.forward(*args,**kwargs) def set_history(self): super().set_history() self.history["correction_weight"] = self.correction_weight.item() self.history["correction_bias"] = self.correction_bias.item() def score(self, *args, **kwargs): if len(self.correction_stream) >= 2: return self.forward_correction(*args, **kwargs) else: if len(args) >= 3: args[2] = "none" else: kwargs["bias_correction"] = "none" return self.forward(*args, **kwargs) class ICARL(LInEx): def __init__(self,input_dim:int,hidden_dim:int,max_slots:int,init_slots:int,device:Union[torch.device, None]=None, **kwargs)->None: super().__init__(input_dim,hidden_dim,max_slots,init_slots,device,**kwargs) self.none_feat = None def set_none_feat(self, dataloader, params=None): self.eval() with torch.no_grad(): ifeat = []; ofeat = []; label = [] num_batches = len(dataloader) for batch in tqdm(dataloader, "collecting exemplar"): batch = batch.to(self.device) loss = self.forward(batch, params=params) ifeat.append(self.outputs["input_features"]) ofeat.append(self.outputs["encoded_features"]) label.append(self.outputs["label"]) ifeat = torch.cat(ifeat, dim=0) ofeat = torch.cat(ofeat, dim=0) label = torch.cat(label, dim=0) nslots = max(self.nslots, torch.max(label).item()+1) exemplar = {} idx = (label == 0) self.none_feat = ofeat[idx].mean(dim=0).cpu() return self.none_feat def score(self, batch, exemplar=None, params=None): if exemplar is None: exemplar_labels, exemplar_features = self.exemplar_labels, self.exemplar_features else: exemplar_labels, exemplar_features = exemplar inputs = self.input_map(batch.features, params=self.get_subdict(params, "input_map")) scores = [] scores.append(- torch.sum((inputs - self.none_feat.to(inputs.device).unsqueeze(0))**2, dim=1)) for i in range(1, self.nslots): label_idx = (exemplar_labels == i) label_features = exemplar_features[label_idx] label_inputs = self.input_map(label_features.to(inputs.device), params=self.get_subdict(params, "input_map")).mean(dim=0, keepdim=True) scores.append(- torch.sum((inputs - label_inputs)**2, dim=1)) scores = torch.stack(scores, dim=0).transpose(0, 1) labels = batch.labels if scores.size(0) != labels.size(0): assert scores.size(0) % labels.size(0) == 0 labels = labels.repeat_interleave(scores.size(0) // labels.size(0), dim=0) pred = torch.argmax(scores, dim=1) acc = torch.mean((pred == labels).float()) labels.masked_fill_(labels >= self.nslots, 0) valid = labels < self.nslots nvalid = torch.sum(valid.float()) if nvalid == 0: loss = 0 else: loss = self.crit(scores[valid], labels[valid]) self.outputs["accuracy"] = acc.item() self.outputs["prediction"] = pred.detach().cpu() self.outputs["label"] = labels.detach().cpu() self.outputs["input_features"] = batch.features.detach().cpu() self.outputs["encoded_features"] = inputs.detach().cpu() return loss def test(): # sanity check m = LInEx(nhead=8,nlayers=3,hidden_dim=512,input_dim=2048,max_slots=30,init_slots=9,device=torch.device("cpu")) if __name__ == "__main__": test()
[ "life4pf@163.com" ]
life4pf@163.com
9dce78213c77274a834e67aa526b49d3187883d8
767318c4ddf2713a8a035aa3bf68cd8260409aa0
/travellow/urls.py
a644dc789f9aa6062a14454cf7171e52bf44b7fb
[]
no_license
sag-coder/travelbooking
704573b145ca04587bbaf2415f4bbdb6ad50b26f
dfc482ca01d1be324aba900075b2a64dc2fd1d88
refs/heads/master
2023-06-11T23:22:44.114545
2021-07-10T23:47:37
2021-07-10T23:47:37
384,562,878
0
0
null
null
null
null
UTF-8
Python
false
false
285
py
from django.urls import path from . import views from django.conf import settings from django.conf.urls.static import static urlpatterns=[ path('', views.index, name='index'), ] urlpatterns = urlpatterns + static(settings.MEDIA_URL, document_root=settings.MEDIA_ROOT)
[ "sagar@sagars-MacBook-Pro.local" ]
sagar@sagars-MacBook-Pro.local
0757a5984a57b6a43f6f9c8dbc15fe28c3b58c96
9f77118ac2fdfbdc3174c1fdfaa32bcf2b5ead40
/ALL_FILES/poxresources.py
71b122023b8723299b5cc25a607861109073ad60
[ "Apache-2.0", "MIT" ]
permissive
gilneidp/FinalProject
309979cb65a115a980c58433decc921e295147bf
ec4f35d154bc4383ccde113126e493c1521ad21a
refs/heads/master
2021-01-01T18:42:30.789219
2015-10-30T10:40:47
2015-10-30T10:40:47
41,107,057
0
0
null
null
null
null
UTF-8
Python
false
false
779
py
import os import sys import datetime from datetime import timedelta os.environ.setdefault("DJANGO_SETTINGS_MODULE", "madapp.settings") from django.core.management import execute_from_command_line from django.db.models import Count, Avg from madapp import settings from madapp.mad.models import * import psutil cpu = psutil.cpu_times() memory_used = psutil.virtual_memory() memory_free = (memory_used.free/1024)/1024 #oxstats = UsageTable.objects.get(servername = 'POX') #oxstats.cpu_usage = cpu #oxstats.save() print cpu #print (((memory_used.used/1024)/1024)*100)/(((memory_used.free/1024)/1024) + ((memory_used.used/1024)/1024)) #print (memory_used.used/1024)/1024 #print ((memory_used.free/1024)/1024) + ((memory_used.used/1024)/1024) #print cpu #print memory_used.percent
[ "gilnei@gilnei.com.br" ]
gilnei@gilnei.com.br
ff8cefcde863d2211483e1d36ab5250a2795db09
667e52d9501e04ad9d301c405a1ffc57dabc439e
/checkio/find_friends.py
dcb04ca3265d5c32614ffd7993960e1e3624939c
[]
no_license
merryjane/python
c9ab15f09b83a0f0069b49fe5f2f7ed6e601831c
423aa3cdf226404b2bf9f5958b5d03fe84b93f74
refs/heads/master
2021-01-23T16:37:05.387564
2014-07-25T14:17:13
2014-07-25T14:17:13
null
0
0
null
null
null
null
UTF-8
Python
false
false
1,690
py
#!/usr/bin/env python3 def look_for_neighbors(network,node): neighbors = set() for couples in network: if couples.split('-')[0] == node: neighbors.add(couples.split('-')[1]) elif couples.split('-')[1] == node: neighbors.add(couples.split('-')[0]) return neighbors def check_connection(network, first, second): visited = set() visited.add(first) future = set() future.update(look_for_neighbors(network,first)) for node in future: if node == second: return True else: visited.add(node) future.remove(node) print(future) print(visited) check_connection( ("dr101-mr99", "mr99-out00", "dr101-out00", "scout1-scout2", "scout3-scout1", "scout1-scout4", "scout4-sscout", "sscout-super"), "scout2", "scout3") ''' if __name__ == '__main__': #These "asserts" using only for self-checking and not necessary for auto-testing assert check_connection( ("dr101-mr99", "mr99-out00", "dr101-out00", "scout1-scout2", "scout3-scout1", "scout1-scout4", "scout4-sscout", "sscout-super"), "scout2", "scout3") == True, "Scout Brotherhood" assert check_connection( ("dr101-mr99", "mr99-out00", "dr101-out00", "scout1-scout2", "scout3-scout1", "scout1-scout4", "scout4-sscout", "sscout-super"), "super", "scout2") == True, "Super Scout" assert check_connection( ("dr101-mr99", "mr99-out00", "dr101-out00", "scout1-scout2", "scout3-scout1", "scout1-scout4", "scout4-sscout", "sscout-super"), "dr101", "sscout") == False, "I don't know any scouts." '''
[ "ibiryulin@qsoft.ru" ]
ibiryulin@qsoft.ru
ef2911b4133217bc48dbf92e02a62bd1d9b5d171
e168a16fdd43d3023d16d8a643ccca318a44c327
/evm/logic/call.py
42acedd0f1791f1cebd63438077524bdee541b46
[]
no_license
DavidKnott/py-evm
c589c88af55c121ea375bfdb0a53ecc6a4836119
66c47f58a62e995b5ce89e47007c8b03796c80b9
refs/heads/master
2021-01-01T04:08:39.921768
2017-07-18T13:03:45
2017-07-18T13:03:45
97,128,228
1
0
null
2017-07-13T13:54:57
2017-07-13T13:54:56
null
UTF-8
Python
false
false
7,349
py
from evm import constants from evm.opcode import ( Opcode, ) from evm.utils.address import ( force_bytes_to_address, ) class BaseCall(Opcode): def compute_msg_gas(self, computation, gas, to, value): raise NotImplementedError("Must be implemented by subclasses") def get_call_params(self, computation): raise NotImplementedError("Must be implemented by subclasses") def __call__(self, computation): computation.gas_meter.consume_gas( self.gas_cost, reason=self.mnemonic, ) ( gas, value, to, sender, code_address, memory_input_start_position, memory_input_size, memory_output_start_position, memory_output_size, should_transfer_value, ) = self.get_call_params(computation) computation.extend_memory(memory_input_start_position, memory_input_size) computation.extend_memory(memory_output_start_position, memory_output_size) call_data = computation.memory.read(memory_input_start_position, memory_input_size) # # Message gas allocation and fees # child_msg_gas, child_msg_gas_fee = self.compute_msg_gas(computation, gas, to, value) computation.gas_meter.consume_gas(child_msg_gas_fee, reason=self.mnemonic) # Pre-call checks sender_balance = computation.state_db.get_balance( computation.msg.storage_address, ) insufficient_funds = should_transfer_value and sender_balance < value stack_too_deep = computation.msg.depth + 1 > constants.STACK_DEPTH_LIMIT if insufficient_funds or stack_too_deep: if self.logger: if insufficient_funds: err_message = "Insufficient Funds: have: {0} | need: {1}".format( sender_balance, value, ) elif stack_too_deep: err_message = "Stack Limit Reached" else: raise Exception("Invariant: Unreachable code path") self.logger.debug( "%s failure: %s", self.mnemonic, err_message, ) computation.gas_meter.return_gas(child_msg_gas) computation.stack.push(0) else: if code_address: code = computation.state_db.get_code(code_address) else: code = computation.state_db.get_code(to) child_msg_kwargs = { 'gas': child_msg_gas, 'value': value, 'to': to, 'data': call_data, 'code': code, 'code_address': code_address, 'should_transfer_value': should_transfer_value, } if sender is not None: child_msg_kwargs['sender'] = sender child_msg = computation.prepare_child_message(**child_msg_kwargs) if child_msg.is_create: child_computation = computation.vm.apply_create_message(child_msg) else: child_computation = computation.vm.apply_message(child_msg) computation.children.append(child_computation) if child_computation.error: computation.stack.push(0) else: actual_output_size = min(memory_output_size, len(child_computation.output)) computation.gas_meter.return_gas(child_computation.gas_meter.gas_remaining) computation.memory.write( memory_output_start_position, actual_output_size, child_computation.output[:actual_output_size], ) computation.stack.push(1) class Call(BaseCall): def compute_msg_gas(self, computation, gas, to, value): account_exists = computation.state_db.account_exists(to) transfer_gas_fee = constants.GAS_CALLVALUE if value else 0 create_gas_fee = constants.GAS_NEWACCOUNT if not account_exists else 0 total_fee = gas + transfer_gas_fee + create_gas_fee child_msg_gas = gas + (constants.GAS_CALLSTIPEND if value else 0) return child_msg_gas, total_fee def get_call_params(self, computation): gas = computation.stack.pop(type_hint=constants.UINT256) to = force_bytes_to_address(computation.stack.pop(type_hint=constants.BYTES)) ( value, memory_input_start_position, memory_input_size, memory_output_start_position, memory_output_size, ) = computation.stack.pop(num_items=5, type_hint=constants.UINT256) return ( gas, value, to, None, # sender None, # code_address memory_input_start_position, memory_input_size, memory_output_start_position, memory_output_size, True, # should_transfer_value, ) class CallCode(BaseCall): def compute_msg_gas(self, computation, gas, to, value): transfer_gas_cost = constants.GAS_CALLVALUE if value else 0 total_fee = transfer_gas_cost + gas child_msg_gas = gas + (constants.GAS_CALLSTIPEND if value else 0) return child_msg_gas, total_fee def get_call_params(self, computation): gas = computation.stack.pop(type_hint=constants.UINT256) code_address = force_bytes_to_address(computation.stack.pop(type_hint=constants.BYTES)) ( value, memory_input_start_position, memory_input_size, memory_output_start_position, memory_output_size, ) = computation.stack.pop(num_items=5, type_hint=constants.UINT256) to = computation.msg.storage_address sender = computation.msg.storage_address return ( gas, value, to, sender, code_address, memory_input_start_position, memory_input_size, memory_output_start_position, memory_output_size, True, # should_transfer_value, ) class DelegateCall(CallCode): def compute_msg_gas(self, computation, gas, to, value): return gas, gas def get_call_params(self, computation): gas = computation.stack.pop(type_hint=constants.UINT256) code_address = force_bytes_to_address(computation.stack.pop(type_hint=constants.BYTES)) ( memory_input_start_position, memory_input_size, memory_output_start_position, memory_output_size, ) = computation.stack.pop(num_items=4, type_hint=constants.UINT256) to = computation.msg.storage_address sender = computation.msg.sender value = computation.msg.value return ( gas, value, to, sender, code_address, memory_input_start_position, memory_input_size, memory_output_start_position, memory_output_size, False, # should_transfer_value, )
[ "pipermerriam@gmail.com" ]
pipermerriam@gmail.com
f0a82e1bd0914b8007e6b425025033a3628ec23c
3badd0d2ea861e56748f5c16beba67e1685dc7c3
/functional_tests.py
75114b2d2c52b96f38a980fbc54472aef4cb095d
[]
no_license
chris-seals/tdd-book
a8b72c0ba83e1a27088783225a3fd30fc32bdb28
28d9742e19d510fd07fe21928ede7c316958f4fa
refs/heads/master
2022-11-22T03:34:23.940099
2020-07-19T14:27:00
2020-07-19T14:27:00
258,937,633
0
0
null
null
null
null
UTF-8
Python
false
false
2,246
py
from selenium import webdriver from selenium.webdriver.common.keys import Keys import time import unittest class NewVisitorTest(unittest.TestCase): def setUp(self): self.browser = webdriver.Firefox() def tearDown(self): self.browser.quit() def test_can_start_a_list_and_retrieve_it_later(self): # Edith has heard about a cool new online to-do app. # She goes to check out the homepage. self.browser.get('http://localhost:8000') # She notices the page title and header mention to-do lists self.assertIn('To-Do', self.browser.title) header_text = self.browser.find_element_by_tag_name('h1').text self.assertIn('To-Do', header_text) # She is invited to enter a to-do item straight away inputbox = self.browser.find_element_by_id('id_new_item') self.assertEqual( inputbox.get_attribute('placeholder'), 'Enter a to-do item' ) # She types "Buy peacock feathers" into a text box (Edith's hobby is # tying fly-fishing lures) inputbox.send_keys('Buy peacock feathers') # When she hits enter, the page updates, and now the page lists # "1: Buy peacock feathers" as an item in a to-do lists inputbox.send_keys(Keys.ENTER) time.sleep(1) table = self.browser.find_element_by_id('id_list_table') rows = table.find_elements_by_tag_name('tr') self.assertTrue( any(row.text == '1: Buy peacock feathers' for row in rows), "New to-do item did not appear in table" ) # There is still a text box inviting her to add another item. She enters # "Use peacock feathers to make a fly" (Edith is very methodical) self.fail('Finish the test!') # The page updates again, and now shows both items on her lists # Edith wonders whether the site will remember her list. Then she sees # that the site has generated a unique URL for her -- there is some # explanatory text to that effect. # She visits that URL - her to-do list is still there. # Satisfied, she goes back to sleep if __name__ == '__main__': unittest.main(warnings='ignore')
[ "christopher.seals@gmail.com" ]
christopher.seals@gmail.com
848257d62f49ecdcc747c38384d79aa0afb7700b
8db1ab4f9a2e47f7e8d69a685837d7e747bf9442
/cocos2d-x-tool/py_tool/syncResToProject.py
0773ceebfd0313dd7ab2c0df0f04cec7b688b661
[]
no_license
tanzuoliang/python
051d6e46cebd7fdb74a0173aca0ca7a2b3ef5986
70f782cf3c72d2b7043727910509eb2d2f2fe065
refs/heads/master
2021-10-20T05:36:03.732738
2019-02-26T02:37:18
2019-02-26T02:37:18
111,288,598
0
0
null
null
null
null
UTF-8
Python
false
false
2,595
py
#!/usr/bin/python #encoding=utf-8 from myutil.utils import syncDir,checkSVNStatus import os fromRoot = '/Users/tanzuoliang/art_resource' toRoot = "../res/new_ui" toLanguage = "../res" ll = [ ('天天坦克/UI\ 效果图+输出\ 20170214\ 优化版/00\ icon','天天坦克/UI 效果图+输出 20170214 优化版/00 icon','icon'), ('天天坦克/UI\ 效果图+输出\ 20170214\ 优化版/00\ button','天天坦克/UI 效果图+输出 20170214 优化版/00 button','button'), ('天天坦克/UI\ 效果图+输出\ 20170214\ 优化版/00\ wenzi','天天坦克/UI 效果图+输出 20170214 优化版/00 wenzi','wenzi'), ('天天坦克/UI\ 效果图+输出\ 20170214\ 优化版/00\ 通用','天天坦克/UI 效果图+输出 20170214 优化版/00 通用','common'), ('天天坦克/UI\ 效果图+输出\ 20170214\ 优化版/00\ 字体','天天坦克/UI 效果图+输出 20170214 优化版/00 字体','fnt'), ('天天坦克/UI\ 效果图+输出\ 20170214\ 优化版/00\ BG','天天坦克/UI 效果图+输出 20170214 优化版/00 BG','bg') ] """ 语言分类资源 """ lll = [ ('天天坦克/UI\ 效果图+输出\ 20170214\ 优化版/00\ 英文翻译','天天坦克/UI 效果图+输出 20170214 优化版/00 英文翻译','lang_en'), ('天天坦克/UI\ 效果图+输出\ 20170214\ 优化版/00\ 翻译原版','天天坦克/UI 效果图+输出 20170214 优化版/00 翻译原版','lang_chs') ] from myutil.utils import getDirsize import os if os.path.exists('../res-new') and getDirsize('../res') < getDirsize('../res-new'): print "当前res是压缩后的" else: os.system('svn up %s'%toLanguage) for tu in ll: fromDir = os.path.join(fromRoot, tu[0]) toDir = os.path.join(toRoot, tu[2]) os.system("svn up %s"%fromDir) fromDir = os.path.join(fromRoot, tu[1]) syncDir(fromDir, toDir,False) checkSVNStatus(toRoot,[tu[2]]) for tu in lll: fromDir = os.path.join(fromRoot, tu[0]) toDir = os.path.join(toLanguage, "language_img", tu[2],"res","new_ui") os.system("svn up %s"%fromDir) fromDir = os.path.join(fromRoot, tu[1]) if not os.path.exists(toDir): os.makedir(toDir) syncDir(fromDir, toDir,False) checkSVNStatus(os.path.join(toLanguage, "language_img"),[tu[2]]) """ 英文引导 """ # os.system("cp %s %s"%(os.path.join(fromRoot,"天天坦克/UI\ 效果图+输出\ 20170214\ 优化版/00\ 英文翻译/Novice\ guide/controlexplain.jpg"),os.path.join(toRoot, "bg/lang_en_controlexplain.jpg"))) os.system("rm -rf %s"%(os.path.join(toLanguage,"language_img/lang_en/res/new_ui/Novice\ guide"))) os.system('svn ci %s -m "同步资源"'%toLanguage)
[ "ysjwdaypm@163.com" ]
ysjwdaypm@163.com
3f60cafc9d44646bf3475d2b1f730a3648b8e27b
863c2fcfd5ebed9153c43a298488abeb6e96d627
/time_series_classification.py
ee0ba8dbea36985a0524ff0d3e4a63708bed7170
[]
no_license
liang112233/time-series-classification
6d51dd7e80b044f6fbc7e64c2dd4e4bf6f1ae8f5
d5e847b302d855bb9dc975e888b2c0e50be32f8e
refs/heads/master
2022-11-25T13:18:35.772054
2020-07-30T15:16:58
2020-07-30T15:16:58
280,761,492
0
0
null
null
null
null
UTF-8
Python
false
false
4,092
py
# https://www.analyticsvidhya.com/blog/2019/01/introduction-time-series-classification/ import os import pandas as pd import numpy as np # matplotlib inline import matplotlib.pyplot as plt from os import listdir import tensorflow as tf gpus = tf.config.experimental.list_physical_devices('GPU') for gpu in gpus: tf.config.experimental.set_memory_growth(gpu, True) tf.config.experimental.set_virtual_device_configuration(gpus[0], [tf.config.experimental.VirtualDeviceConfiguration(memory_limit=4096)]) from keras.preprocessing import sequence import tensorflow as tf from keras.models import Sequential from keras.layers import Dense from keras.layers import LSTM from keras.optimizers import Adam from keras.models import load_model from tensorflow.python.keras.callbacks import ModelCheckpoint from keras.callbacks import ModelCheckpoint df1 = pd.read_csv('/home/liang/PycharmProjects/time-series-classification/MovementAAL/dataset/MovementAAL_RSS_1.csv') df2 = pd.read_csv('/home/liang/PycharmProjects/time-series-classification/MovementAAL/dataset/MovementAAL_RSS_2.csv') path = '/home/liang/PycharmProjects/time-series-classification/MovementAAL/dataset/MovementAAL_RSS_' sequences = list() for i in range(1,315): file_path = path + str(i) + '.csv' print(file_path) df = pd.read_csv(file_path, header=0) values = df.values sequences.append(values) targets = pd.read_csv('/home/liang/PycharmProjects/time-series-classification/MovementAAL/dataset/MovementAAL_target.csv') targets = targets.values[:,1] groups = pd.read_csv('MovementAAL/groups/MovementAAL_DatasetGroup.csv', header=0) groups = groups.values[:,1] len_sequences = [] for one_seq in sequences: len_sequences.append(len(one_seq)) pd.Series(len_sequences).describe() # Padding the sequence with the values in last row to max length to_pad = 129 new_seq = [] for one_seq in sequences: len_one_seq = len(one_seq) last_val = one_seq[-1] n = to_pad - len_one_seq to_concat = np.repeat(one_seq[-1], n).reshape(4, n).transpose() new_one_seq = np.concatenate([one_seq, to_concat]) new_seq.append(new_one_seq) final_seq = np.stack(new_seq) # truncate the sequence to length 60 seq_len = 60 final_seq = sequence.pad_sequences(final_seq, maxlen=seq_len, padding='post', dtype='float', truncating='post') train = [final_seq[i] for i in range(len(groups)) if (groups[i]==2)] validation = [final_seq[i] for i in range(len(groups)) if groups[i]==1] test = [final_seq[i] for i in range(len(groups)) if groups[i]==3] train_target = [targets[i] for i in range(len(groups)) if (groups[i]==2)] validation_target = [targets[i] for i in range(len(groups)) if groups[i]==1] test_target = [targets[i] for i in range(len(groups)) if groups[i]==3] train = np.array(train) validation = np.array(validation) test = np.array(test) train_target = np.array(train_target) train_target = (train_target+1)/2 validation_target = np.array(validation_target) validation_target = (validation_target+1)/2 test_target = np.array(test_target) test_target = (test_target+1)/2 model = Sequential() model.add(LSTM(256, input_shape=(seq_len, 4))) model.add(Dense(1, activation='sigmoid')) model.summary() adam = Adam(lr=0.001) # model_filename = "test-Epoch-{epoch:02d}" # checkpoint_path = os.path.join('models/', model_filename) chk = ModelCheckpoint( 'best_model.pkl', monitor='val_accuracy', verbose=1, save_best_only=False, save_weights_only=False, mode='max') # chk = ModelCheckpoint('best_model.pkl', monitor='val_acc', save_best_only=True, mode='max', verbose=1) model.compile(loss='binary_crossentropy', optimizer=adam, metrics=['accuracy']) model.fit(train, train_target, epochs=5, batch_size=128, callbacks=[chk], validation_data=(validation,validation_target)) # # # #loading the model and checking accuracy on the test data model = load_model('best_model.pkl') # from sklearn.metrics import accuracy_score test_preds = model.predict_classes(test) accuracy_score(test_target, test_preds) print("score",accuracy_score(test_target, test_preds))
[ "sisheng.liang@ttu.edu" ]
sisheng.liang@ttu.edu
87a0d04e73c54c1e0daef6dcf0e338c6af43be21
ef187d259d33e97c7b9ed07dfbf065cec3e41f59
/work/atcoder/abc/abc024/B/answers/111654_Gale.py
b31d17de7f8e5d4c0d019d4cbf95c0c6f7e11513
[]
no_license
kjnh10/pcw
847f7295ea3174490485ffe14ce4cdea0931c032
8f677701bce15517fb9362cc5b596644da62dca8
refs/heads/master
2020-03-18T09:54:23.442772
2018-07-19T00:26:09
2018-07-19T00:26:09
134,586,379
0
0
null
null
null
null
UTF-8
Python
false
false
197
py
n, t = map(int, input().split()) a = [int(input()) for i in range(n)] ans = t for i in range(1, n): ans += t if a[i] <= a[i - 1] + t: ans = ans - (a[i - 1] + t - a[i]) print(ans)
[ "kojinho10@gmail.com" ]
kojinho10@gmail.com
0b583e86f97c1a537be2b27d6980f3a3dd93df1a
528c811306faa4a34bf51fca7955b7a24ac2e30c
/Python/Valid Anagram.py
263508830b33b30fd769bcad02fa5dbf91901f61
[]
no_license
ganjingcatherine/LeetCode-1
1addbd7e4d9254a146601f9d5e28b8becb8235a6
488782d3f1e759da2d32b4e82dbf55b96c431244
refs/heads/master
2021-05-11T03:15:16.810035
2016-02-06T06:19:18
2016-02-06T06:19:18
null
0
0
null
null
null
null
UTF-8
Python
false
false
806
py
""" Given two strings s and t, write a function to determine if t is an anagram of s. For example, s = "anagram", t = "nagaram", return true. s = "rat", t = "car", return false. Note: You may assume the string contains only lowercase alphabets. """ class Solution(object): def isAnagram(self, s, t): """ :type s: str :type t: str :rtype: bool """ if len(s) != len(t): return False table = {} for i in xrange(len(s)): if s[i] not in table: table[s[i]] = 1 else: table[s[i]] += 1 for i in xrange(len(t)): if t[i] in table and table[t[i]] > 0: table[t[i]] -= 1 else: return False return True
[ "anthonyjin0619@gmail.com" ]
anthonyjin0619@gmail.com
626be54fe2c402a3a685abc6d8479c10ea8a75aa
a5a99f646e371b45974a6fb6ccc06b0a674818f2
/CalibMuon/RPCCalibration/python/l1MuonOutputModule_cfi.py
6dbdc357f06e53ed7641a5fc49576123b5f1a25e
[ "Apache-2.0" ]
permissive
cms-sw/cmssw
4ecd2c1105d59c66d385551230542c6615b9ab58
19c178740257eb48367778593da55dcad08b7a4f
refs/heads/master
2023-08-23T21:57:42.491143
2023-08-22T20:22:40
2023-08-22T20:22:40
10,969,551
1,006
3,696
Apache-2.0
2023-09-14T19:14:28
2013-06-26T14:09:07
C++
UTF-8
Python
false
false
419
py
import FWCore.ParameterSet.Config as cms from CalibMuon.RPCCalibration.l1Muon_EventContent_cff import * L1MuonEventContent = cms.OutputModule("PoolOutputModule", l1Muon_EventContent, l1MuonEventSelection, datasets = cms.untracked.PSet( filterName = cms.untracked.string('l1Muon_Filter'), dataTier = cms.untracked.string('USER') ), fileName = cms.untracked.string('l1Muon.root') )
[ "giulio.eulisse@gmail.com" ]
giulio.eulisse@gmail.com
73fb980e24519d85b6a114bf1263927c71ea7335
88deebcc2251f406c0c34325025cd0cbccb142e7
/Drosophila/ABC/melano_simulans_simulate_ms_ABC.py
0c443a1bf2620a4e48345ce53a52e4623d284f5e
[]
no_license
cllong112s/CNN_spDelimitation_Piloso
27dab9269253c4ca360e6f8c4f1d70560bf15d84
5a5cae2498e89291357733f8614c5399558be7c0
refs/heads/master
2023-08-15T20:15:52.292721
2021-09-16T15:08:42
2021-09-16T15:08:42
null
0
0
null
null
null
null
UTF-8
Python
false
false
2,287
py
#!/usr/bin/python ## in order to use this code you have to have ms installed on your computer ## ms can be freely downloaded from: ## http://home.uchicago.edu/rhudson1/source/mksamples.html import random import os import math ### variable declarations #define the number of simulations Priorsize = 10000 ## nDNA sample size of Dmelanogaster. nDNADme = 3 ## nDNA sample size of Dsimulans. nDNADsi = 3 ## nDNA sample sizes (number of alleles). nDNANsam = nDNADme + nDNADsi #number of segregating sites for each marker segsites = [36,36] ## create a file to store parameters and one to store the models parameters = file("parameters.txt","w") models = file("models.txt","w") ### One Species Model for i in range(Priorsize): ### Define parameters ## Theta values from 1 to 15 Theta = random.uniform(5,300) ## divergence time prior set to 0. coalRootDivTime = 0 ## ms commands for s in range(len(segsites)): ## nDNA markers com=os.system("./ms %d 1 -s %d -t %f -I 2 %d %d -ej %f 1 2 | sed '/prob/d' | perl msSS.pl >> simModel1.txt" % (nDNANsam, segsites[s],Theta, nDNADme, nDNADsi, coalRootDivTime)) ## mtDNA marker com=os.system("./ms %d 1 -s 36 -t %f -I 2 %d %d -ej %f 1 2 | sed '/prob/d' | perl msSS.pl >> simModel1.txt" % (nDNANsam, Theta/4, nDNADme, nDNADsi, coalRootDivTime/2)) ## save parameter values and models parameters.write("%f\t%f\n" % (Theta, coalRootDivTime)) models.write("1\n") ### Two Species Model for i in range(Priorsize): ### Define parameters ## Theta values from 5 to 300 Theta = random.uniform(5,300) ## divergence time prior following an uniform distribution from 0.01 to 0.5. coalRootDivTime = random.uniform(0.01,0.5) ## ms commands for s in range(len(segsites)): ## nDNA markers com=os.system("./ms %d 1 -s %d -t %f -I 2 %d %d -ej %f 1 2 | sed '/prob/d' | perl msSS.pl >> simModel2.txt" % (nDNANsam, segsites[s],Theta, nDNADme, nDNADsi, coalRootDivTime)) ## mtDNA marker com=os.system("./ms %d 1 -s 36 -t %f -I 2 %d %d -ej %f 1 2 | sed '/prob/d' | perl msSS.pl >> simModel2.txt" % (nDNANsam, Theta/4, nDNADme, nDNADsi, coalRootDivTime/2)) ## save parameter values and models parameters.write("%f\t%f\n" % (Theta, coalRootDivTime)) models.write("2\n") com=os.system("cat simModel* > SuSt_melano_simulans.txt")
[ "manolo@macbook-pro-de-manolo.home" ]
manolo@macbook-pro-de-manolo.home
fb82d01ee42c4f74d1b66246033ac584b40173c8
9bb28ccac08cbb00b3b973ed1283a4e63db045de
/venv/Scripts/pip3.8-script.py
6f8e9a2a91deacf7a08089af0bac8dcaf76cb0fa
[]
no_license
sanghee5/polls_example2
762202280519efdddff7015cb4399cf80fac94d4
6f0511a6bf0994f46168902c54266ef8f9107519
refs/heads/master
2022-08-21T03:59:38.140488
2020-05-26T11:47:10
2020-05-26T11:47:10
267,026,974
0
0
null
null
null
null
UTF-8
Python
false
false
422
py
#!C:\Users\woo\PycharmProjects\polls_example2\venv\Scripts\python.exe # EASY-INSTALL-ENTRY-SCRIPT: 'pip==19.0.3','console_scripts','pip3.8' __requires__ = 'pip==19.0.3' import re import sys from pkg_resources import load_entry_point if __name__ == '__main__': sys.argv[0] = re.sub(r'(-script\.pyw?|\.exe)?$', '', sys.argv[0]) sys.exit( load_entry_point('pip==19.0.3', 'console_scripts', 'pip3.8')() )
[ "dnstks12345@naver.com" ]
dnstks12345@naver.com
2bff7ce472c638cc2952ee313e844673778ab37c
5faecec9b20d262150e48ac9f31c396f840b1f2f
/migrations/0010_auto_20200804_0913.py
f175b678b5857527caa863cd6db136e7bc3d803b
[]
no_license
binkesi/blogsgn
fb767b0d22e3eb1c32ea7ee8fd0796766e3a8600
579b374f802a5651d20c3b3f85d8ff6a22476bdd
refs/heads/master
2022-11-27T23:24:45.574601
2020-08-04T10:06:28
2020-08-04T10:06:28
283,161,699
0
0
null
null
null
null
UTF-8
Python
false
false
515
py
# Generated by Django 3.0.6 on 2020-08-04 01:13 from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('blogsgn', '0009_auto_20200804_0653'), ] operations = [ migrations.AlterField( model_name='author', name='nation', field=models.CharField(choices=[('CH', 'China'), ('US', 'America'), ('UK', 'England'), ('GE', 'German'), ('CA', 'Canada')], max_length=80, verbose_name='Nationality'), ), ]
[ "sjtusgn@163.com" ]
sjtusgn@163.com
7713fd10c64850e9770370122883e5b6ea01086f
e2ae96b74289a04a2386294bf51bacad92e2a830
/city_scrapers_core/spiders/legistar.py
29c3176db02b4b0851fd939f9f79845a629163c5
[ "MIT" ]
permissive
will-snavely/city-scrapers-core
6afa9d78fb1c325420baaae030633b01111f11bb
cb865069e49d09ab251b7f99247df5e13c5d0241
refs/heads/main
2022-12-11T21:39:03.307347
2020-09-09T13:29:53
2020-09-09T13:29:53
null
0
0
null
null
null
null
UTF-8
Python
false
false
4,265
py
from datetime import datetime from typing import Iterable, List, Mapping, Optional, Tuple from urllib.parse import urlparse import scrapy from legistar.events import LegistarEventsScraper from ..items import Meeting from .spider import CityScrapersSpider LINK_TYPES = ["Agenda", "Minutes", "Video", "Summary", "Captions"] class LegistarSpider(CityScrapersSpider): """Subclass of :class:`CityScrapersSpider` that handles processing Legistar sites, which almost always share the same components and general structure. Uses the `Legistar events scraper <https://github.com/opencivicdata/python-legistar-scraper/blob/master/legistar/events.py>`_ from the ```python-legistar-scraper`` library <https://github.com/opencivicdata/python-legistar-scraper>`. Any methods that don't pull the correct values can be replaced. """ # noqa link_types = [] def parse(self, response: scrapy.http.Response) -> Iterable[Meeting]: """Parse response from the :class:`LegistarEventsScraper`. Ignores the ``scrapy`` :class:`Response` which is still requested to be able to hook into ``scrapy`` broadly. :param response: Scrapy response to be ignored :return: Iterable of processed meetings """ events = self._call_legistar() return self.parse_legistar(events) def parse_legistar( self, events: Iterable[Tuple[Mapping, Optional[str]]] ) -> Iterable[Meeting]: """Method to be implemented by Spider classes that will handle the response from Legistar. Functions similar to ``parse`` for other Spider classes. :param events: Iterable consisting of tuples of a dict-like object of scraped results from legistar and an agenda URL (if available) :raises NotImplementedError: Must be implemented in subclasses :return: [description] """ raise NotImplementedError("Must implement parse_legistar") def _call_legistar( self, since: int = None ) -> Iterable[Tuple[Mapping, Optional[str]]]: les = LegistarEventsScraper() les.BASE_URL = self.base_url les.EVENTSPAGE = f"{self.base_url}/Calendar.aspx" if not since: since = datetime.today().year return les.events(since=since) def legistar_start(self, item: Mapping) -> datetime: """Pulls the start time from a Legistar item :param item: Scraped item from Legistar :return: Meeting start datetime """ start_date = item.get("Meeting Date") start_time = item.get("Meeting Time") if start_date and start_time: try: return datetime.strptime( f"{start_date} {start_time}", "%m/%d/%Y %I:%M %p" ) except ValueError: return datetime.strptime(start_date, "%m/%d/%Y") def legistar_links(self, item: Mapping) -> List[Mapping[str, str]]: """Pulls relevant links from a Legistar item :param item: Scraped item from Legistar :return: List of meeting links """ links = [] for link_type in LINK_TYPES + self.link_types: if isinstance(item.get(link_type), dict) and item[link_type].get("url"): links.append({"href": item[link_type]["url"], "title": link_type}) return links def legistar_source(self, item: Mapping) -> str: """Pulls the source URL from a Legistar item. Pulls a specific meeting URL if available, otherwise defaults to the general Legistar calendar page. :param item: Scraped item from Legistar :return: Source URL """ default_url = f"{self.base_url}/Calendar.aspx" if isinstance(item.get("Name"), dict): return item["Name"].get("url", default_url) if isinstance(item.get("Meeting Details"), dict): return item["Meeting Details"].get("url", default_url) return default_url @property def base_url(self) -> str: """Property with the Legistar site's base URL :return: Legistar base URL """ parsed_url = urlparse(self.start_urls[0]) return f"{parsed_url.scheme}://{parsed_url.netloc}"
[ "pjsier@gmail.com" ]
pjsier@gmail.com
9d101de5dc8616d67f19ec37db6ac2e7ed86d8b1
fde2a3a4858b37cafcd02cf917d3bd69680084b3
/Spacegame/Spacegame/scoreboard.py
7cb820bd2744867f4d0c0ffc58127b56a0c6463a
[]
no_license
CaptainBlowFish/AlienInvasion
34b1e013e97c8f2a69aa82e9e51786ffc5e2f78f
6ec7f399bbaa92901ad8b015047bc6ebe6dc708e
refs/heads/main
2023-08-29T20:04:38.130623
2021-11-12T13:45:19
2021-11-12T13:45:19
422,559,716
0
0
null
null
null
null
UTF-8
Python
false
false
3,082
py
from pygame import font from pygame.sprite import Group from ship import Ship class Scoreboard: """A class to report scoreings information""" def __init__(self, ai_game): """Initialize scorekeeping attributes""" self.ai_game = ai_game self.screen = ai_game.screen self.screen_rect = self.screen.get_rect() self.settings = ai_game.settings self.stats = ai_game.stats # Font settings for scoring information self.text_color = (30,30,30) self.font = font.SysFont(None, 48) # Prepare the initial score image self.prep_score() self.prep_high_score() self.prep_level() self.ships = Group() def prep_score(self): """Turn the score into a rendered image""" rounded_score = round(self.stats.score, -1) score_str = "{:,}".format(rounded_score) self.score_image = self.font.render(score_str, True, self.text_color,None) # Display the se at the top right of the screen self.score_rect = self.score_image.get_rect() self.score_rect.right = self.screen_rect.right - 20 self.score_rect.top = 20 def prep_level(self): """Turn the level into a rendered image.""" level_str = str(self.stats.level) self.level_image = self.font.render(level_str, True, self.text_color, None) # Position the level below the score self.level_rect = self.level_image.get_rect() self.level_rect.right = self.score_rect.right self.level_rect.top = self.score_rect.bottom + 10 def prep_high_score(self): """Turn the high score into a rendered image""" rounded_high_score = round(self.stats.high_score, -1) high_score_str = "{:,}".format(rounded_high_score) self.high_score_image = self.font.render(high_score_str, True, self.text_color,None) # Display the score at the top of the screen self.high_score_rect = self.high_score_image.get_rect() self.high_score_rect.right = self.screen_rect.centerx self.high_score_rect.top = self.score_rect.top def prep_ships(self): """Show howmany ships are left""" self.ships.empty() for ship_number in range(self.stats.ships_left): ship = Ship(self.ai_game) ship.rect.x = 10 + ship_number * ship.rect.width ship.rect.y = 10 self.ships.add(ship) def show_score(self): """Draw the score to the screen.""" self.screen.blit(self.score_image, self.score_rect) self.screen.blit(self.high_score_image, self.high_score_rect) self.screen.blit(self.level_image, self.level_rect) self.ships.draw(self.screen) def check_high_score(self): """Check to see if there's a new high score.""" if self.stats.score > self.stats.high_score: self.stats.high_score = self.stats.score self.prep_high_score()
[ "noreply@github.com" ]
CaptainBlowFish.noreply@github.com
9f0d1351e1fd94cb4f0674d16df5403e9b562544
c9ddbdb5678ba6e1c5c7e64adf2802ca16df778c
/cases/synthetic/sieve-big-7279.py
7b346f67e4f794d398658cee0c9138683fa18c0b
[]
no_license
Virtlink/ccbench-chocopy
c3f7f6af6349aff6503196f727ef89f210a1eac8
c7efae43bf32696ee2b2ee781bdfe4f7730dec3f
refs/heads/main
2023-04-07T15:07:12.464038
2022-02-03T15:42:39
2022-02-03T15:42:39
451,969,776
0
0
null
null
null
null
UTF-8
Python
false
false
31,612
py
# A resizable list of integers class Vector(object): items: [int] = None size: int = 0 def __init__(self:"Vector"): self.items = [0] # Returns current capacity def capacity(self:"Vector") -> int: return len(self.items) # Increases capacity of vector by one element def increase_capacity(self:"Vector") -> int: self.items = self.items + [0] return self.capacity() # Appends one item to end of vector def append(self:"Vector", item: int) -> object: if self.size == self.capacity(): self.increase_capacity() self.items[self.size] = item self.size = self.size + 1 # Appends many items to end of vector def append_all(self:"Vector", new_items: [int]) -> object: item:int = 0 for item in new_items: self.append(item) # Removes an item from the middle of vector def remove_at(self:"Vector", idx: int) -> object: if idx < 0: return while idx < self.size - 1: self.items[idx] = self.items[idx + 1] idx = idx + 1 self.size = self.size - 1 # Retrieves an item at a given index def get(self:"Vector", idx: int) -> int: return self.items[idx] # Retrieves the current size of the vector def length(self:"Vector") -> int: return self.size # A resizable list of integers class Vector2(object): items: [int] = None items2: [int] = None size: int = 0 size2: int = 0 def __init__(self:"Vector2"): self.items = [0] # Returns current capacity def capacity(self:"Vector2") -> int: return len(self.items) # Returns current capacity def capacity2(self:"Vector2") -> int: return len(self.items) # Increases capacity of vector by one element def increase_capacity(self:"Vector2") -> int: self.items = self.items + [0] return self.capacity() # Increases capacity of vector by one element def increase_capacity2(self:"Vector2") -> int: self.items = self.items + [0] return self.capacity() # Appends one item to end of vector def append(self:"Vector2", item: int) -> object: if self.size == self.capacity(): self.increase_capacity() self.items[self.size] = item self.size = self.size + 1 # Appends one item to end of vector def append2(self:"Vector2", item: int, item2: int) -> object: if self.size == self.capacity(): self.increase_capacity() self.items[self.size] = item self.size = self.size + 1 # Appends many items to end of vector def append_all(self:"Vector2", new_items: [int]) -> object: item:int = 0 for item in new_items: self.append(item) # Appends many items to end of vector def append_all2(self:"Vector2", new_items: [int], new_items2: [int]) -> object: item:int = 0 item2:int = 0 for item in new_items: self.append(item) # Removes an item from the middle of vector def remove_at(self:"Vector2", idx: int) -> object: if idx < 0: return while idx < self.size - 1: self.items[idx] = self.items[idx + 1] idx = idx + 1 self.size = self.size - 1 # Removes an item from the middle of vector def remove_at2(self:"Vector2", idx: int, idx2: int) -> object: if idx < 0: return while idx < self.size - 1: self.items[idx] = self.items[idx + 1] idx = idx + 1 self.size = self.size - 1 # Retrieves an item at a given index def get(self:"Vector2", idx: int) -> int: return self.items[idx] # Retrieves an item at a given index def get2(self:"Vector2", idx: int, idx2: int) -> int: return self.items[idx] # Retrieves the current size of the vector def length(self:"Vector2") -> int: return self.size # Retrieves the current size of the vector def length2(self:"Vector2") -> int: return self.size # A resizable list of integers class Vector3(object): items: [int] = None items2: [int] = None items3: [int] = None size: int = 0 size2: int = 0 size3: int = 0 def __init__(self:"Vector3"): self.items = [0] # Returns current capacity def capacity(self:"Vector3") -> int: return len(self.items) # Returns current capacity def capacity2(self:"Vector3") -> int: return len(self.items) # Returns current capacity def capacity3(self:"Vector3") -> int: return len(self.items) # Increases capacity of vector by one element def increase_capacity(self:"Vector3") -> int: self.items = self.items + [0] return self.capacity() # Increases capacity of vector by one element def increase_capacity2(self:"Vector3") -> int: self.items = self.items + [0] return self.capacity() # Increases capacity of vector by one element def increase_capacity3(self:"Vector3") -> int: self.items = self.items + [0] return self.capacity() # Appends one item to end of vector def append(self:"Vector3", item: int) -> object: if self.size == self.capacity(): self.increase_capacity() self.items[self.size] = item self.size = self.size + 1 # Appends one item to end of vector def append2(self:"Vector3", item: int, item2: int) -> object: if self.size == self.capacity(): self.increase_capacity() self.items[self.size] = item self.size = self.size + 1 # Appends one item to end of vector def append3(self:"Vector3", item: int, item2: int, item3: int) -> object: if self.size == self.capacity(): self.increase_capacity() self.items[self.size] = item self.size = self.size + 1 # Appends many items to end of vector def append_all(self:"Vector3", new_items: [int]) -> object: item:int = 0 for item in new_items: self.append(item) # Appends many items to end of vector def append_all2(self:"Vector3", new_items: [int], new_items2: [int]) -> object: item:int = 0 item2:int = 0 for item in new_items: self.append(item) # Appends many items to end of vector def append_all3(self:"Vector3", new_items: [int], new_items2: [int], new_items3: [int]) -> object: item:int = 0 item2:int = 0 item3:int = 0 for item in new_items: self.append(item) # Removes an item from the middle of vector def remove_at(self:"Vector3", idx: int) -> object: if idx < 0: return while idx < self.size - 1: self.items[idx] = self.items[idx + 1] idx = idx + 1 self.size = self.size - 1 # Removes an item from the middle of vector def remove_at2(self:"Vector3", idx: int, idx2: int) -> object: if idx < 0: return while idx < self.size - 1: self.items[idx] = self.items[idx + 1] idx = idx + 1 self.size = self.size - 1 # Removes an item from the middle of vector def remove_at3(self:"Vector3", idx: int, idx2: int, idx3: int) -> object: if idx < 0: return while idx < self.size - 1: self.items[idx] = self.items[idx + 1] idx = idx + 1 self.size = self.size - 1 # Retrieves an item at a given index def get(self:"Vector3", idx: int) -> int: return self.items[idx] # Retrieves an item at a given index def get2(self:"Vector3", idx: int, idx2: int) -> int: return self.items[idx] # Retrieves an item at a given index def get3(self:"Vector3", idx: int, idx2: int, idx3: int) -> int: return self.items[idx] # Retrieves the current size of the vector def length(self:"Vector3") -> int: return self.size # Retrieves the current size of the vector def length2(self:"Vector3") -> int: return self.size # Retrieves the current size of the vector def length3(self:"Vector3") -> int: return self.size # A resizable list of integers class Vector4(object): items: [int] = None items2: [int] = None items3: [int] = None items4: [int] = None size: int = 0 size2: int = 0 size3: int = 0 size4: int = 0 def __init__(self:"Vector4"): self.items = [0] # Returns current capacity def capacity(self:"Vector4") -> int: return len(self.items) # Returns current capacity def capacity2(self:"Vector4") -> int: return len(self.items) # Returns current capacity def capacity3(self:"Vector4") -> int: return len(self.items) # Returns current capacity def capacity4(self:"Vector4") -> int: return len(self.items) # Increases capacity of vector by one element def increase_capacity(self:"Vector4") -> int: self.items = self.items + [0] return self.capacity() # Increases capacity of vector by one element def increase_capacity2(self:"Vector4") -> int: self.items = self.items + [0] return self.capacity() # Increases capacity of vector by one element def increase_capacity3(self:"Vector4") -> int: self.items = self.items + [0] return self.capacity() # Increases capacity of vector by one element def increase_capacity4(self:"Vector4") -> int: self.items = self.items + [0] return self.capacity() # Appends one item to end of vector def append(self:"Vector4", item: int) -> object: if self.size == self.capacity(): self.increase_capacity() self.items[self.size] = item self.size = self.size + 1 # Appends one item to end of vector def append2(self:"Vector4", item: int, item2: int) -> object: if self.size == self.capacity(): self.increase_capacity() self.items[self.size] = item self.size = self.size + 1 # Appends one item to end of vector def append3(self:"Vector4", item: int, item2: int, item3: int) -> object: if self.size == self.capacity(): self.increase_capacity() self.items[self.size] = item self.size = self.size + 1 # Appends one item to end of vector def append4(self:"Vector4", item: int, item2: int, item3: int, item4: int) -> object: if self.size == self.capacity(): self.increase_capacity() self.items[self.size] = item self.size = self.size + 1 # Appends many items to end of vector def append_all(self:"Vector4", new_items: [int]) -> object: item:int = 0 for item in new_items: self.append(item) # Appends many items to end of vector def append_all2(self:"Vector4", new_items: [int], new_items2: [int]) -> object: item:int = 0 item2:int = 0 for item in new_items: self.append(item) # Appends many items to end of vector def append_all3(self:"Vector4", new_items: [int], new_items2: [int], new_items3: [int]) -> object: item:int = 0 item2:int = 0 item3:int = 0 for item in new_items: self.append(item) # Appends many items to end of vector def append_all4(self:"Vector4", new_items: [int], new_items2: [int], new_items3: [int], new_items4: [int]) -> object: item:int = 0 item2:int = 0 item3:int = 0 item4:int = 0 for item in new_items: self.append(item) # Removes an item from the middle of vector def remove_at(self:"Vector4", idx: int) -> object: if idx < 0: return while idx < self.size - 1: self.items[idx] = self.items[idx + 1] idx = idx + 1 self.size = self.size - 1 # Removes an item from the middle of vector def remove_at2(self:"Vector4", idx: int, idx2: int) -> object: if idx < 0: return while idx < self.size - 1: self.items[idx] = self.items[idx + 1] idx = idx + 1 self.size = self.size - 1 # Removes an item from the middle of vector def remove_at3(self:"Vector4", idx: int, idx2: int, idx3: int) -> object: if idx < 0: return while idx < self.size - 1: self.items[idx] = self.items[idx + 1] idx = idx + 1 self.size = self.size - 1 # Removes an item from the middle of vector def remove_at4(self:"Vector4", idx: int, idx2: int, idx3: int, idx4: int) -> object: if idx < 0: return while idx < self.size - 1: self.items[idx] = self.items[idx + 1] idx = idx + 1 self.size = self.size - 1 # Retrieves an item at a given index def get(self:"Vector4", idx: int) -> int: return self.items[idx] # Retrieves an item at a given index def get2(self:"Vector4", idx: int, idx2: int) -> int: return self.items[idx] # Retrieves an item at a given index def get3(self:"Vector4", idx: int, idx2: int, idx3: int) -> int: return self.items[idx] # Retrieves an item at a given index def get4(self:"Vector4", idx: int, idx2: int, idx3: int, idx4: int) -> int: return self.items[idx] # Retrieves the current size of the vector def length(self:"Vector4") -> int: return self.size # Retrieves the current size of the vector def length2(self:"Vector4") -> int: return self.size # Retrieves the current size of the vector def length3(self:"Vector4") -> int: return self.size # Retrieves the current size of the vector def length4(self:"Vector4") -> int: return self.size # A resizable list of integers class Vector5(object): items: [int] = None items2: [int] = None items3: [int] = None items4: [int] = None items5: [int] = None size: int = 0 size2: int = 0 size3: int = 0 size4: int = 0 size5: int = 0 def __init__(self:"Vector5"): self.items = [0] # Returns current capacity def capacity(self:"Vector5") -> int: return len(self.items) # Returns current capacity def capacity2(self:"Vector5") -> int: return len(self.items) # Returns current capacity def capacity3(self:"Vector5") -> int: return len(self.items) # Returns current capacity def capacity4(self:"Vector5") -> int: return len(self.items) # Returns current capacity def capacity5(self:"Vector5") -> int: return len(self.items) # Increases capacity of vector by one element def increase_capacity(self:"Vector5") -> int: self.items = self.items + [0] return self.capacity() # Increases capacity of vector by one element def increase_capacity2(self:"Vector5") -> int: self.items = self.items + [0] return self.capacity() # Increases capacity of vector by one element def increase_capacity3(self:"Vector5") -> int: self.items = self.items + [0] return self.capacity() # Increases capacity of vector by one element def increase_capacity4(self:"Vector5") -> int: self.items = self.items + [0] return self.capacity() # Increases capacity of vector by one element def increase_capacity5(self:"Vector5") -> int: self.items = self.items + [0] return self.capacity() # Appends one item to end of vector def append(self:"Vector5", item: int) -> object: if self.size == self.capacity(): self.increase_capacity() self.items[self.size] = item self.size = self.size + 1 # Appends one item to end of vector def append2(self:"Vector5", item: int, item2: int) -> object: if self.size == self.capacity(): self.increase_capacity() self.items[self.size] = item self.size = self.size + 1 # Appends one item to end of vector def append3(self:"Vector5", item: int, item2: int, item3: int) -> object: if self.size == self.capacity(): self.increase_capacity() self.items[self.size] = item self.size = self.size + 1 # Appends one item to end of vector def append4(self:"Vector5", item: int, item2: int, item3: int, item4: int) -> object: if self.size == self.capacity(): self.increase_capacity() self.items[self.size] = item self.size = self.size + 1 # Appends one item to end of vector def append5(self:"Vector5", item: int, item2: int, item3: int, item4: int, item5: int) -> object: if self.size == self.capacity(): self.increase_capacity() self.items[self.size] = item self.size = self.size + 1 # Appends many items to end of vector def append_all(self:"Vector5", new_items: [int]) -> object: item:int = 0 for item in new_items: self.append(item) # Appends many items to end of vector def append_all2(self:"Vector5", new_items: [int], new_items2: [int]) -> object: item:int = 0 item2:int = 0 for item in new_items: self.append(item) # Appends many items to end of vector def append_all3(self:"Vector5", new_items: [int], new_items2: [int], new_items3: [int]) -> object: item:int = 0 item2:int = 0 item3:int = 0 for item in new_items: self.append(item) # Appends many items to end of vector def append_all4(self:"Vector5", new_items: [int], new_items2: [int], new_items3: [int], new_items4: [int]) -> object: item:int = 0 item2:int = 0 item3:int = 0 item4:int = 0 for item in new_items: self.append(item) # Appends many items to end of vector def append_all5(self:"Vector5", new_items: [int], new_items2: [int], new_items3: [int], new_items4: [int], new_items5: [int]) -> object: item:int = 0 item2:int = 0 item3:int = 0 item4:int = 0 item5:int = 0 for item in new_items: self.append(item) # Removes an item from the middle of vector def remove_at(self:"Vector5", idx: int) -> object: if idx < 0: return while idx < self.size - 1: self.items[idx] = self.items[idx + 1] idx = idx + 1 self.size = self.size - 1 # Removes an item from the middle of vector def remove_at2(self:"Vector5", idx: int, idx2: int) -> object: if idx < 0: return while idx < self.size - 1: self.items[idx] = self.items[idx + 1] idx = idx + 1 self.size = self.size - 1 # Removes an item from the middle of vector def remove_at3(self:"Vector5", idx: int, idx2: int, idx3: int) -> object: if idx < 0: return while idx < self.size - 1: self.items[idx] = self.items[idx + 1] idx = idx + 1 self.size = self.size - 1 # Removes an item from the middle of vector def remove_at4(self:"Vector5", idx: int, idx2: int, idx3: int, idx4: int) -> object: if idx < 0: return while idx < self.size - 1: self.items[idx] = self.items[idx + 1] idx = idx + 1 self.size = self.size - 1 # Removes an item from the middle of vector def remove_at5(self:"Vector5", idx: int, idx2: int, idx3: int, idx4: int, idx5: int) -> object: if idx < 0: return while idx < self.size - 1: self.items[idx] = self.items[idx + 1] idx = idx + 1 self.size = self.size - 1 # Retrieves an item at a given index def get(self:"Vector5", idx: int) -> int: return self.items[idx] # Retrieves an item at a given index def get2(self:"Vector5", idx: int, idx2: int) -> int: return self.items[idx] # Retrieves an item at a given index def get3(self:"Vector5", idx: int, idx2: int, idx3: int) -> int: return self.items[idx] # Retrieves an item at a given index def get4(self:"Vector5", idx: int, idx2: int, idx3: int, idx4: int) -> int: return self.items[idx] # Retrieves an item at a given index def get5(self:"Vector5", idx: int, idx2: int, idx3: int, idx4: int, idx5: int) -> int: return self.items[idx] # Retrieves the current size of the vector def length(self:"Vector5") -> int: return self.size # Retrieves the current size of the vector def length2(self:"Vector5") -> int: return self.size # Retrieves the current size of the vector def length3(self:"Vector5") -> int: return self.size # Retrieves the current size of the vector def length4(self:"Vector5") -> int: return self.size # Retrieves the current size of the vector def length5(self:"Vector5") -> int: return self.size # A faster (but more memory-consuming) implementation of vector class DoublingVector(Vector): doubling_limit:int = 1000 # Overriding to do fewer resizes def increase_capacity(self:"DoublingVector") -> int: if (self.capacity() <= self.doubling_limit // 2): self.items = self.items + self.items else: # If doubling limit has been reached, fall back to # standard capacity increases self.items = self.items + [0] return self.capacity() # A faster (but more memory-consuming) implementation of vector class DoublingVector2(Vector): doubling_limit:int = 1000 doubling_limit2:int = 1000 # Overriding to do fewer resizes def increase_capacity(self:"DoublingVector2") -> int: if (self.capacity() <= self.doubling_limit // 2): self.items = self.items + self.items else: # If doubling limit has been reached, fall back to # standard capacity increases self.items = self.items + [0] return self.capacity() # Overriding to do fewer resizes def increase_capacity2(self:"DoublingVector2") -> int: if (self.capacity() <= self.doubling_limit // 2): self.items = self.items + self.items else: # If doubling limit has been reached, fall back to # standard capacity increases self.items = self.items + [0] return self.capacity() # A faster (but more memory-consuming) implementation of vector class DoublingVector3(Vector): doubling_limit:int = 1000 doubling_limit2:int = 1000 doubling_limit3:int = 1000 # Overriding to do fewer resizes def increase_capacity(self:"DoublingVector3") -> int: if (self.capacity() <= self.doubling_limit // 2): self.items = self.items + self.items else: # If doubling limit has been reached, fall back to # standard capacity increases self.items = self.items + [0] return self.capacity() # Overriding to do fewer resizes def increase_capacity2(self:"DoublingVector3") -> int: if (self.capacity() <= self.doubling_limit // 2): self.items = self.items + self.items else: # If doubling limit has been reached, fall back to # standard capacity increases self.items = self.items + [0] return self.capacity() # Overriding to do fewer resizes def increase_capacity3(self:"DoublingVector3") -> int: if (self.capacity() <= self.doubling_limit // 2): self.items = self.items + self.items else: # If doubling limit has been reached, fall back to # standard capacity increases self.items = self.items + [0] return self.capacity() # A faster (but more memory-consuming) implementation of vector class DoublingVector4(Vector): doubling_limit:int = 1000 doubling_limit2:int = 1000 doubling_limit3:int = 1000 doubling_limit4:int = 1000 # Overriding to do fewer resizes def increase_capacity(self:"DoublingVector4") -> int: if (self.capacity() <= self.doubling_limit // 2): self.items = self.items + self.items else: # If doubling limit has been reached, fall back to # standard capacity increases self.items = self.items + [0] return self.capacity() # Overriding to do fewer resizes def increase_capacity2(self:"DoublingVector4") -> int: if (self.capacity() <= self.doubling_limit // 2): self.items = self.items + self.items $ElifElse return self.capacity() # Overriding to do fewer resizes def increase_capacity3(self:"DoublingVector4") -> int: if (self.capacity() <= self.doubling_limit // 2): self.items = self.items + self.items else: # If doubling limit has been reached, fall back to # standard capacity increases self.items = self.items + [0] return self.capacity() # Overriding to do fewer resizes def increase_capacity4(self:"DoublingVector4") -> int: if (self.capacity() <= self.doubling_limit // 2): self.items = self.items + self.items else: # If doubling limit has been reached, fall back to # standard capacity increases self.items = self.items + [0] return self.capacity() # A faster (but more memory-consuming) implementation of vector class DoublingVector5(Vector): doubling_limit:int = 1000 doubling_limit2:int = 1000 doubling_limit3:int = 1000 doubling_limit4:int = 1000 doubling_limit5:int = 1000 # Overriding to do fewer resizes def increase_capacity(self:"DoublingVector5") -> int: if (self.capacity() <= self.doubling_limit // 2): self.items = self.items + self.items else: # If doubling limit has been reached, fall back to # standard capacity increases self.items = self.items + [0] return self.capacity() # Overriding to do fewer resizes def increase_capacity2(self:"DoublingVector5") -> int: if (self.capacity() <= self.doubling_limit // 2): self.items = self.items + self.items else: # If doubling limit has been reached, fall back to # standard capacity increases self.items = self.items + [0] return self.capacity() # Overriding to do fewer resizes def increase_capacity3(self:"DoublingVector5") -> int: if (self.capacity() <= self.doubling_limit // 2): self.items = self.items + self.items else: # If doubling limit has been reached, fall back to # standard capacity increases self.items = self.items + [0] return self.capacity() # Overriding to do fewer resizes def increase_capacity4(self:"DoublingVector5") -> int: if (self.capacity() <= self.doubling_limit // 2): self.items = self.items + self.items else: # If doubling limit has been reached, fall back to # standard capacity increases self.items = self.items + [0] return self.capacity() # Overriding to do fewer resizes def increase_capacity5(self:"DoublingVector5") -> int: if (self.capacity() <= self.doubling_limit // 2): self.items = self.items + self.items else: # If doubling limit has been reached, fall back to # standard capacity increases self.items = self.items + [0] return self.capacity() # Makes a vector in the range [i, j) def vrange(i:int, j:int) -> Vector: v:Vector = None v = DoublingVector() while i < j: v.append(i) i = i + 1 return v def vrange2(i:int, j:int, i2:int, j2:int) -> Vector: v:Vector = None v2:Vector = None v = DoublingVector() while i < j: v.append(i) i = i + 1 return v def vrange3(i:int, j:int, i2:int, j2:int, i3:int, j3:int) -> Vector: v:Vector = None v2:Vector = None v3:Vector = None v = DoublingVector() while i < j: v.append(i) i = i + 1 return v def vrange4(i:int, j:int, i2:int, j2:int, i3:int, j3:int, i4:int, j4:int) -> Vector: v:Vector = None v2:Vector = None v3:Vector = None v4:Vector = None v = DoublingVector() while i < j: v.append(i) i = i + 1 return v def vrange5(i:int, j:int, i2:int, j2:int, i3:int, j3:int, i4:int, j4:int, i5:int, j5:int) -> Vector: v:Vector = None v2:Vector = None v3:Vector = None v4:Vector = None v5:Vector = None v = DoublingVector() while i < j: v.append(i) i = i + 1 return v # Sieve of Eratosthenes (not really) def sieve(v:Vector) -> object: i:int = 0 j:int = 0 k:int = 0 while i < v.length(): k = v.get(i) j = i + 1 while j < v.length(): if v.get(j) % k == 0: v.remove_at(j) else: j = j + 1 i = i + 1 def sieve2(v:Vector, v2:Vector) -> object: i:int = 0 i2:int = 0 j:int = 0 j2:int = 0 k:int = 0 k2:int = 0 while i < v.length(): k = v.get(i) j = i + 1 while j < v.length(): if v.get(j) % k == 0: v.remove_at(j) else: j = j + 1 i = i + 1 def sieve3(v:Vector, v2:Vector, v3:Vector) -> object: i:int = 0 i2:int = 0 i3:int = 0 j:int = 0 j2:int = 0 j3:int = 0 k:int = 0 k2:int = 0 k3:int = 0 while i < v.length(): k = v.get(i) j = i + 1 while j < v.length(): if v.get(j) % k == 0: v.remove_at(j) else: j = j + 1 i = i + 1 def sieve4(v:Vector, v2:Vector, v3:Vector, v4:Vector) -> object: i:int = 0 i2:int = 0 i3:int = 0 i4:int = 0 j:int = 0 j2:int = 0 j3:int = 0 j4:int = 0 k:int = 0 k2:int = 0 k3:int = 0 k4:int = 0 while i < v.length(): k = v.get(i) j = i + 1 while j < v.length(): if v.get(j) % k == 0: v.remove_at(j) else: j = j + 1 i = i + 1 def sieve5(v:Vector, v2:Vector, v3:Vector, v4:Vector, v5:Vector) -> object: i:int = 0 i2:int = 0 i3:int = 0 i4:int = 0 i5:int = 0 j:int = 0 j2:int = 0 j3:int = 0 j4:int = 0 j5:int = 0 k:int = 0 k2:int = 0 k3:int = 0 k4:int = 0 k5:int = 0 while i < v.length(): k = v.get(i) j = i + 1 while j < v.length(): if v.get(j) % k == 0: v.remove_at(j) else: j = j + 1 i = i + 1 # Input parameter n:int = 50 n2:int = 50 n3:int = 50 n4:int = 50 n5:int = 50 # Data v:Vector = None v2:Vector = None v3:Vector = None v4:Vector = None v5:Vector = None i:int = 0 i2:int = 0 i3:int = 0 i4:int = 0 i5:int = 0 # Crunch v = vrange(2, n) v2 = vrange(2, n) v3 = vrange(2, n) v4 = vrange(2, n) v5 = vrange(2, n) sieve(v) # Print while i < v.length(): print(v.get(i)) i = i + 1
[ "647530+Virtlink@users.noreply.github.com" ]
647530+Virtlink@users.noreply.github.com
3230d5448ef48ac2a50e98f9791b15a0ed770f9f
0147677b611e40ac695ba07f914264b3470a7401
/src/mac_address_info.py
4ad9f15c04cbbf5909df457315f089a8c5f1a0cb
[]
no_license
mblomdahl/sniffer
a2aed3ee37bb9a39d3c13ad8455ce7c7a2fc58c7
9101c59f958bb94fe1443fd90e95d333a02b785f
refs/heads/master
2021-01-24T00:23:30.318623
2015-08-14T12:56:33
2015-08-14T12:56:33
41,627,533
0
0
null
2015-08-30T12:11:01
2015-08-30T12:11:01
null
UTF-8
Python
false
false
5,190
py
import json import urllib2 import os class MacAddressInfo: def __init__(self): self.mac_address = "" self.company = "" self.address1 = "" self.address2 = "" self.address3 = "" self.country = "" class MacAddressStorage: def __init__(self): self.data = [] # creates a new empty list def mac_address_lookup_from_internet(self, mac_address): try: print "Load from Internet %s" % mac_address # Set the request URL http://www.macvendorlookup.com/api/v2/08-86-3B-D4-90-C0 url = 'http://www.macvendorlookup.com/api/v2/' + mac_address # Send the GET request response = urllib2.urlopen(url) resp = response.read() mac_object = MacAddressInfo data = [] if resp: # Interpret the JSON response #data = json.loads(resp.decode('utf8')) data = json.loads(resp) mac_object.mac_address = mac_address for company in data: mac_object.company = company['company'] for address1 in data: mac_object.address1 = address1['addressL1'] for address2 in data: mac_object.address2 = address2['addressL2'] for address3 in data: mac_object.address3 = address3['addressL3'] for country in data: mac_object.country = country['country'] else: mac_object.mac_address = mac_address mac_object.company = "" mac_object.address1 = "" mac_object.address2 = "" mac_object.address3 = "" mac_object.country = "" return mac_object except : print "Unexpected error:", url, resp return None def mac_address_lookup_from_cache(self, mac_address): try: self.load_data_from_file() count = len( self.data["mac addresses"] ) for index in range(count): if self.data["mac addresses"][index]["macaddress"] == mac_address: mac_object = MacAddressInfo mac_object.mac_address = mac_address mac_object.company = self.data["mac addresses"][index]["company"] mac_object.address1 = self.data["mac addresses"][index]["address1"] mac_object.address2 = self.data["mac addresses"][index]["address2"] mac_object.address3 = self.data["mac addresses"][index]["address3"] mac_object.country = self.data["mac addresses"][index]["country"] return mac_object return None except : print "mac_address_lookup_from_cache error:" return None def mac_address_lookup(self, mac_address): try: mac_object = self.mac_address_lookup_from_cache(mac_address) if mac_object is None : mac_object = self.mac_address_lookup_from_internet(mac_address) if mac_object is not None : #self.load_data_from_file() print mac_address self.data["mac addresses"].append( {"macaddress":mac_address, "company":mac_object.company, "address1":mac_object.address1, "address2":mac_object.address2, "address3":mac_object.address3, "country":mac_object.country} ) self.store_data_to_file() else : return None return mac_object except : print "mac_address_lookup error:" return None def load_data_from_file(self): if len( self.data ) == 0: if os.path.exists("/home/pi/sniffer/mac_addresses.json"): file_handel = open('/home/pi/sniffer/mac_addresses.json', 'r') self.data = json.load(file_handel) #print "Load" else: #file_handel = open('/home/pi/sniffer/mac_addresses.json', 'w') self.data.append( {"mac addresses":[]} ) #print "new" def store_data_to_file(self): file_handel = open('/home/pi/sniffer/mac_addresses.json', 'w') json.dump(self.data, file_handel, sort_keys=True, indent=2) #file_handel.write('\n') if __name__ == '__main__': storage = MacAddressStorage() mac_object = MacAddressInfo() #mac_object = storage.mac_address_lookup("08:86:3B:D4:90:C0") #mac_object = storage.mac_address_lookup("6C:F3:73:E6:0A:11") mac_object = storage.mac_address_lookup("9C:6C:15:97:76:04") #print storage.mac_address_lookup("08-86-3B-D4-90-C0").mac_address if mac_object : print mac_object.mac_address print mac_object.company print mac_object.address1 print mac_object.address2 print mac_object.address3 print mac_object.country else : print "Error"
[ "=" ]
=
49d5dabd0c8205c8eed6e8ec4291d98ee9d5549f
6ff024bfa3078c0e7f9f2f502c27bb066ae4a862
/ML_geron_1.py
f94f8cdbfe32e016666c5443242e0b396e8f9765
[]
no_license
emvil/MLtraining
920b0f6808e9b3c675a62c90c548608beae8c292
6c9f37d6c6151bcb75d07b3cd6ee684062c5a7ed
refs/heads/master
2020-06-25T01:28:15.043324
2019-08-26T20:11:07
2019-08-26T20:11:07
199,154,180
0
0
null
null
null
null
UTF-8
Python
false
false
3,377
py
# References: # Book: Hands_on Machine Learning with Scikit-Learn and TensorFlow, Aurélien Géron, 2017. # Chapter 2: End to end Machine Learning Project # Objective: predict the district's median housing price in any block group of California. # Current solution: manually estimated by experts, with a typical error of 15%. # Framework: supervised multivariate regression problem, using batch learning. # Performance measure: RMSE or MAE import os import tarfile from six.moves import urllib import pandas as pd import matplotlib.pyplot as plt from ML_tools import preprocessing from sklearn.preprocessing import Imputer, LabelEncoder, OneHotEncoder, LabelBinarizer # GET DATA # DOWNLOAD_ROOT = "https://github.com/ageron/handson-ml/tree/master/" HOUSING_PATH = "datasets/housing" # HOUSING_URL = DOWNLOAD_ROOT + HOUSING_PATH + "/housing.tgz" # -> downloaded def load_housing_data(housing_path=HOUSING_PATH): csv_path = os.path.join(housing_path, "housing.csv") return pd.read_csv(csv_path) # Explore data housing_dataset = load_housing_data() print(housing_dataset.head()) print(housing_dataset.info()) print(housing_dataset["ocean_proximity"].value_counts()) # stats for each variable print(housing_dataset.describe()) # display histograms of variables housing_dataset.hist(bins=50, figsize=(20, 15)) # plt.show() # CREATE A TEST SET housing_train, housing_test, tmp, tmp2 = preprocessing.split_data(housing_dataset, None, training_split_rate=0.80, display=True) # PREPARE THE DATA FOR ML ALGORITHMS (p.59) housing_labels = housing_train.loc[:, "median_house_value"].copy() housing = housing_train.drop(labels="median_house_value", axis=1) print(housing_labels.head()) print(housing.info()) # Data cleaning """ 3 options to deal with missing values - get rid of corresponding rows - get rid of the whole attribute (columns) - set the values to some value (zero, mean, median, etc) """ median_bedrooms = housing['total_bedrooms'].median() print(median_bedrooms) housing['total_bedrooms'].fillna(median_bedrooms, inplace=True) print(housing.info()) # Data cleaning : other method imputer = Imputer(strategy="median") # compute df with only numerical values housing_num = housing.drop('ocean_proximity', axis=1) # fit the imputer with the training data imputer.fit(housing_num) print(imputer.statistics_) print(housing_num.median().values) # apply the imputer to all the numerical values (result is a numpy array) X = imputer.transform(housing_num) housing_tr = pd.DataFrame(X, columns=housing_num.columns) # HANDLING TEXT AND CATEGORICAL ATTRIBUTES (p.62) # to convert text labels into numbers encoder = LabelEncoder() housing_cat = housing["ocean_proximity"] housing_cat_encoded = encoder.fit_transform(housing_cat) print(housing_cat_encoded) print(encoder.classes_) # One-hot encoding : binary classes from the encoded category vector, output is sparse matrix encoder = OneHotEncoder() housing_cat_1hot = encoder.fit_transform(housing_cat_encoded.reshape(-1, 1)) print(housing_cat_1hot) # do integer encoding and binary encoding more quickly, output is a dense numpy array by default # unless you add sparse_output=True encoder = LabelBinarizer() housing_cat_1hot = encoder.fit_transform(housing_cat) print(housing_cat_1hot) # CUSTOM TRANSFORMERS (p.64)
[ "e.villeneuve@outlook.com" ]
e.villeneuve@outlook.com
4d6b4640228f33cec47a54841e46353e8e26d85d
94f4b8a12cc09e3056cfc8d5304e5937b33ea6ec
/StemmingGUI.py
df3b3186cf7d4574b4fdb63f3d24f57296236b3b
[]
no_license
Ehtisham09/Machine-Learning-Projects
8f1aaa7489bd84333491dbab8432cc76ed62915a
dd09308e555cc7aee0db74d91af6f5140e41d689
refs/heads/master
2020-07-13T03:12:27.513696
2019-10-20T18:25:43
2019-10-20T18:25:43
204,975,654
0
0
null
null
null
null
UTF-8
Python
false
false
2,634
py
from tkinter import * from tkinter import filedialog from tkinter import messagebox import tkinter.scrolledtext as tkscrolled root = Tk() root.title("Steming") root.geometry("1160x600") root.resizable(0,0) def showOriginalText(): pass def showStopWords(): pass def showUniqueWords(): pass def showInfix(): pass def showPrefix(): pass def showPostfix(): pass def customization(): pass def showPostProcessing(): pass tabbuttons = Frame(root) b1 = Button(tabbuttons,text="Original Text", command=showOriginalText, height=1, width=20) b1.grid(row=1, column=0) b2 = Button(tabbuttons,text="Stop Words", command=showStopWords, height=1, width=20) b2.grid(row=1, column=1) b3 = Button(tabbuttons,text="Unique Words", command=showUniqueWords, height=1, width=20) b3.grid(row=1, column=2) b4 = Button(tabbuttons,text="Prefix", command=showPrefix, height=1, width=20) b4.grid(row=1, column=3) b5 = Button(tabbuttons,text="Postfix", command=showPostfix, height=1, width=20) b5.grid(row=1, column=4) b6 = Button(tabbuttons,text="Post-Processing", command=showPostProcessing, height=1, width=20) b6.grid(row=1, column=5) b7 = Button(tabbuttons,text="Infix", command=showInfix, height=1, width=20) b7.grid(row=1, column=6) tabbuttons.grid(row=1, pady=(30,0)) textbox = tkscrolled.ScrolledText(root, height=20, width=132) textbox.grid(row=2, pady=(0,20), padx=50) def InputFile(): # Function For Browsing File root.filename = filedialog.askopenfilename(title = "Select File", filetypes = [('Text files', '*.txt')]) f = open(root.filename,encoding="utf8") content = f.read() # data = str(content) print(type(content)) textbox.insert(INSERT,content) # print(content) # data = content.split("۔") def stemData(): pass buttons = Frame(root) # Three Main Buttons Frame clear = Button(buttons, text= "Clear" , command= lambda: textbox.delete(1.0,END), height=2, width=20) # Clear Button browsebutton = Button(buttons, text ="Browse From Computer",command = InputFile, height=2) # Browse Button browsebutton.grid(row=3,column=1, padx=4) clear.grid(row=3,column=2, padx=4) buttons.grid() submitButton = Button(root, text="Stemming", command= stemData, width= 20, height= 2, bg = "yellow", font='bold') # Submit Button submitButton.config(font=("Calibri", 15)) submitButton.grid(pady=(20,15)) root.mainloop()
[ "noreply@github.com" ]
Ehtisham09.noreply@github.com
f6d96854d2e988c2c0d5f5a1574b5cc6c67840af
f31630adc1e677065975a4b57902db5e0700e4e9
/Exploratory Analysis.py
a0e87bf56446b76e8896020163885234c2d39435
[]
no_license
kkoehncke/ML-Cheat-Sheets
8a5847a88b49e8d6b92ce5168a637347453fb3c5
2e85180a432a3b4b9aa591f6ac65dcbfe7721d2c
refs/heads/master
2020-03-11T20:08:33.822196
2018-04-19T14:40:55
2018-04-19T14:40:55
130,228,713
0
1
null
null
null
null
UTF-8
Python
false
false
1,689
py
Exploratory Analysis: # Dataframe dimensions df.shape # Column datatypes df.dtypes # Summarize numerical features df.describe() # Summarize categorical features df.describe(include=['object']) #Display first 5 rows; can speciy number inside to show n rows df.head() # Display last 5 rows of data df.tail() # Filter and display only df.dtypes that are 'object' df.dtypes[df.dtypes == 'object'] # Segment by <> and display the means within each class; can do the same with .std() df.groupby('<>').mean() # Segment by <> and display the means and standard deviations within each class df.groupby('<>').agg([np.mean, np.std]) # Loop through categorical feature names and print each one for feature_names in df.dtypes[df.dtypes == 'object'].index: print (feature_names) # Plot bar plot for each categorical feature for feature_names in df.dtypes[df.dtypes == 'object'].index: sns.countplot(y = feature_names, data=df) plt.show() # Plot histogram grid df.hist(figsize=(14,14), xrot=-45) # Clear the text "residue" plt.show() # Bar plot for '<insert column name>' sns.countplot(y = '<>', data=df) # Boxplot of <> and <> sns.boxplot(x = '<>', y = '<>', data = df) # Violinplot of <> and <> sns.violinplot(y = '<>', x = '<>', data = df) # Make the figsize 10 x 8 plt.figure(figsize=(9,8)) # Plot heatmap of annotated correlations sns.heatmap(correlations*100,annot = True ,fmt='.0f', cbar=False) #For classification problems (bivariate) sns.lmplot(x='<>', y='<>', hue='<binary target variable>', data=df, fit_reg=False) # If we want scatter of only one of the target variables sns.lmplot(x='<>', y='<>', data=df[df.<target column> == '<target value>'], fit_reg=False)
[ "kkoehncke@captechventures.com" ]
kkoehncke@captechventures.com
1f905bbe4853784c2555aed917cc9767ada33140
5d0986d930daf567100130c66ae403e0d0cd043f
/train_segmentation.py
8f4eb9123b4c3a0861ff829789f87178289f8f38
[]
no_license
and-jonas/HaloDetector
40d8f7d73cdd89f6c8aa51731ea29d79909a7ab1
ba0a1bd6f9d060d5ba1e035cb63d7d16a513e60c
refs/heads/master
2023-09-06T09:54:26.983910
2021-11-11T10:04:02
2021-11-11T10:04:02
426,910,093
0
0
null
null
null
null
UTF-8
Python
false
false
6,049
py
# ====================================================================================================================== # Author: Jonas Anderegg, jonas.anderegg@usys.ethz.ch # Extract training data from labelled patches and train segmentation algorithm # Last modified: 2021-11-10 # ====================================================================================================================== from pathlib import Path import pickle import numpy as np import pandas as pd import glob from sklearn.ensemble import RandomForestClassifier from sklearn.model_selection import RandomizedSearchCV # Number of trees in random forest from sklearn.model_selection import GridSearchCV # Create the parameter grid based on the results of random search import SegmentationFunctions # ====================================================================================================================== # (1) extract features and save to .csv # ====================================================================================================================== workdir = 'Z:/Public/Jonas/001_LesionZoo/MRE' # set directories to previously selected training patches dir_positives = f'{workdir}/train_data_segmentation/Positives' dir_negatives = f'{workdir}/train_data_segmentation/Negatives' # extract feature data for all pixels in all patches # output is stores in .csv files in the same directories SegmentationFunctions.iterate_patches(dir_positives, dir_negatives) # ====================================================================================================================== # (2) combine training data from all patches into single file # ====================================================================================================================== # import all training data # get list of files files_pos = glob.glob(f'{dir_positives}/*.csv') files_neg = glob.glob(f'{dir_negatives}/*.csv') all_files = files_pos + files_neg # load data train_data = [] for i, file in enumerate(all_files): print(i) data = pd.read_csv(file) data = data.iloc[::10, :] # only keep every 10th pixel of the patch train_data.append(data) # to single df train_data = pd.concat(train_data) # export, this may take a while train_data.to_csv(f'{workdir}/train_data_segmentation/training_data.csv', index=False) # ====================================================================================================================== # (3) train random forest classifier # ====================================================================================================================== train_data = pd.read_csv(f'{workdir}/train_data_segmentation/training_data.csv') # OPTIONAL: sample an equal number of rows per class for training n_pos = train_data.groupby('response').count().iloc[0, 0] n_neg = train_data.groupby('response').count().iloc[1, 0] n_min = min(n_pos, n_neg) train_data = train_data.groupby(['response']).apply(lambda grp: grp.sample(n=n_min)) n_estimators = [int(x) for x in np.linspace(start=20, stop=200, num=10)] # Number of features to consider at every split max_features = ['auto', 'sqrt'] # Maximum number of levels in tree max_depth = [int(x) for x in np.linspace(10, 110, num=11)] max_depth.append(None) # Minimum number of samples required to split a node min_samples_split = [2, 5, 10] # Minimum number of samples required at each leaf node min_samples_leaf = [1, 2, 4, 8] # Method of selecting samples for training each tree bootstrap = [True, False] # Create the random grid random_grid = {'n_estimators': n_estimators, 'max_features': max_features, 'max_depth': max_depth, 'min_samples_split': min_samples_split, 'min_samples_leaf': min_samples_leaf, 'bootstrap': bootstrap} # Use the random grid to search for best hyperparameters # First create the base model to tune rf = RandomForestClassifier() # Random search of parameters, using 3 fold cross validation, # search across 100 different combinations, and use all available cores rf_random = RandomizedSearchCV(estimator=rf, param_distributions= random_grid, n_iter=100, cv=10, verbose=3, random_state=42, n_jobs=-1) # Fit the random search model # predictor matrix X = np.asarray(train_data)[:, 0:21] # response vector y = np.asarray(train_data)[:, 21] model = rf_random.fit(X, y) rf_random.best_params_ best_random = rf_random.best_estimator_ param_grid = { 'bootstrap': [True], 'max_depth': [60, 70, 80], 'max_features': [2, 4, 6, 8, 10], 'min_samples_leaf': [2, 4, 6, 8], 'min_samples_split': [8, 10, 12], 'n_estimators': [200, 300, 400] } # Create a based model rf = RandomForestClassifier() # Instantiate the grid search model grid_search = GridSearchCV(estimator=rf, param_grid=param_grid, cv=10, n_jobs=-1, verbose=3) # Fit the grid search to the data grid_search.fit(X, y) grid_search.best_params_ # specify model hyper-parameters clf = RandomForestClassifier( max_depth=80, # maximum depth of 95 decision nodes for each decision tree max_features=2, # maximum of 6 features (channels) are considered when forming a decision node min_samples_leaf=8, # minimum of 6 samples needed to form a final leaf min_samples_split=12, # minimum 4 samples needed to create a decision node n_estimators=200, # maximum of 55 decision trees bootstrap=False, # don't reuse samples random_state=1, n_jobs=-1 ) # fit random forest model = clf.fit(X, y) score = model.score(X, y) # save model path = f'{workdir}/Output/Models' if not Path(path).exists(): Path(path).mkdir(parents=True, exist_ok=True) pkl_filename = f'{path}/rf_segmentation.pkl' with open(pkl_filename, 'wb') as file: pickle.dump(model, file) # ======================================================================================================================
[ "anjonas@student.ethz.ch" ]
anjonas@student.ethz.ch
720b1ea6f006745bee5abcb72de9509ce028d5e6
a73cbe345107af388253ab3437992f09b1a033ed
/Aula 8/Aula8-Tuplas.py
04c116c1b911114ca5cfc5c9a746ea0c7bd79cc6
[ "MIT" ]
permissive
ohanamirella/TrabalhosPython
09c586f897cf4ae33a1e49c8340883b17df832d2
453a97848654b0391d0d717bf102f6c466f79b3d
refs/heads/master
2020-09-05T22:54:46.245904
2020-02-03T14:51:39
2020-02-03T14:51:39
220,237,145
1
0
null
null
null
null
UTF-8
Python
false
false
335
py
# Aula 8 - 18-11-2019 # Tuplas numeros = [1,4,6] usuario = {'nome':'user', 'passwd':123456 } pessoa = ('maykon','granemann',0, 45.5, numeros) # print(numeros) # print(usuario) # print(pessoa) lista = [1] numeros[1] = 5 usuario['passwd'] = 456123 lista_pessoas = [] lista_pessoas.append(pessoa) #pessoa[4][1] = 6 print(pessoa[4][1])
[ "900153@proway.treina" ]
900153@proway.treina
aa787dba2580b471a9d3b39d3d1b56d66c681e27
2d09e7e0b8c11e7a604580e54eb980fe7c97cffb
/ArrayOperation/canFinish.py
93d31a2bfc9c42d83e9db164208fbb1a546f3466
[]
no_license
Philex5/codingPractice
8b67e53186c8f087c02214008269b3eb5d235d44
95dddb78bccd169d9d219a473627361fe739ab5e
refs/heads/master
2020-04-24T02:22:05.663438
2019-10-09T02:19:25
2019-10-09T02:19:25
171,634,380
0
0
null
null
null
null
UTF-8
Python
false
false
2,652
py
# -*- coding: utf-8 -*- # @Time : 2019/6/14 下午9:22 # @Author : Philex # @File : canFinish.py # @Software: PyCharm """ 给定一个整数数组和一个整数 k,你需要找到该数组中和为 k 的连续的子数组的个数。 示例 1 : 输入:nums = [1,1,1], k = 2 输出: 2 , [1,1] 与 [1,1] 为两种不同的情况。 说明 : 数组的长度为 [1, 20,000]。 数组中元素的范围是 [-1000, 1000] ,且整数 k 的范围是 [-1e7, 1e7]。 """ class Solution: # def subarraySum(self, nums, k): # """ # 思路:连续子数组的和为一确定值 # 暴力法: 确定开始位置,向后遍历,累计从开始位置到当前位置的数字之和, 等于加入解集,大于终止, # 把当前开始位置下一个位置作为开始位置 # 考虑的情况有点复杂,比如最后一个数字等于k,j=l不符合要求,不会进入while循环 # 还是会少考虑一些情况 # 直接简化,不用while,虽然会引入更多的计算 # 时间复杂度为: O(n^2) # 空间复杂度为: O(1) # 超出时间限制 # """ # if not nums or len(nums) == 0: # return 0 # l = len(nums) # res = 0 # for i in range(l): # sums = 0 # for j in range(i, l): # sums += nums[j] # res += sums == k # return res def subarraySum(self, nums, k): """ 思路: cumulative纪录从第一个位置到数组其他位置的和, cumulative[j] - cumulative[i]表示第j个元素 到第i个元素的和. 然后通过双重循环计算所有 两个元素之间的累计和 """ if nums is None or len(nums) == 0: return 0 l = len(nums) cumu = [0] * (l+1) res = 0 for i in range(1, l+1): cumu[i] = cumu[i-1] + nums[i-1] for i in range(l): for j in range(i+1, l): res += cumu[j+1] - cumu[i] == k return res def subarraySum1(self, nums, k): """ 思路: 哈希表 利用字典存储 sum-k 出现的次数, sum-k = n -> sum = k + n, n存在, k肯定也存在 注意初始化0:1,保证结果能够累加 """ sum, res, cul = 0, 0, {} cul[0] = 1 for i in range(len(nums)): sum += nums[i] if sum - k in cul: res += cul[sum-k] if sum not in cul: cul[sum] = 0 cul[sum] += 1 return res so = Solution() print(so.subarraySum1([1, 2, 3, 5], 3))
[ "Philexwoo@gmail.com" ]
Philexwoo@gmail.com
f4e31986896cc7a6b49e731bcc2f5d073c79c3e2
0a643bb9bb8e864a439b7302ac9432eae8313de1
/main_app/migrations/0001_initial.py
3bbcc8ba2afffa21afce1c97cae20d1a61f24659
[]
no_license
jamesgarcia2020/assess_3
f396a036cb04bb2f2b5b51c02702c8aabf0660ee
2644481ca7affa66fa7258dfc31f2b1443e51c43
refs/heads/master
2022-12-17T14:08:44.615862
2020-09-25T21:39:59
2020-09-25T21:39:59
298,646,492
0
0
null
null
null
null
UTF-8
Python
false
false
545
py
# Generated by Django 3.1.1 on 2020-09-25 18:54 from django.db import migrations, models class Migration(migrations.Migration): initial = True dependencies = [ ] operations = [ migrations.CreateModel( name='Widget', fields=[ ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), ('description', models.CharField(max_length=50)), ('quantity', models.IntegerField()), ], ), ]
[ "jamesgarcia@Jamess-MacBook-Air.local" ]
jamesgarcia@Jamess-MacBook-Air.local
4d6647ad50459cf616a4eaaa782651b3b18edf2d
1af1f89eb9a178b95d1ba023b209b7538fb151f0
/Algorithms/146. LRU Cache.py
09989dad11e2256fc8b9ce1d4d8a754a15563be9
[]
no_license
0xtinyuk/LeetCode
77d690161cc52738e63a4c4b6595a6012fa5c21e
08bc96a0fc2b672282cda348c833c02218c356f1
refs/heads/master
2023-02-21T16:58:39.881908
2021-01-25T08:00:13
2021-01-25T08:00:13
292,037,842
0
0
null
null
null
null
UTF-8
Python
false
false
826
py
from collections import OrderedDict class LRUCache(): def __init__(self, capacity): self.memory = OrderedDict() self.capacity = capacity def get(self, key): if key not in self.memory: return - 1 self.memory.move_to_end(key) return self.memory[key] def put(self, key, value): if key in self.memory: self.memory.move_to_end(key) self.memory[key] = value if len(self.memory) > self.capacity: self.memory.popitem(last = False) # Your LRUCache object will be instantiated and called as such: # obj = LRUCache(capacity) # param_1 = obj.get(key) # obj.put(key,value) # Your LRUCache object will be instantiated and called as such: # obj = LRUCache(capacity) # param_1 = obj.get(key) # obj.put(key,value)
[ "xliu301@uottawa.ca" ]
xliu301@uottawa.ca
e64030f4bfdc9f2ecd066eaf1ad8e5e2b067c849
e0eb81aef84ee0929aa3dfc166f29a343251c35b
/seafile-pro-server-7.0.10/pro/python/seafevents/tests/conftest.py
eadb20a149347ce0520fe528caab3c2e5768a7bc
[]
no_license
Sandra-Z/filesharing
414fc56abe2f87b80ea390e0814a0bf86148a2bf
0e4e637f0c78f96949796b480b51df72d859c4ff
refs/heads/master
2022-12-11T12:28:07.155281
2019-11-19T08:39:51
2019-11-19T08:39:51
null
0
0
null
null
null
null
UTF-8
Python
false
false
6,559
py
import os import sys import ConfigParser import subprocess from sqlalchemy import create_engine, text from sqlalchemy.exc import DisconnectionError from sqlalchemy.event import contains as has_event_listener, listen as add_event_listener from urllib import quote_plus from pytest import yield_fixture from sqlalchemy.pool import Pool from sqlalchemy.orm import sessionmaker SEAHUB_DBNAME = '' SEAFEVENTS_DBNAME = '' TEST_DBNAME = '' @yield_fixture(scope="module") def test_db(): delete_all_table_if_exists() # copy_db_from_seahub_with_no_data() # copy_db_from_seafevent_with_no_data() apply_tables() yield None # delete_all_table_if_exists() def generate_tables_sql(): seahub_db = read_db_conf('SEAHUBDB') seafevents_db = read_db_conf('SEAFEVENTSDB') connection_data = [seahub_db[0]] connection_data.extend(seahub_db[2:]) connection_data = tuple(connection_data) cmd = "mysqldump -h%s -u%s -p%s --skip-add-locks --no-data --skip-add-drop-table --skip-comments %s > seahub.sql" % connection_data cwd = ["bash", "-c", cmd] subprocess.check_call(cwd, stdout=None, stderr=None) connection_data = [seafevents_db[0]] connection_data.extend(seafevents_db[2:]) connection_data = tuple(connection_data) cmd = "mysqldump -h%s -u%s -p%s --skip-add-locks --no-data --skip-add-drop-table --skip-comments %s > seafevents.sql" % connection_data cwd = ["bash", "-c", cmd] subprocess.check_call(cwd, stdout=None, stderr=None) merge_sql_file('raw_table_sql.sql') def merge_sql_file(filename): with open(filename, 'w') as fp: for fname in ['seahub.sql', 'seafevents.sql']: with open(fname) as tfp: fp.write(tfp.read()) fp.write('\n') def apply_tables(): seafevents_db = read_db_conf('TESTDB') full_path = os.path.join(os.path.dirname(os.path.realpath(__file__)), 'raw_table_sql.sql') cmd = "mysql -h %s -u%s -p%s %s < %s" % (seafevents_db[0], seafevents_db[2], seafevents_db[3], seafevents_db[4], full_path) cwd = ["bash", "-c", cmd] try: subprocess.check_call(cwd, stdout=subprocess.PIPE, stderr=subprocess.STDOUT) except Exception as e: print e.output def delete_all_table_if_exists(): session = None try: session = get_db_session('TESTDB') session = get_db_session('TESTDB') sql = text('SET FOREIGN_KEY_CHECKS = 0;') session.execute(sql) sql = text('SELECT table_name FROM information_schema.tables where table_schema= :db;') tables = session.execute(sql, {'db': TEST_DBNAME}).fetchall() if tables: for tablename in tables: del_sql = text('drop table %s' % tablename[0]) session.execute(del_sql) sql = text('SET FOREIGN_KEY_CHECKS = 1;') session.execute(sql) except Exception as e: sys.stdout.write(str(e)) finally: if session: session.close() def copy_db_from_seahub_with_no_data(): test_session = None seahub_session = None try: test_session = get_db_session('TESTDB') seahub_session = get_db_session('SEAHUBDB') sql = text('SELECT table_name FROM information_schema.tables where table_schema= :db') tables = seahub_session.execute(sql, {'db': SEAHUB_DBNAME}).fetchall() if tables: for t_name in tables: create_sql = text('create table %s like %s' % (t_name[0], "{0}.{1}".format(SEAHUB_DBNAME, t_name[0]))) test_session.execute(create_sql) except Exception as e: sys.stdout.write(str(e)) finally: if seahub_session: seahub_session.close() if test_session: test_session.close() def copy_db_from_seafevent_with_no_data(): test_session = None seahub_session = None try: test_session = get_db_session('TESTDB') seahub_session = get_db_session('SEAFEVENTSDB') sql = text('SELECT table_name FROM information_schema.tables where table_schema= :db') tables = seahub_session.execute(sql, {'db': SEAFEVENTS_DBNAME}).fetchall() if tables: for t_name in tables: create_sql = text('create table %s like %s' % (t_name[0], "{0}.{1}".format(SEAFEVENTS_DBNAME, t_name[0]))) test_session.execute(create_sql) except Exception as e: sys.stdout.write(str(e)) finally: if seahub_session: seahub_session.close() if test_session: test_session.close() def get_db_session(section): config = ConfigParser.ConfigParser() config.read('./db.cnf') if not config.has_section(section): sys.stdout.write("no section: %s" % section) return host, port, username, passwd, dbname = read_db_conf(section) db_url = "mysql+mysqldb://%s:%s@%s:%s/%s?charset=utf8" % (username, quote_plus(passwd), host, port, dbname) global SEAHUB_DBNAME, SEAFEVENTS_DBNAME, TEST_DBNAME if section == 'TESTDB': TEST_DBNAME = dbname elif section == 'SEAFEVENTSDB': SEAFEVENTS_DBNAME = dbname elif section == 'SEAHUBDB': SEAHUB_DBNAME = dbname kwargs = dict(pool_recycle=300, echo=False, echo_pool=False) engine = create_engine(db_url, **kwargs) if not has_event_listener(Pool, 'checkout', ping_connection): add_event_listener(Pool, 'checkout', ping_connection) Session = sessionmaker(bind=engine) return Session() def read_db_conf(section): config = ConfigParser.ConfigParser() config.read('./db.cnf') if not config.has_section(section): sys.stdout.write("no section: %s" % section) return if config.has_option(section, 'host'): host = config.get(section, 'host').lower() else: host = 'localhost' if config.has_option(section, 'port'): port = config.getint(section, 'port') else: port = 3306 username = config.get(section, 'username') passwd = config.get(section, 'password') dbname = config.get(section, 'name') return (host, port, username, passwd, dbname) def ping_connection(dbapi_connection, connection_record, connection_proxy): # pylint: disable=unused-argument cursor = dbapi_connection.cursor() try: cursor.execute("SELECT 1") cursor.close() except: connection_proxy._pool.dispose() # pylint: disable=protected-access # Raise DisconnectionError so the pool would create a new connection raise DisconnectionError()
[ "deutschland.gray@gmail.com" ]
deutschland.gray@gmail.com
85b871b51e585a914eea3800d452e2101a966e14
e211fdfc6fe8b79840409f7e2a2ee5e738bf9393
/main/migrations/0002_wishlist.py
4fcac7a1a83bc649d9aa74f9fdd14ab4a171ac52
[]
no_license
Kuuhaku11/wishlist
ec28416c628d1df2c1e4a4f2ec8f767e255c1d3f
c97346c30c364da30d224edccf87a548b396a24c
refs/heads/master
2023-08-14T23:18:30.906645
2021-09-13T19:04:40
2021-09-13T19:04:40
404,098,496
0
0
null
null
null
null
UTF-8
Python
false
false
897
py
# Generated by Django 3.2.7 on 2021-09-10 12:44 from django.conf import settings from django.db import migrations, models import django.db.models.deletion class Migration(migrations.Migration): dependencies = [ migrations.swappable_dependency(settings.AUTH_USER_MODEL), ('main', '0001_initial'), ] operations = [ migrations.CreateModel( name='Wishlist', fields=[ ('id', models.BigAutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), ('title', models.CharField(max_length=120)), ('is_hidden', models.BooleanField(default=True)), ('owner', models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to=settings.AUTH_USER_MODEL)), ('product', models.ManyToManyField(to='main.Product')), ], ), ]
[ "kuuhaku112121@gmail.com" ]
kuuhaku112121@gmail.com
b5ca6f0bd9969bb338ebb2304402697b16cd66b8
78eea6b9b9ebada2d02360ff8747c804e4d3f2db
/load_image.py
0ed8141232b37f4d890dce638696d3666a93d76d
[]
no_license
erkanercan/pyLogicGate
74c81cd714dff969c3aae496aab3cc297f6c295f
df19777da4299720131565eece40eebb93546a6c
refs/heads/master
2021-09-10T15:41:45.197118
2018-03-28T18:56:51
2018-03-28T18:56:51
null
0
0
null
null
null
null
UTF-8
Python
false
false
501
py
import pygame,os def load_image(name, colorkey=None): fullname = os.path.join('data', name) print(name) try: image = pygame.image.load(fullname) except pygame.error as message: print('Cannot load image:', fullname) raise SystemExit(message) image = image.convert() if colorkey is not None: if colorkey is -1: colorkey = image.get_at((0,0)) image.set_colorkey(colorkey, pygame.RLEACCEL) return image, image.get_rect()
[ "abajuradam@yahoo.com" ]
abajuradam@yahoo.com
d090e5080697eb9ddc699d37c4656032fc8ef74a
31f85926c1bbafdb0621a43b320f48be2a1090ff
/matrix-cuda/extract_gpu.py
315584be16bc7aedece6334b2b66eb32c6cc9b13
[]
no_license
subratpp/jetson_yolo
f1c3e32812c0c9c65883a4d58b817f5c0bdcc833
7e4c0edb55a70353a86e733914819077903b3f00
refs/heads/main
2023-03-03T10:28:01.331540
2021-02-12T17:17:22
2021-02-12T17:17:22
337,465,387
0
0
null
null
null
null
UTF-8
Python
false
false
342
py
import re def gpu_workload(filename): max_use = 0 with open(filename) as fp: lines = fp.readlines() for line in lines: gpu = re.search(r'GR3D_FREQ (.*?)%', line).group(1) if int(gpu) > max_use: max_use = float(gpu) return max_use print(gpu_worklaod("matmul1000.txt"))
[ "subratprasad.mail@gmail.com" ]
subratprasad.mail@gmail.com