blob_id stringlengths 40 40 | directory_id stringlengths 40 40 | path stringlengths 3 281 | content_id stringlengths 40 40 | detected_licenses listlengths 0 57 | license_type stringclasses 2 values | repo_name stringlengths 6 116 | snapshot_id stringlengths 40 40 | revision_id stringlengths 40 40 | branch_name stringclasses 313 values | visit_date timestamp[us] | revision_date timestamp[us] | committer_date timestamp[us] | github_id int64 18.2k 668M ⌀ | star_events_count int64 0 102k | fork_events_count int64 0 38.2k | gha_license_id stringclasses 17 values | gha_event_created_at timestamp[us] | gha_created_at timestamp[us] | gha_language stringclasses 107 values | src_encoding stringclasses 20 values | language stringclasses 1 value | is_vendor bool 2 classes | is_generated bool 2 classes | length_bytes int64 4 6.02M | extension stringclasses 78 values | content stringlengths 2 6.02M | authors listlengths 1 1 | author stringlengths 0 175 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
698eef67e3fa9abc62d773f0930fe68d851e7987 | 84b99a8efdf30d7e80a1a7dc2e25970de30b3a3f | /mockGrill.py | 97f07f81bde9858105834e36584b476041c2fc56 | [] | no_license | netgio/mockGrill | 3d6a3253fc14d7af48671c7453f940d3fc81dedc | f8006ed037810923b4a5fd310bf8db968c8b04b1 | refs/heads/master | 2021-01-19T10:34:49.362756 | 2015-02-21T18:20:30 | 2015-02-21T18:20:30 | 31,134,773 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 1,382 | py | #!/usr/bin/python
import sys
import subprocess
import os
import json
import time
prev = 0
current = 0
path = "./media"
picPath = "./api/static/shot.jpg"
print "Running...path: " + path
while (1):
## reload the status from disk
print "Check grill status..."
f = open("status.json")
status = json.load(f)
print "Status: " + str(status)
## read status from JSON data
grillLid = status['lid'] ### "open" or "closed"
grillRotisserie = status['rotisserie'] ### "on" or "off"
grillBurner = status['burner'] ### or "high", "med" or "low"
### compose filename for correct video
grillVideo = "grill_" + grillLid + "_" + grillRotisserie + "_" + grillBurner + ".m4v"
grillVideoPath = os.path.join(path, grillVideo)
print "Playing: " + grillVideoPath
### a = subprocess.Popen( [ "omxplayer" , "-o", "hdmi", grillVideoPath ] )
current = subprocess.Popen( [ "omxplayer", "--win", "50 250 1000 750", grillVideoPath ], stdin=subprocess.PIPE, stdout=subprocess.PIPE )
print "Playback Procees: " + str(current.pid)
### splitting the sleep allows the process to shutdown gracefully
time.sleep(1) #all videos are about 5 seconds long so sleep and leave 1 second overlap
#picProc = subprocess.Popen( [ "fswebcam", "-D", "1", picPath ], stdin=subprocess.PIPE, stdout=subprocess.PIPE )
time.sleep(3.1) #all videos are about 5 seconds long so sleep and leave 1 second overlap
| [
"gary@netgio.com"
] | gary@netgio.com |
025e92b046f4c40d049ff8c62b670e257da64222 | 698e105281d34806561ef8c466a12a5dc6c67132 | /web/frontend/submissions.py | fbe7f583a1d9d0c2de5c629d0221ebc445b29933 | [] | no_license | Podshot/3DPrinterSystem | 75c8dcafc5f926415efd28a6b1ce24458d1e217c | bbf225ad76e29f783f13d543fcd72b94d4640db2 | refs/heads/master | 2021-01-01T05:10:32.567917 | 2016-12-24T17:27:52 | 2016-12-24T17:27:52 | 59,046,789 | 2 | 0 | null | null | null | null | UTF-8 | Python | false | false | 3,345 | py | import tornado.web
from profile import BaseProfileHandler
from utils import SQLWrapper, directories, DropboxWrapper # @UnresolvedImport
import os
from datetime import datetime
import uuid
from gzip import GzipFile
class NewSubmissionHandler(BaseProfileHandler):
@tornado.web.authenticated
def get(self):
name = tornado.escape.xhtml_escape(self.current_user)
account = SQLWrapper.get_account(name)
self.render("new_submission.html", is_robotics=account.is_robotics)
class SubmitHandler(BaseProfileHandler):
detail_map = {"high": 0.1, "normal": 0.2, "low": 0.3}
@tornado.web.authenticated
def post(self):
name = tornado.escape.xhtml_escape(self.current_user)
info = self.request.files["file"][0]
filename = info["filename"]
filename = tornado.web.escape.xhtml_escape(os.path.basename(filename)).replace("$$", "_._")
self.redirect("profile")
if filename.lower().endswith(".stl") and info["content_type"] == "application/octet-stream":
submission_data = {
"title": self.get_argument("title"),
"date": datetime.strftime(datetime.now(), "%B %d, %Y at %I:%M %p"),
"status": "pending",
"priority": 1,
"options": {
"detail_name": self.get_argument("quality"),
"detail_height": float(self.detail_map.get(self.get_argument("quality"), self.get_argument("custom_quality"))),
"color": self.get_argument("color"),
"rafts": (True if self.get_argument("rafts", default="off") == "on" else False),
"supports": (True if self.get_argument("supports", default="off") == "on" else False),
"infill": int(self.get_argument("infill")),
},
"assignment": {},
"for_robotics": (True if self.get_argument("for_robotics", default="off") == "on" else False),
}
if self.get_argument("assignment", default="off") == "on":
submission_data["assignment"]["class_name"] = self.get_argument("class_name")
submission_data['assignment']['teacher'] = self.get_argument('teacher')
date = self.get_argument('due_date').split("-")
submission_data['assignment']['due_date'] = "{}-{}-{}".format(date[1], date[2], date[0])
submission_id = str(uuid.uuid4())
SQLWrapper.add_submission(submission_id, name, submission_data)
file_path = os.path.join(directories.upload_directory, submission_id)
file_gz_obj = GzipFile(file_path + ".stl.gz", 'wb')
file_gz_obj.write(info['body'])
file_gz_obj.close()
with open('{}.stl.gz'.format(file_path), 'rb') as _in:
DropboxWrapper.add_submission(submission_id, _in.read())
os.remove('{}.stl.gz'.format(file_path)) | [
"ben.gothard3@gmail.com"
] | ben.gothard3@gmail.com |
0ae138ca2c573f41b61fb6fee4e9ecb781b2fc89 | dc62838e9bd214d16dd0b165a219cc658f7b07c3 | /cloud/bin/rst2xetex.py | 081aeaad419428b5578a864e63407d74a1fde9a5 | [] | no_license | zzzzssss/FitKeeper | 976a17b3d6d44c65331d83789848fbabc87d3f85 | 9b16df589151540f94c3f14a98744647a5dcf2de | refs/heads/master | 2022-06-25T15:13:31.229381 | 2017-06-02T16:08:58 | 2017-06-02T16:08:58 | 87,112,882 | 0 | 0 | null | 2017-04-03T19:31:52 | 2017-04-03T19:31:52 | null | UTF-8 | Python | false | false | 925 | py | #!/home/yanan/Desktop/COMS6998CloudComputing/FitKeeper/cloud/bin/python2
# $Id: rst2xetex.py 7847 2015-03-17 17:30:47Z milde $
# Author: Guenter Milde
# Copyright: This module has been placed in the public domain.
"""
A minimal front end to the Docutils Publisher, producing Lua/XeLaTeX code.
"""
try:
import locale
locale.setlocale(locale.LC_ALL, '')
except:
pass
from docutils.core import publish_cmdline
description = ('Generates LaTeX documents from standalone reStructuredText '
'sources for compilation with the Unicode-aware TeX variants '
'XeLaTeX or LuaLaTeX. '
'Reads from <source> (default is stdin) and writes to '
'<destination> (default is stdout). See '
'<http://docutils.sourceforge.net/docs/user/latex.html> for '
'the full reference.')
publish_cmdline(writer_name='xetex', description=description)
| [
"shuyang@ZhaoShuyangsMBP.home"
] | shuyang@ZhaoShuyangsMBP.home |
788b1114cf8da3899edd4800a1fbc676bf8142ee | 1577e1cf4e89584a125cffb855ca50a9654c6d55 | /pyobjc/pyobjc/pyobjc-framework-Quartz-2.5.1/Examples/Programming with Quartz/BasicDrawing/MyAppController.py | 7108ddb749d657bf205c4db6e76aba0164427919 | [
"MIT"
] | permissive | apple-open-source/macos | a4188b5c2ef113d90281d03cd1b14e5ee52ebffb | 2d2b15f13487673de33297e49f00ef94af743a9a | refs/heads/master | 2023-08-01T11:03:26.870408 | 2023-03-27T00:00:00 | 2023-03-27T00:00:00 | 180,595,052 | 124 | 24 | null | 2022-12-27T14:54:09 | 2019-04-10T14:06:23 | null | UTF-8 | Python | false | false | 4,062 | py | from Cocoa import *
import objc
import PDFHandling
import BitmapContext
import Utilities
# Initial defaults
_dpi = 144
_useQT = False
def getURLToExport(suffix):
savePanel = NSSavePanel.savePanel()
initialFileName = "BasicDrawing.%s"%(suffix,)
if savePanel.runModalForDirectory_file_(None, initialFileName) == NSFileHandlingPanelOKButton:
return savePanel.URL()
return None
class MyAppController (NSObject):
theView = objc.IBOutlet()
currentDPIMenuItem = objc.IBOutlet()
currentExportStyleMenuItem = objc.IBOutlet()
@objc.IBAction
def print_(self, sender):
self.theView.print_(sender)
def updateDPIMenu_(self, sender):
if self.currentDPIMenuItem is not sender:
# Uncheck the previous item.
if self.currentDPIMenuItem is not None:
self.currentDPIMenuItem.setState_(NSOffState)
# Update to the current item.
self.currentDPIMenuItem = sender
# Check new menu item.
self.currentDPIMenuItem.setState_(NSOnState)
def updateExportStyleMenu_(self, sender):
if self.currentExportStyleMenuItem is not sender:
# Uncheck the previous item.
if self.currentExportStyleMenuItem is not None:
self.currentExportStyleMenuItem.setState_(NSOffState)
# Update to the current item.
self.currentExportStyleMenuItem = sender
# Check new menu item.
self.currentExportStyleMenuItem.setState_(NSOnState)
@objc.IBAction
def setExportResolution_(self, sender):
global _dpi
_dpi = sender.tag()
self.updateDPIMenu_(sender)
@objc.IBAction
def setUseQT_(self, sender):
global _useQT
_useQT = True
self.updateExportStyleMenu_(sender)
@objc.IBAction
def setUseCGImageSource_(self, sender):
global _useQT
_useQT = False
self.updateExportStyleMenu_(sender)
def setupExportInfo_(self, exportInfoP):
# Use the printable version of the current command. This produces
# the best results for exporting.
exportInfoP.command = self.theView.currentPrintableCommand()
exportInfoP.fileType = ' ' # unused
exportInfoP.useQTForExport = _useQT
exportInfoP.dpi = _dpi
@objc.IBAction
def exportAsPDF_(self, sender):
url = getURLToExport("pdf")
if url is not None:
exportInfo = Utilities.ExportInfo()
self.setupExportInfo_(exportInfo)
PDFHandling.MakePDFDocument(url, exportInfo)
@objc.IBAction
def exportAsPNG_(self, sender):
url = getURLToExport("png")
if url is not None:
exportInfo = Utilities.ExportInfo()
self.setupExportInfo_(exportInfo)
BitmapContext.MakePNGDocument(url, exportInfo)
@objc.IBAction
def exportAsTIFF_(self, sender):
url = getURLToExport("tif")
if url is not None:
exportInfo = Utilities.ExportInfo()
self.setupExportInfo_(exportInfo)
BitmapContext.MakeTIFFDocument(url, exportInfo)
@objc.IBAction
def exportAsJPEG_(self, sender):
url = getURLToExport("jpg")
if url is not None:
exportInfo = Utilities.ExportInfo()
self.setupExportInfo_(exportInfo)
BitmapContext.MakeJPEGDocument(url, exportInfo)
def validateMenuItem_(self, menuItem):
if menuItem.tag == _dpi:
currentDPIMenuItem = menuItem
menuItem.setState_(True)
elif menuItem.action() == 'setUseQT:':
if _useQT:
self.currentDPIMenuItem = menuItem
menuItem.setState_(True)
else:
menuItem.setState_(False)
elif menuItem.action() == 'setUseCGImageSource:':
if _useQT:
currentDPIMenuItem = menuItem
menuItem.setState_(True)
else:
menuItem.setState_(False)
return True
| [
"opensource@apple.com"
] | opensource@apple.com |
41ddd091df6ea055f01a6a9169e98ab77a7ceedd | 2af6a5c2d33e2046a1d25ae9dd66d349d3833940 | /res/scripts/client/gui/app_loader/decorators.py | 49d57664e5a33eceff808d2385710e8a082e19e6 | [] | no_license | webiumsk/WOT-0.9.12-CT | e6c8b5bb106fad71b5c3056ada59fb1aebc5f2b2 | 2506e34bd6634ad500b6501f4ed4f04af3f43fa0 | refs/heads/master | 2021-01-10T01:38:38.080814 | 2015-11-11T00:08:04 | 2015-11-11T00:08:04 | 45,803,240 | 0 | 0 | null | null | null | null | WINDOWS-1250 | Python | false | false | 1,256 | py | # 2015.11.10 21:25:12 Střední Evropa (běžný čas)
# Embedded file name: scripts/client/gui/app_loader/decorators.py
from gui.app_loader.loader import g_appLoader
from gui.app_loader.settings import APP_NAME_SPACE as _SPACE
class app_getter(property):
def __init__(self, fget = None, doc = None, space = None):
super(app_getter, self).__init__(fget=fget, doc=doc)
self._space = space
def __get__(self, obj, objType = None):
return g_appLoader.getApp(self._space)
class def_lobby(property):
def __get__(self, obj, objType = None):
return g_appLoader.getDefLobbyApp()
class def_battle(property):
def __get__(self, obj, objType = None):
return g_appLoader.getDefBattleApp()
class sf_lobby(app_getter):
def __init__(self, fget = None, doc = None):
super(sf_lobby, self).__init__(fget, doc, _SPACE.SF_LOBBY)
class sf_battle(app_getter):
def __init__(self, fget = None, doc = None):
super(sf_battle, self).__init__(fget, doc, _SPACE.SF_BATTLE)
# okay decompyling c:\Users\PC\wotsources\files\originals\res\scripts\client\gui\app_loader\decorators.pyc
# decompiled 1 files: 1 okay, 0 failed, 0 verify failed
# 2015.11.10 21:25:12 Střední Evropa (běžný čas)
| [
"info@webium.sk"
] | info@webium.sk |
f74082804a8763e11216fbb269463a97ccf36bc7 | f904cf1371790b39902cc9a5f24b1903f5828dc7 | /feature5_3_1.py | 3433bf38c56574aef488e727cb2ec87d29acf811 | [] | no_license | Birdie9/Sarcasm-Detection- | af8c4b2791a1e15015fb24649e8c47b809d9d74d | dca5f3f01ca3eddc046440197f37e0c2a707c105 | refs/heads/master | 2020-04-17T13:44:11.523637 | 2016-09-08T03:55:58 | 2016-09-08T03:55:58 | 67,667,097 | 3 | 0 | null | null | null | null | UTF-8 | Python | false | false | 2,332 | py | import time
import datetime
from time import mktime
from datetime import datetime
def getnegscore(tweet, sentidict):
ret=0.0
for word in tweet:
if word in sentidict:
if sentidict[word]<0:
ret+=sentidict[word]
return -ret
def getposscore(tweet,sentidict):
ret=0.0
for word in tweet:
if word in sentidict:
if sentidict[word]>0:
ret+=sentidict[word]
return ret
def mood(tweet,past_data, sentidict,times):
features=[]
pos_score={}
neg_score={}
bucket_size=[1,2,5,10,20,len(past_data)]
buckets=[]
for pastt in past_data:
pos_score[''.join(pastt)]=getposscore(pastt,sentidict)
neg_score[''.join(pastt)]=getnegscore(pastt,sentidict)
for n in bucket_size:
buckets.append(past_data[0:n])
for buck in buckets:
pos_sum=0
neg_sum=0
p=1
q=0
n_pos=0
n_neg=0
n_neut=0
for past_tweet in buck:
pos_sum += pos_score[''.join(past_tweet)]
neg_sum += neg_score[''.join(past_tweet)]
if pos_score[''.join(past_tweet)]>neg_score[''.join(past_tweet)]:
n_pos +=1
elif pos_score[''.join(past_tweet)]<neg_score[''.join(past_tweet)]:
n_neg +=1
else:
n_neut +=1
if pos_sum<neg_sum:
p=-1
maxn = max(n_neg,n_pos,n_neut)
if maxn==n_pos:
q=1
elif maxn==n_neg:
q=-1
else:
q=0
features.extend((pos_sum, neg_sum, p, max(pos_sum,neg_sum), n_pos,n_neg,n_neut,n_neg+n_pos+n_neut,q,maxn))
time_intervals=[1, 2, 5, 10, 20, 60, 720, 1440] #minutes
buckets=[]
for interval in time_intervals:
buck=[]
for i in range(1,len(past_data)):
if (datetime.fromtimestamp(mktime(times[0]))-datetime.fromtimestamp(mktime(times[i]))).seconds<=(interval*60):
buck.append(past_data[i])
buckets.append(buck)
for buck in buckets:
pos_sum=0
neg_sum=0
p=1
q=0
n_pos=0
n_neg=0
n_neut=0
for past_tweet in buck:
pos_sum += pos_score[''.join(past_tweet)]
neg_sum += neg_score[''.join(past_tweet)]
if pos_score[''.join(past_tweet)]>neg_score[''.join(past_tweet)]:
n_pos +=1
elif pos_score[''.join(past_tweet)]<neg_score[''.join(past_tweet)]:
n_neg +=1
else:
n_neut +=1
if pos_sum<neg_sum:
p=-1
maxn = max(n_neg,n_pos,n_neut)
if maxn==n_pos:
q=1
elif maxn==n_neg:
q=-1
else:
q=0
features.extend((pos_sum, neg_sum, p, max(pos_sum,neg_sum), n_pos,n_neg,n_neut,n_neg+n_pos+n_neut,q,maxn))
return features
| [
"jayati.009@gmail.com"
] | jayati.009@gmail.com |
7884029160be242dc95fcc490d6dc0e07a0442ff | 4e2198902a9e07896d4af8817430b2d44a706b25 | /Python Scripts/25.py | 9644a36c79c664519d03663ed85ad52cbfeb8243 | [] | no_license | emilkloeden/Project-Euler-Solutions | e5993472edec1b38fc9b0e8cf83916b0a034af57 | b44fb51d52e9e046c825281a28beacb51331426a | refs/heads/master | 2022-10-01T07:05:08.905374 | 2022-09-12T10:27:31 | 2022-09-12T10:27:31 | 25,626,295 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 735 | py | # A permutation is an ordered arrangement of objects. For example, 3124 is one possible permutation of the digits 1, 2, 3 and 4. If all of the permutations are listed numerically or alphabetically, we call it lexicographic order. The lexicographic permutations of 0, 1 and 2 are:
# 012 021 102 120 201 210
# What is the millionth lexicographic permutation of the digits 0, 1, 2, 3, 4, 5, 6, 7, 8 and 9?
# Answer: 2783915460
# Elapsed time: 0.9259 seconds
from euler import iterative_fibonacci as fib
from timer import timer
@timer
def main():
limit = 1000
i = 1
while len(str(fib(i))) <= limit:
if len(str(fib(i))) >= limit:
return i
i += 1
if __name__ == "__main__":
main() | [
"emilkloeden@gmail.com"
] | emilkloeden@gmail.com |
5b2787c83a0a8eb0caae96635e595e2bc7f9dbed | bc441bb06b8948288f110af63feda4e798f30225 | /database_delivery_sdk/api/sqlpkgs/update_pb2.py | 0f09d354715753abc6b91286cff95f9f6a2d58bf | [
"Apache-2.0"
] | permissive | easyopsapis/easyops-api-python | 23204f8846a332c30f5f3ff627bf220940137b6b | adf6e3bad33fa6266b5fa0a449dd4ac42f8447d0 | refs/heads/master | 2020-06-26T23:38:27.308803 | 2020-06-16T07:25:41 | 2020-06-16T07:25:41 | 199,773,131 | 5 | 0 | null | null | null | null | UTF-8 | Python | false | true | 14,764 | py | # -*- coding: utf-8 -*-
# Generated by the protocol buffer compiler. DO NOT EDIT!
# source: update.proto
import sys
_b=sys.version_info[0]<3 and (lambda x:x) or (lambda x:x.encode('latin1'))
from google.protobuf import descriptor as _descriptor
from google.protobuf import message as _message
from google.protobuf import reflection as _reflection
from google.protobuf import symbol_database as _symbol_database
# @@protoc_insertion_point(imports)
_sym_db = _symbol_database.Default()
from database_delivery_sdk.model.database_delivery import sql_package_version_pb2 as database__delivery__sdk_dot_model_dot_database__delivery_dot_sql__package__version__pb2
from database_delivery_sdk.model.database_delivery import app_pb2 as database__delivery__sdk_dot_model_dot_database__delivery_dot_app__pb2
from database_delivery_sdk.model.database_delivery import dbservice_pb2 as database__delivery__sdk_dot_model_dot_database__delivery_dot_dbservice__pb2
DESCRIPTOR = _descriptor.FileDescriptor(
name='update.proto',
package='sqlpkgs',
syntax='proto3',
serialized_options=None,
serialized_pb=_b('\n\x0cupdate.proto\x12\x07sqlpkgs\x1aGdatabase_delivery_sdk/model/database_delivery/sql_package_version.proto\x1a\x37\x64\x61tabase_delivery_sdk/model/database_delivery/app.proto\x1a=database_delivery_sdk/model/database_delivery/dbservice.proto\"\xbd\x01\n\x17UpdateSQLPackageRequest\x12\r\n\x05pkgId\x18\x01 \x01(\t\x12\x43\n\x0cupdateSqlpkg\x18\x02 \x01(\x0b\x32-.sqlpkgs.UpdateSQLPackageRequest.UpdateSqlpkg\x1aN\n\x0cUpdateSqlpkg\x12\r\n\x05\x61ppId\x18\x01 \x01(\t\x12\x13\n\x0b\x64\x62ServiceId\x18\x02 \x01(\t\x12\x0c\n\x04name\x18\x03 \x01(\t\x12\x0c\n\x04memo\x18\x04 \x01(\t\"\xa1\x02\n\x18UpdateSQLPackageResponse\x12\x39\n\x0bversionList\x18\x01 \x03(\x0b\x32$.database_delivery.SQLPackageVersion\x12+\n\x03\x41PP\x18\x02 \x03(\x0b\x32\x1e.database_delivery.Application\x12/\n\tDBSERVICE\x18\x03 \x03(\x0b\x32\x1c.database_delivery.DBService\x12\n\n\x02id\x18\x04 \x01(\t\x12\x0c\n\x04name\x18\x05 \x01(\t\x12\x0c\n\x04memo\x18\x06 \x01(\t\x12\x0f\n\x07\x63reator\x18\x07 \x01(\t\x12\r\n\x05\x63time\x18\x08 \x01(\x03\x12\r\n\x05mtime\x18\t \x01(\x03\x12\x15\n\rrepoPackageId\x18\n \x01(\t\"\x84\x01\n\x1fUpdateSQLPackageResponseWrapper\x12\x0c\n\x04\x63ode\x18\x01 \x01(\x05\x12\x13\n\x0b\x63odeExplain\x18\x02 \x01(\t\x12\r\n\x05\x65rror\x18\x03 \x01(\t\x12/\n\x04\x64\x61ta\x18\x04 \x01(\x0b\x32!.sqlpkgs.UpdateSQLPackageResponseb\x06proto3')
,
dependencies=[database__delivery__sdk_dot_model_dot_database__delivery_dot_sql__package__version__pb2.DESCRIPTOR,database__delivery__sdk_dot_model_dot_database__delivery_dot_app__pb2.DESCRIPTOR,database__delivery__sdk_dot_model_dot_database__delivery_dot_dbservice__pb2.DESCRIPTOR,])
_UPDATESQLPACKAGEREQUEST_UPDATESQLPKG = _descriptor.Descriptor(
name='UpdateSqlpkg',
full_name='sqlpkgs.UpdateSQLPackageRequest.UpdateSqlpkg',
filename=None,
file=DESCRIPTOR,
containing_type=None,
fields=[
_descriptor.FieldDescriptor(
name='appId', full_name='sqlpkgs.UpdateSQLPackageRequest.UpdateSqlpkg.appId', index=0,
number=1, type=9, cpp_type=9, label=1,
has_default_value=False, default_value=_b("").decode('utf-8'),
message_type=None, enum_type=None, containing_type=None,
is_extension=False, extension_scope=None,
serialized_options=None, file=DESCRIPTOR),
_descriptor.FieldDescriptor(
name='dbServiceId', full_name='sqlpkgs.UpdateSQLPackageRequest.UpdateSqlpkg.dbServiceId', index=1,
number=2, type=9, cpp_type=9, label=1,
has_default_value=False, default_value=_b("").decode('utf-8'),
message_type=None, enum_type=None, containing_type=None,
is_extension=False, extension_scope=None,
serialized_options=None, file=DESCRIPTOR),
_descriptor.FieldDescriptor(
name='name', full_name='sqlpkgs.UpdateSQLPackageRequest.UpdateSqlpkg.name', index=2,
number=3, type=9, cpp_type=9, label=1,
has_default_value=False, default_value=_b("").decode('utf-8'),
message_type=None, enum_type=None, containing_type=None,
is_extension=False, extension_scope=None,
serialized_options=None, file=DESCRIPTOR),
_descriptor.FieldDescriptor(
name='memo', full_name='sqlpkgs.UpdateSQLPackageRequest.UpdateSqlpkg.memo', index=3,
number=4, type=9, cpp_type=9, label=1,
has_default_value=False, default_value=_b("").decode('utf-8'),
message_type=None, enum_type=None, containing_type=None,
is_extension=False, extension_scope=None,
serialized_options=None, file=DESCRIPTOR),
],
extensions=[
],
nested_types=[],
enum_types=[
],
serialized_options=None,
is_extendable=False,
syntax='proto3',
extension_ranges=[],
oneofs=[
],
serialized_start=330,
serialized_end=408,
)
_UPDATESQLPACKAGEREQUEST = _descriptor.Descriptor(
name='UpdateSQLPackageRequest',
full_name='sqlpkgs.UpdateSQLPackageRequest',
filename=None,
file=DESCRIPTOR,
containing_type=None,
fields=[
_descriptor.FieldDescriptor(
name='pkgId', full_name='sqlpkgs.UpdateSQLPackageRequest.pkgId', index=0,
number=1, type=9, cpp_type=9, label=1,
has_default_value=False, default_value=_b("").decode('utf-8'),
message_type=None, enum_type=None, containing_type=None,
is_extension=False, extension_scope=None,
serialized_options=None, file=DESCRIPTOR),
_descriptor.FieldDescriptor(
name='updateSqlpkg', full_name='sqlpkgs.UpdateSQLPackageRequest.updateSqlpkg', index=1,
number=2, type=11, cpp_type=10, label=1,
has_default_value=False, default_value=None,
message_type=None, enum_type=None, containing_type=None,
is_extension=False, extension_scope=None,
serialized_options=None, file=DESCRIPTOR),
],
extensions=[
],
nested_types=[_UPDATESQLPACKAGEREQUEST_UPDATESQLPKG, ],
enum_types=[
],
serialized_options=None,
is_extendable=False,
syntax='proto3',
extension_ranges=[],
oneofs=[
],
serialized_start=219,
serialized_end=408,
)
_UPDATESQLPACKAGERESPONSE = _descriptor.Descriptor(
name='UpdateSQLPackageResponse',
full_name='sqlpkgs.UpdateSQLPackageResponse',
filename=None,
file=DESCRIPTOR,
containing_type=None,
fields=[
_descriptor.FieldDescriptor(
name='versionList', full_name='sqlpkgs.UpdateSQLPackageResponse.versionList', index=0,
number=1, type=11, cpp_type=10, label=3,
has_default_value=False, default_value=[],
message_type=None, enum_type=None, containing_type=None,
is_extension=False, extension_scope=None,
serialized_options=None, file=DESCRIPTOR),
_descriptor.FieldDescriptor(
name='APP', full_name='sqlpkgs.UpdateSQLPackageResponse.APP', index=1,
number=2, type=11, cpp_type=10, label=3,
has_default_value=False, default_value=[],
message_type=None, enum_type=None, containing_type=None,
is_extension=False, extension_scope=None,
serialized_options=None, file=DESCRIPTOR),
_descriptor.FieldDescriptor(
name='DBSERVICE', full_name='sqlpkgs.UpdateSQLPackageResponse.DBSERVICE', index=2,
number=3, type=11, cpp_type=10, label=3,
has_default_value=False, default_value=[],
message_type=None, enum_type=None, containing_type=None,
is_extension=False, extension_scope=None,
serialized_options=None, file=DESCRIPTOR),
_descriptor.FieldDescriptor(
name='id', full_name='sqlpkgs.UpdateSQLPackageResponse.id', index=3,
number=4, type=9, cpp_type=9, label=1,
has_default_value=False, default_value=_b("").decode('utf-8'),
message_type=None, enum_type=None, containing_type=None,
is_extension=False, extension_scope=None,
serialized_options=None, file=DESCRIPTOR),
_descriptor.FieldDescriptor(
name='name', full_name='sqlpkgs.UpdateSQLPackageResponse.name', index=4,
number=5, type=9, cpp_type=9, label=1,
has_default_value=False, default_value=_b("").decode('utf-8'),
message_type=None, enum_type=None, containing_type=None,
is_extension=False, extension_scope=None,
serialized_options=None, file=DESCRIPTOR),
_descriptor.FieldDescriptor(
name='memo', full_name='sqlpkgs.UpdateSQLPackageResponse.memo', index=5,
number=6, type=9, cpp_type=9, label=1,
has_default_value=False, default_value=_b("").decode('utf-8'),
message_type=None, enum_type=None, containing_type=None,
is_extension=False, extension_scope=None,
serialized_options=None, file=DESCRIPTOR),
_descriptor.FieldDescriptor(
name='creator', full_name='sqlpkgs.UpdateSQLPackageResponse.creator', index=6,
number=7, type=9, cpp_type=9, label=1,
has_default_value=False, default_value=_b("").decode('utf-8'),
message_type=None, enum_type=None, containing_type=None,
is_extension=False, extension_scope=None,
serialized_options=None, file=DESCRIPTOR),
_descriptor.FieldDescriptor(
name='ctime', full_name='sqlpkgs.UpdateSQLPackageResponse.ctime', index=7,
number=8, type=3, cpp_type=2, label=1,
has_default_value=False, default_value=0,
message_type=None, enum_type=None, containing_type=None,
is_extension=False, extension_scope=None,
serialized_options=None, file=DESCRIPTOR),
_descriptor.FieldDescriptor(
name='mtime', full_name='sqlpkgs.UpdateSQLPackageResponse.mtime', index=8,
number=9, type=3, cpp_type=2, label=1,
has_default_value=False, default_value=0,
message_type=None, enum_type=None, containing_type=None,
is_extension=False, extension_scope=None,
serialized_options=None, file=DESCRIPTOR),
_descriptor.FieldDescriptor(
name='repoPackageId', full_name='sqlpkgs.UpdateSQLPackageResponse.repoPackageId', index=9,
number=10, type=9, cpp_type=9, label=1,
has_default_value=False, default_value=_b("").decode('utf-8'),
message_type=None, enum_type=None, containing_type=None,
is_extension=False, extension_scope=None,
serialized_options=None, file=DESCRIPTOR),
],
extensions=[
],
nested_types=[],
enum_types=[
],
serialized_options=None,
is_extendable=False,
syntax='proto3',
extension_ranges=[],
oneofs=[
],
serialized_start=411,
serialized_end=700,
)
_UPDATESQLPACKAGERESPONSEWRAPPER = _descriptor.Descriptor(
name='UpdateSQLPackageResponseWrapper',
full_name='sqlpkgs.UpdateSQLPackageResponseWrapper',
filename=None,
file=DESCRIPTOR,
containing_type=None,
fields=[
_descriptor.FieldDescriptor(
name='code', full_name='sqlpkgs.UpdateSQLPackageResponseWrapper.code', index=0,
number=1, type=5, cpp_type=1, label=1,
has_default_value=False, default_value=0,
message_type=None, enum_type=None, containing_type=None,
is_extension=False, extension_scope=None,
serialized_options=None, file=DESCRIPTOR),
_descriptor.FieldDescriptor(
name='codeExplain', full_name='sqlpkgs.UpdateSQLPackageResponseWrapper.codeExplain', index=1,
number=2, type=9, cpp_type=9, label=1,
has_default_value=False, default_value=_b("").decode('utf-8'),
message_type=None, enum_type=None, containing_type=None,
is_extension=False, extension_scope=None,
serialized_options=None, file=DESCRIPTOR),
_descriptor.FieldDescriptor(
name='error', full_name='sqlpkgs.UpdateSQLPackageResponseWrapper.error', index=2,
number=3, type=9, cpp_type=9, label=1,
has_default_value=False, default_value=_b("").decode('utf-8'),
message_type=None, enum_type=None, containing_type=None,
is_extension=False, extension_scope=None,
serialized_options=None, file=DESCRIPTOR),
_descriptor.FieldDescriptor(
name='data', full_name='sqlpkgs.UpdateSQLPackageResponseWrapper.data', index=3,
number=4, type=11, cpp_type=10, label=1,
has_default_value=False, default_value=None,
message_type=None, enum_type=None, containing_type=None,
is_extension=False, extension_scope=None,
serialized_options=None, file=DESCRIPTOR),
],
extensions=[
],
nested_types=[],
enum_types=[
],
serialized_options=None,
is_extendable=False,
syntax='proto3',
extension_ranges=[],
oneofs=[
],
serialized_start=703,
serialized_end=835,
)
_UPDATESQLPACKAGEREQUEST_UPDATESQLPKG.containing_type = _UPDATESQLPACKAGEREQUEST
_UPDATESQLPACKAGEREQUEST.fields_by_name['updateSqlpkg'].message_type = _UPDATESQLPACKAGEREQUEST_UPDATESQLPKG
_UPDATESQLPACKAGERESPONSE.fields_by_name['versionList'].message_type = database__delivery__sdk_dot_model_dot_database__delivery_dot_sql__package__version__pb2._SQLPACKAGEVERSION
_UPDATESQLPACKAGERESPONSE.fields_by_name['APP'].message_type = database__delivery__sdk_dot_model_dot_database__delivery_dot_app__pb2._APPLICATION
_UPDATESQLPACKAGERESPONSE.fields_by_name['DBSERVICE'].message_type = database__delivery__sdk_dot_model_dot_database__delivery_dot_dbservice__pb2._DBSERVICE
_UPDATESQLPACKAGERESPONSEWRAPPER.fields_by_name['data'].message_type = _UPDATESQLPACKAGERESPONSE
DESCRIPTOR.message_types_by_name['UpdateSQLPackageRequest'] = _UPDATESQLPACKAGEREQUEST
DESCRIPTOR.message_types_by_name['UpdateSQLPackageResponse'] = _UPDATESQLPACKAGERESPONSE
DESCRIPTOR.message_types_by_name['UpdateSQLPackageResponseWrapper'] = _UPDATESQLPACKAGERESPONSEWRAPPER
_sym_db.RegisterFileDescriptor(DESCRIPTOR)
UpdateSQLPackageRequest = _reflection.GeneratedProtocolMessageType('UpdateSQLPackageRequest', (_message.Message,), {
'UpdateSqlpkg' : _reflection.GeneratedProtocolMessageType('UpdateSqlpkg', (_message.Message,), {
'DESCRIPTOR' : _UPDATESQLPACKAGEREQUEST_UPDATESQLPKG,
'__module__' : 'update_pb2'
# @@protoc_insertion_point(class_scope:sqlpkgs.UpdateSQLPackageRequest.UpdateSqlpkg)
})
,
'DESCRIPTOR' : _UPDATESQLPACKAGEREQUEST,
'__module__' : 'update_pb2'
# @@protoc_insertion_point(class_scope:sqlpkgs.UpdateSQLPackageRequest)
})
_sym_db.RegisterMessage(UpdateSQLPackageRequest)
_sym_db.RegisterMessage(UpdateSQLPackageRequest.UpdateSqlpkg)
UpdateSQLPackageResponse = _reflection.GeneratedProtocolMessageType('UpdateSQLPackageResponse', (_message.Message,), {
'DESCRIPTOR' : _UPDATESQLPACKAGERESPONSE,
'__module__' : 'update_pb2'
# @@protoc_insertion_point(class_scope:sqlpkgs.UpdateSQLPackageResponse)
})
_sym_db.RegisterMessage(UpdateSQLPackageResponse)
UpdateSQLPackageResponseWrapper = _reflection.GeneratedProtocolMessageType('UpdateSQLPackageResponseWrapper', (_message.Message,), {
'DESCRIPTOR' : _UPDATESQLPACKAGERESPONSEWRAPPER,
'__module__' : 'update_pb2'
# @@protoc_insertion_point(class_scope:sqlpkgs.UpdateSQLPackageResponseWrapper)
})
_sym_db.RegisterMessage(UpdateSQLPackageResponseWrapper)
# @@protoc_insertion_point(module_scope)
| [
"service@easyops.cn"
] | service@easyops.cn |
9ac659ed774916b83e4235fa8eecb1f0508c3ea5 | bad62c2b0dfad33197db55b44efeec0bab405634 | /sdk/storage/azure-mgmt-storage/azure/mgmt/storage/v2021_08_01/aio/_storage_management_client.py | 70166161a2d224894dce9e90a48fb8f889e68c78 | [
"LicenseRef-scancode-generic-cla",
"MIT",
"LGPL-2.1-or-later"
] | permissive | test-repo-billy/azure-sdk-for-python | 20c5a2486456e02456de17515704cb064ff19833 | cece86a8548cb5f575e5419864d631673be0a244 | refs/heads/master | 2022-10-25T02:28:39.022559 | 2022-10-18T06:05:46 | 2022-10-18T06:05:46 | 182,325,031 | 0 | 0 | MIT | 2019-07-25T22:28:52 | 2019-04-19T20:59:15 | Python | UTF-8 | Python | false | false | 10,437 | py | # coding=utf-8
# --------------------------------------------------------------------------
# Copyright (c) Microsoft Corporation. All rights reserved.
# Licensed under the MIT License. See License.txt in the project root for license information.
# Code generated by Microsoft (R) AutoRest Code Generator.
# Changes may cause incorrect behavior and will be lost if the code is regenerated.
# --------------------------------------------------------------------------
from copy import deepcopy
from typing import Any, Awaitable, TYPE_CHECKING
from msrest import Deserializer, Serializer
from azure.core.rest import AsyncHttpResponse, HttpRequest
from azure.mgmt.core import AsyncARMPipelineClient
from .. import models
from ._configuration import StorageManagementClientConfiguration
from .operations import BlobContainersOperations, BlobInventoryPoliciesOperations, BlobServicesOperations, DeletedAccountsOperations, EncryptionScopesOperations, FileServicesOperations, FileSharesOperations, LocalUsersOperations, ManagementPoliciesOperations, ObjectReplicationPoliciesOperations, Operations, PrivateEndpointConnectionsOperations, PrivateLinkResourcesOperations, QueueOperations, QueueServicesOperations, SkusOperations, StorageAccountsOperations, TableOperations, TableServicesOperations, UsagesOperations
if TYPE_CHECKING:
# pylint: disable=unused-import,ungrouped-imports
from azure.core.credentials_async import AsyncTokenCredential
class StorageManagementClient: # pylint: disable=too-many-instance-attributes
"""The Azure Storage Management API.
:ivar operations: Operations operations
:vartype operations: azure.mgmt.storage.v2021_08_01.aio.operations.Operations
:ivar skus: SkusOperations operations
:vartype skus: azure.mgmt.storage.v2021_08_01.aio.operations.SkusOperations
:ivar storage_accounts: StorageAccountsOperations operations
:vartype storage_accounts:
azure.mgmt.storage.v2021_08_01.aio.operations.StorageAccountsOperations
:ivar deleted_accounts: DeletedAccountsOperations operations
:vartype deleted_accounts:
azure.mgmt.storage.v2021_08_01.aio.operations.DeletedAccountsOperations
:ivar usages: UsagesOperations operations
:vartype usages: azure.mgmt.storage.v2021_08_01.aio.operations.UsagesOperations
:ivar management_policies: ManagementPoliciesOperations operations
:vartype management_policies:
azure.mgmt.storage.v2021_08_01.aio.operations.ManagementPoliciesOperations
:ivar blob_inventory_policies: BlobInventoryPoliciesOperations operations
:vartype blob_inventory_policies:
azure.mgmt.storage.v2021_08_01.aio.operations.BlobInventoryPoliciesOperations
:ivar private_endpoint_connections: PrivateEndpointConnectionsOperations operations
:vartype private_endpoint_connections:
azure.mgmt.storage.v2021_08_01.aio.operations.PrivateEndpointConnectionsOperations
:ivar private_link_resources: PrivateLinkResourcesOperations operations
:vartype private_link_resources:
azure.mgmt.storage.v2021_08_01.aio.operations.PrivateLinkResourcesOperations
:ivar object_replication_policies: ObjectReplicationPoliciesOperations operations
:vartype object_replication_policies:
azure.mgmt.storage.v2021_08_01.aio.operations.ObjectReplicationPoliciesOperations
:ivar local_users: LocalUsersOperations operations
:vartype local_users: azure.mgmt.storage.v2021_08_01.aio.operations.LocalUsersOperations
:ivar encryption_scopes: EncryptionScopesOperations operations
:vartype encryption_scopes:
azure.mgmt.storage.v2021_08_01.aio.operations.EncryptionScopesOperations
:ivar blob_services: BlobServicesOperations operations
:vartype blob_services: azure.mgmt.storage.v2021_08_01.aio.operations.BlobServicesOperations
:ivar blob_containers: BlobContainersOperations operations
:vartype blob_containers:
azure.mgmt.storage.v2021_08_01.aio.operations.BlobContainersOperations
:ivar file_services: FileServicesOperations operations
:vartype file_services: azure.mgmt.storage.v2021_08_01.aio.operations.FileServicesOperations
:ivar file_shares: FileSharesOperations operations
:vartype file_shares: azure.mgmt.storage.v2021_08_01.aio.operations.FileSharesOperations
:ivar queue_services: QueueServicesOperations operations
:vartype queue_services: azure.mgmt.storage.v2021_08_01.aio.operations.QueueServicesOperations
:ivar queue: QueueOperations operations
:vartype queue: azure.mgmt.storage.v2021_08_01.aio.operations.QueueOperations
:ivar table_services: TableServicesOperations operations
:vartype table_services: azure.mgmt.storage.v2021_08_01.aio.operations.TableServicesOperations
:ivar table: TableOperations operations
:vartype table: azure.mgmt.storage.v2021_08_01.aio.operations.TableOperations
:param credential: Credential needed for the client to connect to Azure.
:type credential: ~azure.core.credentials_async.AsyncTokenCredential
:param subscription_id: The ID of the target subscription.
:type subscription_id: str
:param base_url: Service URL. Default value is "https://management.azure.com".
:type base_url: str
:keyword api_version: Api Version. Default value is "2021-08-01". Note that overriding this
default value may result in unsupported behavior.
:paramtype api_version: str
:keyword int polling_interval: Default waiting time between two polls for LRO operations if no
Retry-After header is present.
"""
def __init__(
self,
credential: "AsyncTokenCredential",
subscription_id: str,
base_url: str = "https://management.azure.com",
**kwargs: Any
) -> None:
self._config = StorageManagementClientConfiguration(credential=credential, subscription_id=subscription_id, **kwargs)
self._client = AsyncARMPipelineClient(base_url=base_url, config=self._config, **kwargs)
client_models = {k: v for k, v in models.__dict__.items() if isinstance(v, type)}
self._serialize = Serializer(client_models)
self._deserialize = Deserializer(client_models)
self._serialize.client_side_validation = False
self.operations = Operations(
self._client, self._config, self._serialize, self._deserialize
)
self.skus = SkusOperations(
self._client, self._config, self._serialize, self._deserialize
)
self.storage_accounts = StorageAccountsOperations(
self._client, self._config, self._serialize, self._deserialize
)
self.deleted_accounts = DeletedAccountsOperations(
self._client, self._config, self._serialize, self._deserialize
)
self.usages = UsagesOperations(
self._client, self._config, self._serialize, self._deserialize
)
self.management_policies = ManagementPoliciesOperations(
self._client, self._config, self._serialize, self._deserialize
)
self.blob_inventory_policies = BlobInventoryPoliciesOperations(
self._client, self._config, self._serialize, self._deserialize
)
self.private_endpoint_connections = PrivateEndpointConnectionsOperations(
self._client, self._config, self._serialize, self._deserialize
)
self.private_link_resources = PrivateLinkResourcesOperations(
self._client, self._config, self._serialize, self._deserialize
)
self.object_replication_policies = ObjectReplicationPoliciesOperations(
self._client, self._config, self._serialize, self._deserialize
)
self.local_users = LocalUsersOperations(
self._client, self._config, self._serialize, self._deserialize
)
self.encryption_scopes = EncryptionScopesOperations(
self._client, self._config, self._serialize, self._deserialize
)
self.blob_services = BlobServicesOperations(
self._client, self._config, self._serialize, self._deserialize
)
self.blob_containers = BlobContainersOperations(
self._client, self._config, self._serialize, self._deserialize
)
self.file_services = FileServicesOperations(
self._client, self._config, self._serialize, self._deserialize
)
self.file_shares = FileSharesOperations(
self._client, self._config, self._serialize, self._deserialize
)
self.queue_services = QueueServicesOperations(
self._client, self._config, self._serialize, self._deserialize
)
self.queue = QueueOperations(
self._client, self._config, self._serialize, self._deserialize
)
self.table_services = TableServicesOperations(
self._client, self._config, self._serialize, self._deserialize
)
self.table = TableOperations(
self._client, self._config, self._serialize, self._deserialize
)
def _send_request(
self,
request: HttpRequest,
**kwargs: Any
) -> Awaitable[AsyncHttpResponse]:
"""Runs the network request through the client's chained policies.
>>> from azure.core.rest import HttpRequest
>>> request = HttpRequest("GET", "https://www.example.org/")
<HttpRequest [GET], url: 'https://www.example.org/'>
>>> response = await client._send_request(request)
<AsyncHttpResponse: 200 OK>
For more information on this code flow, see https://aka.ms/azsdk/python/protocol/quickstart
:param request: The network request you want to make. Required.
:type request: ~azure.core.rest.HttpRequest
:keyword bool stream: Whether the response payload will be streamed. Defaults to False.
:return: The response of your network call. Does not do error handling on your response.
:rtype: ~azure.core.rest.AsyncHttpResponse
"""
request_copy = deepcopy(request)
request_copy.url = self._client.format_url(request_copy.url)
return self._client.send_request(request_copy, **kwargs)
async def close(self) -> None:
await self._client.close()
async def __aenter__(self) -> "StorageManagementClient":
await self._client.__aenter__()
return self
async def __aexit__(self, *exc_details) -> None:
await self._client.__aexit__(*exc_details)
| [
"noreply@github.com"
] | noreply@github.com |
1b8d32d4bbb24d8091951ee9ed23715c039cb4ae | f48b406244631f84f64e8b61a8892c19010ac554 | /cd4ml/pipeline_params.py | 27c69293c0650845cf322d8d87019437fd815d70 | [
"MIT"
] | permissive | heyestom/CD4ML-Scenarios | c271119fe0dc6a87ece160cae51b4890dcd83a26 | 5fb2edb7098f9c0bf5010199a87dbad39673f431 | refs/heads/master | 2022-11-21T15:29:33.421848 | 2020-07-16T13:27:56 | 2020-07-16T13:27:56 | 276,595,703 | 1 | 0 | MIT | 2020-07-02T08:48:11 | 2020-07-02T08:48:10 | null | UTF-8 | Python | false | false | 728 | py | # parameters for running the pipeline
from cd4ml.ml_model_params import model_parameters
# TODO: add some security protocols around the key?
pipeline_params = {'model_name': 'random_forest',
'days_back': 57,
'acceptance_metric': 'r2_score',
'acceptance_threshold_min': 0.60,
'acceptance_threshold_max': 1.0,
'data_source': 'file',
'model_params': model_parameters,
'download_data_info': {
'key': 'store47-2016.csv',
'gcs_bucket': 'continuous-intelligence',
'base_url': 'https://storage.googleapis.com'
}}
| [
"dajohnst@thoughtworks.com"
] | dajohnst@thoughtworks.com |
c36e62063a94a409390144111aa8b1febb637d79 | 1c594498900dd6f25e0a598b4c89b3e33cec5840 | /iqps/search/views.py | c6c5dfb564a3088854e3a4badd988789e7fb6d3b | [
"MIT"
] | permissive | thealphadollar/iqps | cef42ed8c86e4134e724a5f4967e96a83d672fcd | 187f6b134d82e2dce951b356cb0c7151994ca3ab | refs/heads/master | 2023-07-14T04:41:13.190595 | 2020-06-25T14:51:17 | 2020-06-25T14:51:17 | 277,360,692 | 0 | 0 | MIT | 2020-07-05T18:29:17 | 2020-07-05T18:29:16 | null | UTF-8 | Python | false | false | 3,320 | py | from django.shortcuts import render
from django.db import connection
from django.http import JsonResponse
from iqps.settings import DATABASES
#from .processors import SearchCursor
#Use this with sqlite
#db_name = DATABASES['default']['NAME']
def sqlite_search(subject, year=0, department="", paper_type=""):
year_filter = "AND p.year = {}".format(year) if year > 0 else ""
dep_filter = "AND d.code = '{}'".format(department) if department != "" else ""
type_filter = "AND p.paper_type = '{}'".format(paper_type) if paper_type != "" else ""
if subject == "":
return []
query =\
"""SELECT p.subject, p.year, p.department_id, d.id, d.code, p.paper_type, p.link, SIMILARITYSCORE(p.subject, '{}') AS s
FROM papers p JOIN departments d ON p.department_id = d.id
WHERE s > 70 {} {} {} ORDER BY s DESC;""".format(subject, year_filter, dep_filter, type_filter)
results = []
with SearchCursor(db_name) as c:
c.execute(query)
for row in c.fetchall():
results.append(row)
return results
def _search(subject, year=0, department="", paper_type="", keywords=""):
year_filter = "AND p.year = {}".format(year) if year > 0 else ""
dep_filter = "AND d.code = '{}'".format(department) if department != "" else ""
type_filter = "AND p.paper_type = '{}'".format(paper_type) if paper_type != "" else ""
keyword_filter = "AND kt.text IN {}".format(keywords) if keywords != "" else ""
if subject == "":
return []
if keyword_filter == "":
query =\
"""SELECT p.subject, p.year, d.code, p.paper_type, p.link, p.id
FROM papers p JOIN departments d ON p.department_id = d.id
WHERE SOUNDEX(SUBSTRING(p.subject, 1, LENGTH('{}'))) = SOUNDEX('{}') {} {} {} ORDER BY year DESC LIMIT 30;""".format(subject, subject, year_filter, dep_filter, type_filter)
else:
query =\
"""SELECT p.subject, p.year, d.code, p.paper_type, p.link, p.id, GROUP_CONCAT(kt.text) AS keywords
FROM papers AS p JOIN departments AS d ON p.department_id = d.id
LEFT OUTER JOIN (
SELECT pk.paper_id, k.text FROM papers_keywords AS pk JOIN keywords AS k ON pk.keyword_id = k.id
) AS kt
ON p.id = kt.paper_id
WHERE SOUNDEX(SUBSTRING(p.subject, 1, LENGTH('{}'))) = SOUNDEX('{}')
{} {} {} {}
ORDER BY p.year DESC LIMIT 30;
""".format(subject, subject, year_filter, dep_filter, type_filter, keyword_filter)
results = []
with connection.cursor() as c:
c.execute(query)
for row in c.fetchall():
results.append(row)
return results
def hitSearch(request):
"""
Meant to be an independent API.
Request args:
q -> subject name
year -> year filter
dep -> department filter
typ -> paper_type filter
"""
q = request.GET.get('q', "")
year = request.GET.get('year', 0)
dep = request.GET.get('dep', "")
typ = request.GET.get('typ', "")
keywords = request.GET.get('keys', "")
try:
year = int(year)
except:
year = 0
results = _search(q, year=year, department=dep, paper_type=typ, keywords=keywords)
response = JsonResponse({"papers": results})
response["Access-Control-Allow-Origin"] = "*" #For CORS
return response
| [
"smishra99.iitkgp@gmail.com"
] | smishra99.iitkgp@gmail.com |
e077f429daff201e907044fe1dafc3a66af86952 | 26fc334777ce27d241c67d97adc1761e9d23bdba | /tests/django_tests/tests/middleware_exceptions/tests.py | 0c39f09f9156cf2b9787fa67ac627a5c7dd4a653 | [
"BSD-3-Clause"
] | permissive | alihoseiny/djongo | 1434c9e78c77025d7e0b3330c3a40e9ea0029877 | e2edf099e398573faa90e5b28a32c3d7f1c5f1e9 | refs/heads/master | 2020-03-27T23:27:02.530397 | 2018-08-30T14:44:37 | 2018-08-30T14:44:37 | 147,317,771 | 2 | 1 | BSD-3-Clause | 2018-09-04T09:00:53 | 2018-09-04T09:00:53 | null | UTF-8 | Python | false | false | 6,887 | py | from django.conf import settings
from django.core.exceptions import MiddlewareNotUsed
from django.test import RequestFactory, SimpleTestCase, override_settings
from django.test.utils import patch_logger
from . import middleware as mw
@override_settings(ROOT_URLCONF='middleware_exceptions.urls')
class MiddlewareTests(SimpleTestCase):
def tearDown(self):
mw.log = []
@override_settings(MIDDLEWARE=['middleware_exceptions.middleware.ProcessViewNoneMiddleware'])
def test_process_view_return_none(self):
response = self.client.get('/middleware_exceptions/view/')
self.assertEqual(mw.log, ['processed view normal_view'])
self.assertEqual(response.content, b'OK')
@override_settings(MIDDLEWARE=['middleware_exceptions.middleware.ProcessViewMiddleware'])
def test_process_view_return_response(self):
response = self.client.get('/middleware_exceptions/view/')
self.assertEqual(response.content, b'Processed view normal_view')
@override_settings(MIDDLEWARE=[
'middleware_exceptions.middleware.ProcessViewTemplateResponseMiddleware',
'middleware_exceptions.middleware.LogMiddleware',
])
def test_templateresponse_from_process_view_rendered(self):
"""
TemplateResponses returned from process_view() must be rendered before
being passed to any middleware that tries to access response.content,
such as middleware_exceptions.middleware.LogMiddleware.
"""
response = self.client.get('/middleware_exceptions/view/')
self.assertEqual(response.content, b'Processed view normal_view\nProcessViewTemplateResponseMiddleware')
@override_settings(MIDDLEWARE=[
'middleware_exceptions.middleware.ProcessViewTemplateResponseMiddleware',
'middleware_exceptions.middleware.TemplateResponseMiddleware',
])
def test_templateresponse_from_process_view_passed_to_process_template_response(self):
"""
TemplateResponses returned from process_view() should be passed to any
template response middleware.
"""
response = self.client.get('/middleware_exceptions/view/')
expected_lines = [
b'Processed view normal_view',
b'ProcessViewTemplateResponseMiddleware',
b'TemplateResponseMiddleware',
]
self.assertEqual(response.content, b'\n'.join(expected_lines))
@override_settings(MIDDLEWARE=['middleware_exceptions.middleware.TemplateResponseMiddleware'])
def test_process_template_response(self):
response = self.client.get('/middleware_exceptions/template_response/')
self.assertEqual(response.content, b'template_response OK\nTemplateResponseMiddleware')
@override_settings(MIDDLEWARE=['middleware_exceptions.middleware.LogMiddleware'])
def test_view_exception_converted_before_middleware(self):
response = self.client.get('/middleware_exceptions/permission_denied/')
self.assertEqual(mw.log, [(response.status_code, response.content)])
self.assertEqual(response.status_code, 403)
@override_settings(MIDDLEWARE=['middleware_exceptions.middleware.ProcessExceptionMiddleware'])
def test_view_exception_handled_by_process_exception(self):
response = self.client.get('/middleware_exceptions/error/')
self.assertEqual(response.content, b'Exception caught')
@override_settings(MIDDLEWARE=[
'middleware_exceptions.middleware.ProcessExceptionLogMiddleware',
'middleware_exceptions.middleware.ProcessExceptionMiddleware',
])
def test_response_from_process_exception_short_circuits_remainder(self):
response = self.client.get('/middleware_exceptions/error/')
self.assertEqual(mw.log, [])
self.assertEqual(response.content, b'Exception caught')
@override_settings(MIDDLEWARE=[
'middleware_exceptions.middleware.LogMiddleware',
'middleware_exceptions.middleware.NotFoundMiddleware',
])
def test_exception_in_middleware_converted_before_prior_middleware(self):
response = self.client.get('/middleware_exceptions/view/')
self.assertEqual(mw.log, [(404, response.content)])
self.assertEqual(response.status_code, 404)
@override_settings(MIDDLEWARE=['middleware_exceptions.middleware.ProcessExceptionMiddleware'])
def test_exception_in_render_passed_to_process_exception(self):
response = self.client.get('/middleware_exceptions/exception_in_render/')
self.assertEqual(response.content, b'Exception caught')
@override_settings(ROOT_URLCONF='middleware_exceptions.urls')
class RootUrlconfTests(SimpleTestCase):
@override_settings(ROOT_URLCONF=None)
def test_missing_root_urlconf(self):
# Removing ROOT_URLCONF is safe, as override_settings will restore
# the previously defined settings.
del settings.ROOT_URLCONF
with self.assertRaises(AttributeError):
self.client.get("/middleware_exceptions/view/")
class MyMiddleware:
def __init__(self, get_response=None):
raise MiddlewareNotUsed
def process_request(self, request):
pass
class MyMiddlewareWithExceptionMessage:
def __init__(self, get_response=None):
raise MiddlewareNotUsed('spam eggs')
def process_request(self, request):
pass
@override_settings(
DEBUG=True,
ROOT_URLCONF='middleware_exceptions.urls',
MIDDLEWARE=['django.middleware.common.CommonMiddleware'],
)
class MiddlewareNotUsedTests(SimpleTestCase):
rf = RequestFactory()
def test_raise_exception(self):
request = self.rf.get('middleware_exceptions/view/')
with self.assertRaises(MiddlewareNotUsed):
MyMiddleware().process_request(request)
@override_settings(MIDDLEWARE=['middleware_exceptions.tests.MyMiddleware'])
def test_log(self):
with patch_logger('django.request', 'debug') as calls:
self.client.get('/middleware_exceptions/view/')
self.assertEqual(len(calls), 1)
self.assertEqual(
calls[0],
"MiddlewareNotUsed: 'middleware_exceptions.tests.MyMiddleware'"
)
@override_settings(MIDDLEWARE=['middleware_exceptions.tests.MyMiddlewareWithExceptionMessage'])
def test_log_custom_message(self):
with patch_logger('django.request', 'debug') as calls:
self.client.get('/middleware_exceptions/view/')
self.assertEqual(len(calls), 1)
self.assertEqual(
calls[0],
"MiddlewareNotUsed('middleware_exceptions.tests.MyMiddlewareWithExceptionMessage'): spam eggs"
)
@override_settings(DEBUG=False)
def test_do_not_log_when_debug_is_false(self):
with patch_logger('django.request', 'debug') as calls:
self.client.get('/middleware_exceptions/view/')
self.assertEqual(len(calls), 0)
| [
"nesdis@gmail.com"
] | nesdis@gmail.com |
92c94e59ce4623a858dc7e53049807390bbac34f | 35093a037837e5155464782ab0300d64cf1ce31a | /Automata.py | 659c3e1e2e738c60ae8d0780bffd739fb501d189 | [] | no_license | ChenSaijilahu/Simulation-traffic | eeb0a38169a60ffcc50e859be554a165a3a31c97 | a97a8f63d1382ffb59874cb209bdf29ea9625c57 | refs/heads/master | 2023-03-15T00:04:51.159032 | 2021-01-05T11:30:04 | 2021-01-05T11:30:04 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 6,655 | py | import random
import numpy as np
import matplotlib.pyplot as plt
class Cellular_Automata(object):
# Initialize the constructor
# Read from the keyboard the input values: the number of cells for the road, car density and the number of iterations
# The road will be modified just below, in the constructor. The declaration has been made here for easier code handling
def __init__(self, no_cells, density, no_iter):
# The road is initialized as a matrix of two numpy arrays. This is to outline the road initially and after each car has gone a certain step distance
self.road = np.zeros( (2, no_cells) )
# Another array has been defined for the state of the cell. In this task, it s taken into account whether the car has previously moved or not.
# Initially, all the states of the cells within the state array are 1
self.state = np.zeros( (1, no_cells) )
# Some random cells are selected, standing for the cars (the filled cells)
# Work out the number of cars as function of the given traffic density and the desired number of cells
no_cars = round( density * no_cells )
self.car = no_cars
for i in range (no_cars):
# If the selected cell is already filled, try another value, until an empty one has been found
r = random.randint(0, no_cells - 1)
while ( self.road[0][r] == 1):
r = random.randint(0, no_cells - 1)
# Once an empty cell has been found, it is filled
self.road[0][r] = 1
# State array with cells 0 involves empty cells, 1 are used for stationary cars
# and -1 for cars which previously moved. Initially, all cars that are created have state 1
self.state[0][r] = 1
print("The number of the cars is: " + str(no_cars))
def modify_step(self, no_cells, p_move, p_stay):
# Modify the next step for the traffic using the rule specified below
# It is taken into account whether the car has previously moved or not
moving_cars = 0
for i in range (no_cells):
# If the cell has state 1
if(self.road[0][i] == 1):
# Analyse the case where the next cell is empty
# When the last cell is reached, the next value will be the first cell. Hence, the modulo operation.
if(self.road[0][ (i+1) % no_cells ] == 0):
# Analyse if the car has previously moved. This is why two arrays have been created. self.road array retains only the curent state of the cell.
# On the other hand, self.state array also takes into account the previous state of the cell. If the car has previously moved and
# is now located at position i, then self.state[0][i] will have value -1. If the car has not moved and is at position i, then self.state[0][i]
# will have value 0. The two cases are analysed below
if(self.state[0][i] == -1):
r = random.random()
if(r <= p_move):
# If the car moves, update the values of the road and state numpy arrays. The car leaves the current cell at position i and moves to position (i+1)
# or 0 if the car is located at the last cell
self.road[1][i] = 0
self.road[1][(i+1) % no_cells] = 1
self.state[0][i] = 0
self.state[0][ (i+1) % no_cells ] = -1
# Update the number of cars which have moved
moving_cars += 1
if(self.state[0][i] == 1):
# In this case, the car has not previously moved
r = random.random()
if(r <= p_stay):
self.road[1][i] = 0
self.road[1][(i+1) % no_cells] = 1
self.state[0][i] = 0
self.state[0][ (i+1) % no_cells ] = -1
moving_cars += 1
else:
self.road[1][i] = 1
self.state[0][i] = 1
# If the cell has state 0
if(self.road[0][i] == 0):
# The same cases are analysed here, ony from the motion of the cars behind cell located at position i
if(self.road[0][ (i-1) % no_cells ] == 1):
if(self.state[0][i] == -1):
r = random.random()
if(r <= p_move):
self.road[1][i] = 1
self.road[1][(i-1) % no_cells] = 0
self.state[0][i] = -1
self.state[0][ (i-1) % no_cells ] = 0
if(self.state[0][i] == 1):
r = random.random()
if(r <= p_stay):
self.road[1][i] = 1
self.road[1][(i-1) % no_cells] = 0
self.state[0][i] = -1
self.state[0][ (i-1) % no_cells ] = 0
else:
self.road[1][i] = 0
self.state[0][i] = 0
# The first row of the road array represents the initial state of the traffic and the second row represents the final state of the traffic.
# Update the the initial state to be the same as the final state, in order to repeat the process
self.road[0] = self.road[1]
return moving_cars
def plot_normal_traffic(self, no_cells, no_iter, p_move, p_stay):
# Modify the traffic for the specified number of iterations and calculate the average speed of cars for each timestep
avg_speed = []
timestep = []
traffic = np.zeros( (no_iter,no_cells) )
for i in range (no_iter):
timestep.append(i)
moving_cars = self.modify_step(no_cells, p_move, p_stay)
traffic[i] = self.road[0]
avg = moving_cars / self.car
avg_speed.append(avg)
return traffic, timestep, avg_speed
def output(self):
# Print out the final grid by using string
s = ""
for i in range (len(self.road)):
for j in range (len(self.road[i])):
s += str(self.road[i][j]) + " "
s += "\n"
print(s)
| [
"noreply@github.com"
] | noreply@github.com |
0917883314d3c71d9c5f509602e435b21d824cce | 6d70cd7f3e15375ae4c8fa8619a5a29f98b629ee | /Redes3/Parcial2/RRDTools/TrendLineal/TrendGraph.py | 21a7f4e0c3a45f73c7e27fccc3dba559cfb722fb | [] | no_license | AlinaMarianaPerezSoberanes/Archivos | 5155aa47243d0010395046f1cb37f20af2e99dd8 | 7628ae1e6644215b3c8148f590923502714fc10e | refs/heads/master | 2020-04-20T08:35:55.008022 | 2019-02-02T16:59:38 | 2019-02-02T16:59:38 | 168,744,044 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 1,357 | py | import sys
import rrdtool
import time
while 1:
ret = rrdtool.graph( "trend.png",
"--start",'1539382500',
"--end","1539384500",
"--vertical-label=Carga CPU",
"--title=Uso de CPU",
"--color", "ARROW#009900",
'--vertical-label', "Uso de CPU (%)",
'--lower-limit', '0',
'--upper-limit', '100',
"DEF:carga=trend.rrd:CPUload:AVERAGE",
"AREA:carga#00FF00:CPU load",
"LINE1:30",
"AREA:5#ff000022:stack",
"VDEF:CPUlast=carga,LAST",
"VDEF:CPUmin=carga,MINIMUM",
"VDEF:CPUavg=carga,AVERAGE",
"VDEF:CPUmax=carga,MAXIMUM",
"COMMENT: Now Min Avg Max//n",
"GPRINT:CPUlast:%12.0lf%s",
"GPRINT:CPUmin:%10.0lf%s",
"GPRINT:CPUavg:%13.0lf%s",
"GPRINT:CPUmax:%13.0lf%s",
"VDEF:a=carga,LSLSLOPE",
"VDEF:b=carga,LSLINT",
'CDEF:avg2=carga,POP,a,COUNT,*,b,+',
"LINE2:avg2#FFBB00" )
time.sleep(15)
| [
"psoberanes_mariana@hotmail.com"
] | psoberanes_mariana@hotmail.com |
ec45f92fa7feccad2e8ebffe1dfe13c66eefb4c8 | b4f80293a9230925429bccf9ca80df404f830466 | /char-rnns/microgradchar.py | 90cbf45553a52b3692d0fbe815caa415f50e10ab | [] | no_license | jcanode/nlp_fun | cac737c43559aaa7bca55984b849adfbea4e1314 | f81a75fe6f325eeaf531d3c6cfa64ef12ada4fc2 | refs/heads/master | 2023-01-11T15:34:04.888871 | 2020-11-11T04:56:52 | 2020-11-11T04:56:52 | 303,938,286 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 463 | py | from micrograd.engine import Value
a = Value(-4.0)
b = Value(2.0)
c = a + b
d = a * b + b**3
c += c + 1
c += 1 + c + (-a)
d += d * 2 + (b + a).relu()
d += 3 * d + (b - a).relu()
e = c - d
f = e**2
g = f / 2.0
g += 10.0 / f
print(f'{g.data:.4f}') # prints 24.7041, the outcome of this forward pass
g.backward()
print(f'{a.grad:.4f}') # prints 138.8338, i.e. the numerical value of dg/da
print(f'{b.grad:.4f}') # prints 645.5773, i.e. the numerical value of dg/dbc | [
"45806280+jcanode@users.noreply.github.com"
] | 45806280+jcanode@users.noreply.github.com |
522c84f8db34b50d58ffb6b486b23961e9cc2994 | 5556424ae28e1965ccf0bb60b8203d76e812c3e3 | /app/settings_prod.py | 3db6997577ab1a3ec887588bf20072da76a50cb7 | [] | no_license | astromitts/borkle | e0c36ebb0c6575d123680342efcf20383d592102 | ff5f67a1fb81bc0dbf4afb5e8fda3cbe86ec9b63 | refs/heads/main | 2023-01-01T02:48:54.321538 | 2020-10-25T13:43:45 | 2020-10-25T13:43:45 | 307,006,472 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 139 | py | import os
from app.settings import * # noqa
DEBUG = True
STATICFILES_STORAGE = 'whitenoise.storage.CompressedManifestStaticFilesStorage'
| [
"bomorin-id@MacBook-Pro.local"
] | bomorin-id@MacBook-Pro.local |
9e99850e135e9250610331794779c4d4a0c8881e | 449c29b00f44e441f285638eea46fe957a042424 | /MainApp/mixins.py | d01891068c8582665e8aca033474b0576eb0d1dc | [] | no_license | InnocenceNerevarine/Diploma | d4260cf55041583a6e199f75884bd46e3e5ffd3c | 7a2ea9976ec7e0730c7731a4a0613cd7c58013af | refs/heads/master | 2023-04-26T13:00:53.634118 | 2021-05-20T06:52:07 | 2021-05-20T06:52:07 | 368,245,756 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 984 | py | from django.views.generic import View
from .models import Cart, Customer
class CartMixin(View):
def dispatch(self, request, *args, **kwargs):
if not request.session.session_key:
request.session.save()
self.session = request.session
if request.user.is_authenticated:
customer = Customer.objects.filter(user=request.user).first()
if not customer:
customer = Customer.objects.create(user=request.user)
cart = Cart.objects.filter(owner=customer, in_order=False).first()
if not cart:
cart = Cart.objects.create(owner=customer)
else:
cart = Cart.objects.filter(session_key=self.session.session_key, for_anonymous_user=True).first()
if not cart:
cart = Cart.objects.create(session_key=self.session.session_key, for_anonymous_user=True)
self.cart = cart
return super().dispatch(request, *args, **kwargs)
| [
"ilya.sidorov.2014@gmail.com"
] | ilya.sidorov.2014@gmail.com |
1265e8a3612b796c07f5ab24327fecb0bbb2d1b8 | b9a607c121c8e36c3c1dec47003ec2fbfc660ff1 | /sendmail/views.py | 6a5b9c7c278a35a31939020d3becfa191cb0c296 | [] | no_license | AustralianSynchrotron/send-mail-server | fe8bc8a6e5c076d9980b8b6df25f50a10f875927 | ea3c3eff1e229bfeb35a008b6a4c116a86113af0 | refs/heads/master | 2021-01-19T11:45:06.921363 | 2017-07-15T14:00:24 | 2017-07-15T14:00:24 | 82,261,123 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 1,969 | py | from sendmail import app,mail
from flask_mail import Message
from flask import request
import logging
logger = logging.getLogger(__name__)
logger.setLevel(logging.DEBUG)
st = logging.StreamHandler()
st.setLevel(logging.DEBUG)
formatter = logging.Formatter("%(asctime)s [%(name)s] %(levelname)s :: %(message)s")
st.setFormatter(formatter)
logger.addHandler(st)
from jinja2 import Template
@app.route('/')
def index():
return Template("<link rel='shortcut icon' href='/static/img/favicon.png' >" +
"<pre>curl --data 'subject=<subject>&body=<body>" +
"&recipients=<recipient[@synchrotron.org.au]>[,one][,two][,etc...]' " +
"hostAddress:Port/sendmail/</pre>").render()
@app.route('/sendmail/', methods=['POST'])
def post_the_mail():
# required to send message
# subject<string>, body<string>, recipients<list>
subject_string = str(request.form.get('subject'))
body_string = str(request.form.get('body'))
tmp_list = str(request.form.get('recipients')).split(',')
from_string = str(request.form.get('from'))
logger.info(from_string)
if not from_string:
from_string = 'email_robot'
recipient_list = []
for recipient in tmp_list:
r = recipient
if '@' not in recipient:
r = r + '@synchrotron.org.au'
recipient_list.append(r)
logger.info("%s, %s, %s, %s" % (from_string, subject_string, body_string, tmp_list))
with mail.connect() as con:
logger.info('Sending Email Message to the following users: ')
for user in recipient_list:
logger.info(user)
msg = Message(subject=subject_string,
sender=from_string + '@synchrotron.org.au',
html=body_string,
recipients=[user])
con.send(msg)
logger.info("Email Sent, you don't have anymore messages!")
return "Done! "
| [
"cameron.rodda@synchrotron.org.au"
] | cameron.rodda@synchrotron.org.au |
183aba71c348509e64b9916b1b2e0d84103f3be3 | 322726c2e15389d0381692676156a5e87c8ad0f9 | /site/bin/painter.py | 342e538912085fb724da01907eca35899f87f5b4 | [] | no_license | Trostnick/PSDB | e2cddd3f5c4518098e38935d441e0112b178aa14 | fd4c4a209d60c1c029916b0328ac9d7372daed83 | refs/heads/master | 2020-05-27T13:01:55.726212 | 2018-03-20T14:32:12 | 2018-03-20T14:32:12 | 124,147,394 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 2,119 | py | #!/home/meeg/site/bin/python2
#
# The Python Imaging Library
# $Id$
#
# this demo script illustrates pasting into an already displayed
# photoimage. note that the current version of Tk updates the whole
# image every time we paste, so to get decent performance, we split
# the image into a set of tiles.
#
try:
from tkinter import Tk, Canvas, NW
except ImportError:
from Tkinter import Tk, Canvas, NW
from PIL import Image, ImageTk
import sys
#
# painter widget
class PaintCanvas(Canvas):
def __init__(self, master, image):
Canvas.__init__(self, master, width=image.size[0], height=image.size[1])
# fill the canvas
self.tile = {}
self.tilesize = tilesize = 32
xsize, ysize = image.size
for x in range(0, xsize, tilesize):
for y in range(0, ysize, tilesize):
box = x, y, min(xsize, x+tilesize), min(ysize, y+tilesize)
tile = ImageTk.PhotoImage(image.crop(box))
self.create_image(x, y, image=tile, anchor=NW)
self.tile[(x, y)] = box, tile
self.image = image
self.bind("<B1-Motion>", self.paint)
def paint(self, event):
xy = event.x - 10, event.y - 10, event.x + 10, event.y + 10
im = self.image.crop(xy)
# process the image in some fashion
im = im.convert("L")
self.image.paste(im, xy)
self.repair(xy)
def repair(self, box):
# update canvas
dx = box[0] % self.tilesize
dy = box[1] % self.tilesize
for x in range(box[0]-dx, box[2]+1, self.tilesize):
for y in range(box[1]-dy, box[3]+1, self.tilesize):
try:
xy, tile = self.tile[(x, y)]
tile.paste(self.image.crop(xy))
except KeyError:
pass # outside the image
self.update_idletasks()
#
# main
if len(sys.argv) != 2:
print("Usage: painter file")
sys.exit(1)
root = Tk()
im = Image.open(sys.argv[1])
if im.mode != "RGB":
im = im.convert("RGB")
PaintCanvas(root, im).pack()
root.mainloop()
| [
"trostnick97@mail.ru"
] | trostnick97@mail.ru |
fbe152c2c7005c6aebce424912f910e0be53e37a | 65ce70d806e379f75683244722b2975e6f511bbb | /another program/Map() in Python.py | 32bbde9e8dc5fdc8bc20989ea35087a257880b03 | [] | no_license | Mahedi2150/python | dfc2ad6c01d16cf5a6da85864d5fb9c478e645ef | 37fbd2b1fe0b606cc91a5bc8a47a0fc9f2ab0eec | refs/heads/master | 2022-12-30T17:20:23.705238 | 2020-10-20T20:04:55 | 2020-10-20T20:04:55 | 293,569,171 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 267 | py | """def mulFiveTimes(number):
return number*5
result = []
num = [3,5,7,9,1,5]
for i in num:
result.append(mulFiveTimes(i))
print(result)"""
def mulFiveTimes(number):
return number*5
num = [3,5,7,9,1,5]
print(list(map(mulFiveTimes,num))) | [
"noreply@github.com"
] | noreply@github.com |
46b2a84ea85072fd8b8c7365f2bcc70d57327e12 | 8509927b6281647a5400f92a2199cb82860f2997 | /code/grid_search/run_grid_search.py | b60615d1712fba4315ad782c693ded31690fffc0 | [] | no_license | egeodaci/comp551-2020-p2_classification_of_textual_data | 02dbd55dd61a098081aed27202ce5d653ca46dee | 13b6e169b5b965b7185de49294c33c35c7da9b65 | refs/heads/master | 2022-11-21T08:15:45.700213 | 2020-07-09T17:38:22 | 2020-07-09T17:38:22 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 23,838 | py | import json
import logging
import os
from time import time
from sklearn import metrics
from sklearn.ensemble import AdaBoostClassifier
from sklearn.ensemble import GradientBoostingClassifier
from sklearn.ensemble import RandomForestClassifier
from sklearn.feature_extraction.text import TfidfVectorizer
from sklearn.linear_model import LogisticRegression
from sklearn.linear_model import PassiveAggressiveClassifier
from sklearn.linear_model import Perceptron
from sklearn.linear_model import RidgeClassifier
from sklearn.model_selection import GridSearchCV
from sklearn.model_selection import cross_validate
from sklearn.naive_bayes import BernoulliNB
from sklearn.naive_bayes import ComplementNB
from sklearn.naive_bayes import MultinomialNB
from sklearn.neighbors import KNeighborsClassifier
from sklearn.neighbors import NearestCentroid
from sklearn.pipeline import Pipeline
from sklearn.svm import LinearSVC
from sklearn.tree import DecisionTreeClassifier
from datasets.load_dataset import load_twenty_news_groups, load_imdb_reviews
from model_selection.ml_algorithm_pair_list import JSON_FOLDER
from utils.dataset_enum import Dataset
from utils.ml_classifiers_enum import Classifier
def get_classifier_with_best_parameters(classifier_enum, best_parameters):
if classifier_enum == Classifier.ADA_BOOST_CLASSIFIER:
return AdaBoostClassifier(**best_parameters)
elif classifier_enum == Classifier.BERNOULLI_NB:
return BernoulliNB(**best_parameters)
elif classifier_enum == Classifier.COMPLEMENT_NB:
return ComplementNB(**best_parameters)
elif classifier_enum == Classifier.DECISION_TREE_CLASSIFIER:
return DecisionTreeClassifier(**best_parameters)
elif classifier_enum == Classifier.GRADIENT_BOOSTING_CLASSIFIER:
return GradientBoostingClassifier(**best_parameters)
elif classifier_enum == Classifier.K_NEIGHBORS_CLASSIFIER:
return KNeighborsClassifier(**best_parameters)
elif classifier_enum == Classifier.LINEAR_SVC:
return LinearSVC(**best_parameters)
elif classifier_enum == Classifier.LOGISTIC_REGRESSION:
return LogisticRegression(**best_parameters)
elif classifier_enum == Classifier.MULTINOMIAL_NB:
return MultinomialNB(**best_parameters)
elif classifier_enum == Classifier.NEAREST_CENTROID:
return NearestCentroid(**best_parameters)
elif classifier_enum == Classifier.PASSIVE_AGGRESSIVE_CLASSIFIER:
return PassiveAggressiveClassifier(**best_parameters)
elif classifier_enum == Classifier.PERCEPTRON:
return Perceptron(**best_parameters)
elif classifier_enum == Classifier.RANDOM_FOREST_CLASSIFIER:
return RandomForestClassifier(**best_parameters)
elif classifier_enum == Classifier.RIDGE_CLASSIFIER:
return RidgeClassifier(**best_parameters)
def run_classifier_grid_search(classifer, classifier_enum, param_grid, dataset, final_classification_table_default_parameters, final_classification_table_best_parameters, imdb_multi_class, save_json_with_best_parameters):
if param_grid is None:
return
if dataset == Dataset.TWENTY_NEWS_GROUPS:
remove = ('headers', 'footers', 'quotes')
data_train = \
load_twenty_news_groups(subset='train', categories=None, shuffle=True, random_state=0, remove=remove)
data_test = \
load_twenty_news_groups(subset='test', categories=None, shuffle=True, random_state=0, remove=remove)
X_train, y_train = data_train.data, data_train.target
X_test, y_test = data_test.data, data_test.target
target_names = data_train.target_names
elif dataset == Dataset.IMDB_REVIEWS:
db_parent_path = os.getcwd()
db_parent_path = db_parent_path.replace('grid_search', '')
if imdb_multi_class:
X_train, y_train = \
load_imdb_reviews(subset='train', multi_class_labels=True, verbose=False, shuffle=True, random_state=0,
db_parent_path=db_parent_path)
X_test, y_test = \
load_imdb_reviews(subset='test', multi_class_labels=True, verbose=False, shuffle=True, random_state=0,
db_parent_path=db_parent_path)
else:
X_train, y_train = \
load_imdb_reviews(subset='train', multi_class_labels=False, verbose=False, shuffle=True, random_state=0,
db_parent_path=db_parent_path)
X_test, y_test = \
load_imdb_reviews(subset='test', multi_class_labels=False, verbose=False, shuffle=True, random_state=0,
db_parent_path=db_parent_path)
if imdb_multi_class:
# IMDB_REVIEWS dataset
# If binary classification: 0 = neg and 1 = pos.
# If multi-class classification use the review scores: 1, 2, 3, 4, 7, 8, 9, 10
target_names = ['1', '2', '3', '4', '7', '8', '9', '10']
else:
# IMDB_REVIEWS dataset
# If binary classification: 0 = neg and 1 = pos.
# If multi-class classification use the review scores: 1, 2, 3, 4, 7, 8, 9, 10
target_names = ['0', '1']
try:
# Extracting features
vectorizer = TfidfVectorizer(stop_words='english', strip_accents='unicode', analyzer='word', binary=True)
X_train = vectorizer.fit_transform(X_train)
X_test = vectorizer.transform(X_test)
# Create pipeline
pipeline = Pipeline([('classifier', classifer)])
# Create grid search object
grid_search = GridSearchCV(pipeline, param_grid=param_grid, cv=5, verbose=True, n_jobs=-1)
logging.info("\n\nPerforming grid search...\n")
logging.info("Parameters:")
logging.info(param_grid)
t0 = time()
grid_search.fit(X_train, y_train)
logging.info("\tDone in %0.3fs" % (time() - t0))
# Get best parameters
logging.info("\tBest score: %0.3f" % grid_search.best_score_)
logging.info("\tBest parameters set:")
best_parameters = grid_search.best_estimator_.get_params()
new_parameters = {}
for param_name in sorted(param_grid.keys()):
logging.info("\t\t%s: %r" % (param_name, best_parameters[param_name]))
key = param_name.replace('classifier__', '')
value = best_parameters[param_name]
new_parameters[key] = value
if save_json_with_best_parameters:
if dataset == Dataset.TWENTY_NEWS_GROUPS:
json_path = os.path.join(os.getcwd(), JSON_FOLDER, dataset.name, classifier_enum.name + ".json")
with open(json_path, 'w') as outfile:
json.dump(new_parameters, outfile)
else:
if imdb_multi_class:
json_path = os.path.join(os.getcwd(), JSON_FOLDER, dataset.name, 'multi_class_classification', classifier_enum.name + ".json")
with open(json_path, 'w') as outfile:
json.dump(new_parameters, outfile)
else:
json_path = os.path.join(os.getcwd(), JSON_FOLDER, dataset.name, 'binary_classification', classifier_enum.name + ".json")
with open(json_path, 'w') as outfile:
json.dump(new_parameters, outfile)
logging.info('\n\nUSING {} WITH DEFAULT PARAMETERS'.format(classifier_enum.name))
clf = classifer
final_classification_report(clf, X_train, y_train, X_test, y_test, target_names, classifier_enum, final_classification_table_default_parameters)
logging.info('\n\nUSING {} WITH BEST PARAMETERS: {}'.format(classifier_enum.name, new_parameters))
clf = get_classifier_with_best_parameters(classifier_enum, new_parameters)
final_classification_report(clf, X_train, y_train, X_test, y_test, target_names, classifier_enum, final_classification_table_best_parameters)
except MemoryError as error:
# Output expected MemoryErrors.
logging.error(error)
except Exception as exception:
# Output unexpected Exceptions.
logging.error(exception)
def final_classification_report(clf, X_train, y_train, X_test, y_test, target_names, classifier_enum, final_classification_table):
# Fit on data
logging.info('_' * 80)
logging.info("Training: ")
logging.info(clf)
t0 = time()
clf.fit(X_train, y_train)
train_time = time() - t0
logging.info("Train time: %0.3fs" % train_time)
# Predict
t0 = time()
y_pred = clf.predict(X_test)
test_time = time() - t0
logging.info("Test time: %0.3fs" % test_time)
accuracy_score = metrics.accuracy_score(y_test, y_pred)
logging.info("Accuracy score: %0.3f" % accuracy_score)
logging.info("\n\n===> Classification Report:\n")
# logging.info(metrics.classification_report(y_test, y_pred, target_names=target_names))
logging.info(metrics.classification_report(y_test, y_pred))
n_splits = 5
logging.info("\n\nCross validation:")
scoring = ['accuracy', 'precision_macro', 'precision_micro', 'precision_weighted', 'recall_macro', 'recall_micro',
'recall_weighted', 'f1_macro', 'f1_micro', 'f1_weighted', 'jaccard_macro']
cross_val_scores = cross_validate(clf, X_train, y_train, scoring=scoring, cv=n_splits,
n_jobs=-1, verbose=True)
cv_test_accuracy = cross_val_scores['test_accuracy']
logging.info("\taccuracy: {}-fold cross validation: {}".format(5, cv_test_accuracy))
cv_accuracy_score_mean_std = "%0.2f (+/- %0.2f)" % (cv_test_accuracy.mean() * 100, cv_test_accuracy.std() * 2 * 100)
logging.info("\ttest accuracy: {}-fold cross validation accuracy: {}".format(n_splits, cv_accuracy_score_mean_std))
final_classification_table[classifier_enum.value] = classifier_enum.name, format(accuracy_score, ".2%"), str(cv_test_accuracy), cv_accuracy_score_mean_std, format(train_time, ".4"), format(test_time, ".4")
def get_classifier_with_default_parameters(classifier_enum):
if classifier_enum == Classifier.ADA_BOOST_CLASSIFIER:
'''
AdaBoostClassifier(algorithm='SAMME.R', base_estimator=None, learning_rate=1.0,
n_estimators=50, random_state=None)
'''
clf = AdaBoostClassifier()
parameters = {
'classifier__learning_rate': [0.1, 1],
'classifier__n_estimators': [200, 500]
}
elif classifier_enum == Classifier.BERNOULLI_NB:
'''
BernoulliNB(alpha=1.0, binarize=0.0, class_prior=None, fit_prior=True)
'''
clf = BernoulliNB()
parameters = {
'classifier__alpha': [0.0001, 0.001, 0.01, 0.1, 0.5, 1.0, 2.0, 4.0, 6.0, 8.0, 10.0],
'classifier__binarize': [0.0001, 0.001, 0.01, 0.1, 0.5, 1.0, 2.0, 4.0, 6.0, 8.0, 10.0],
'classifier__fit_prior': [False, True]
}
elif classifier_enum == Classifier.COMPLEMENT_NB:
'''
ComplementNB(alpha=1.0, class_prior=None, fit_prior=True, norm=False)
'''
clf = ComplementNB()
parameters = {
'classifier__alpha': [0.0001, 0.001, 0.01, 0.1, 0.5, 1.0, 2.0, 4.0, 6.0, 8.0, 10.0],
'classifier__fit_prior': [False, True],
'classifier__norm': [False, True]
}
elif classifier_enum == Classifier.DECISION_TREE_CLASSIFIER:
'''
DecisionTreeClassifier(ccp_alpha=0.0, class_weight=None, criterion='gini',
max_depth=None, max_features=None, max_leaf_nodes=None,
min_impurity_decrease=0.0, min_impurity_split=None,
min_samples_leaf=1, min_samples_split=2,
min_weight_fraction_leaf=0.0, presort='deprecated',
random_state=None, splitter='best')
'''
clf = DecisionTreeClassifier()
parameters = {
'classifier__criterion': ["entropy", "gini"],
'classifier__splitter': ["best", "random"],
'classifier__min_samples_split': [2, 100, 250]
}
elif classifier_enum == Classifier.K_NEIGHBORS_CLASSIFIER:
'''
KNeighborsClassifier(algorithm='auto', leaf_size=30, metric='minkowski',
metric_params=None, n_jobs=None, n_neighbors=5, p=2,
weights='uniform')
'''
clf = KNeighborsClassifier()
parameters = {
'classifier__leaf_size': [5, 30],
'classifier__metric': ['euclidean', 'minkowski'],
'classifier__n_neighbors': [3, 50],
'classifier__weights': ['uniform', 'distance']
}
elif classifier_enum == Classifier.LINEAR_SVC:
'''
LinearSVC(C=1.0, class_weight=None, dual=True, fit_intercept=True,
intercept_scaling=1, loss='squared_hinge', max_iter=1000,
multi_class='ovr', penalty='l2', random_state=None, tol=0.0001,
verbose=0)
'''
clf = LinearSVC()
parameters = {
'classifier__C': [0.01, 1.0],
'classifier__multi_class': ['ovr', 'crammer_singer'],
'classifier__tol': [0.0001, 0.001]
}
elif classifier_enum == Classifier.LOGISTIC_REGRESSION:
'''
LogisticRegression(C=1.0, class_weight=None, dual=False, fit_intercept=True,
intercept_scaling=1, l1_ratio=None, max_iter=100,
multi_class='auto', n_jobs=None, penalty='l2',
random_state=None, solver='lbfgs', tol=0.0001, verbose=0,
warm_start=False)
'''
clf = LogisticRegression()
parameters = {
'classifier__C': [1, 10],
'classifier__tol': [0.001, 0.01]
}
elif classifier_enum == Classifier.MULTINOMIAL_NB:
'''
MultinomialNB(alpha=1.0, class_prior=None, fit_prior=True)
'''
clf = MultinomialNB()
parameters = {
'classifier__alpha': [0.0001, 0.001, 0.01, 0.1, 0.5, 1.0, 2.0, 4.0, 6.0, 8.0, 10.0],
'classifier__fit_prior': [False, True]
}
elif classifier_enum == Classifier.NEAREST_CENTROID:
'''
NearestCentroid(metric='euclidean', shrink_threshold=None)
'''
clf = NearestCentroid()
parameters = {
'classifier__metric': ['euclidean', 'cosine']
}
elif classifier_enum == Classifier.PASSIVE_AGGRESSIVE_CLASSIFIER:
'''
PassiveAggressiveClassifier(C=1.0, average=False, class_weight=None,
early_stopping=False, fit_intercept=True,
loss='hinge', max_iter=1000, n_iter_no_change=5,
n_jobs=None, random_state=None, shuffle=True,
tol=0.001, validation_fraction=0.1, verbose=0,
warm_start=False)
'''
clf = PassiveAggressiveClassifier()
parameters = {
'classifier__C': [0.01, 1.0],
'classifier__early_stopping': [False, True],
'classifier__tol': [0.0001, 0.001, 0.01],
'classifier__validation_fraction': [0.0001, 0.01]
}
elif classifier_enum == Classifier.PERCEPTRON:
'''
Perceptron(alpha=0.0001, class_weight=None, early_stopping=False, eta0=1.0,
fit_intercept=True, max_iter=1000, n_iter_no_change=5, n_jobs=None,
penalty=None, random_state=0, shuffle=True, tol=0.001,
validation_fraction=0.1, verbose=0, warm_start=False)
'''
clf = Perceptron()
parameters = {
'classifier__early_stopping': [True],
'classifier__max_iter': [100],
'classifier__n_iter_no_change': [3, 15],
'classifier__penalty': ['l2'],
'classifier__tol': [0.0001, 0.1],
'classifier__validation_fraction': [0.0001, 0.01]
}
elif classifier_enum == Classifier.RANDOM_FOREST_CLASSIFIER:
'''
RandomForestClassifier(bootstrap=True, ccp_alpha=0.0, class_weight=None,
criterion='gini', max_depth=None, max_features='auto',
max_leaf_nodes=None, max_samples=None,
min_impurity_decrease=0.0, min_impurity_split=None,
min_samples_leaf=1, min_samples_split=2,
min_weight_fraction_leaf=0.0, n_estimators=100,
n_jobs=None, oob_score=False, random_state=None,
verbose=0, warm_start=False)
'''
clf = RandomForestClassifier()
parameters = {
'classifier__min_samples_leaf': [1, 2],
'classifier__min_samples_split': [2, 5],
'classifier__n_estimators': [100, 200]
}
elif classifier_enum == Classifier.RIDGE_CLASSIFIER:
'''
RidgeClassifier(alpha=1.0, class_weight=None, copy_X=True, fit_intercept=True,
max_iter=None, normalize=False, random_state=None,
solver='auto', tol=0.001)
'''
clf = RidgeClassifier()
parameters = {
'classifier__alpha': [0.5, 1.0],
'classifier__tol': [0.0001, 0.001]
}
elif classifier_enum == Classifier.GRADIENT_BOOSTING_CLASSIFIER:
'''
GradientBoostingClassifier(ccp_alpha=0.0, criterion='friedman_mse', init=None,
learning_rate=0.1, loss='deviance', max_depth=3,
max_features=None, max_leaf_nodes=None,
min_impurity_decrease=0.0, min_impurity_split=None,
min_samples_leaf=1, min_samples_split=2,
min_weight_fraction_leaf=0.0, n_estimators=100,
n_iter_no_change=None, presort='deprecated',
random_state=None, subsample=1.0, tol=0.0001,
validation_fraction=0.1, verbose=0,
warm_start=False)
'''
clf = GradientBoostingClassifier()
parameters = {
'classifier__learning_rate': [0.01, 0.1],
'classifier__n_estimators': [100, 200]
}
return clf, parameters
def run_grid_search(save_logs_in_file, just_imdb_dataset, imdb_multi_class, save_json_with_best_parameters=False):
if imdb_multi_class:
if save_logs_in_file:
if not os.path.exists('grid_search/just_imdb_using_multi_class_classification'):
os.mkdir('grid_search/just_imdb_using_multi_class_classification')
logging.basicConfig(filename='grid_search/just_imdb_using_multi_class_classification/all.log', format='%(asctime)s - %(levelname)s - %(message)s',
level=logging.INFO, datefmt='%m/%d/%Y %I:%M:%S %p')
else:
logging.basicConfig(format='%(asctime)s - %(levelname)s - %(message)s', level=logging.INFO,
datefmt='%m/%d/%Y %I:%M:%S %p')
else:
if save_logs_in_file:
if not os.path.exists('grid_search/20newsgroups_and_imdb_using_binary_classification'):
os.mkdir('grid_search/20newsgroups_and_imdb_using_binary_classification')
logging.basicConfig(filename='grid_search/20newsgroups_and_imdb_using_binary_classification/all.log', format='%(asctime)s - %(levelname)s - %(message)s',
level=logging.INFO, datefmt='%m/%d/%Y %I:%M:%S %p')
else:
logging.basicConfig(format='%(asctime)s - %(levelname)s - %(message)s', level=logging.INFO,
datefmt='%m/%d/%Y %I:%M:%S %p')
classifier_list = [
Classifier.ADA_BOOST_CLASSIFIER,
Classifier.DECISION_TREE_CLASSIFIER,
Classifier.LINEAR_SVC,
Classifier.LOGISTIC_REGRESSION,
Classifier.RANDOM_FOREST_CLASSIFIER,
Classifier.BERNOULLI_NB,
Classifier.COMPLEMENT_NB,
Classifier.MULTINOMIAL_NB,
Classifier.NEAREST_CENTROID,
Classifier.PASSIVE_AGGRESSIVE_CLASSIFIER,
Classifier.K_NEIGHBORS_CLASSIFIER,
Classifier.PERCEPTRON,
Classifier.RIDGE_CLASSIFIER,
Classifier.GRADIENT_BOOSTING_CLASSIFIER
]
if just_imdb_dataset:
dataset_list = [
Dataset.IMDB_REVIEWS
]
else:
dataset_list = [
Dataset.IMDB_REVIEWS,
Dataset.TWENTY_NEWS_GROUPS
]
logging.info("\n>>> GRID SEARCH")
for dataset in dataset_list:
c_count = 1
final_classification_table_default_parameters = {}
final_classification_table_best_parameters = {}
for classifier_enum in classifier_list:
logging.info("\n")
logging.info("#" * 80)
if save_logs_in_file:
print("#" * 80)
logging.info("{})".format(c_count))
clf, parameters = get_classifier_with_default_parameters(classifier_enum)
logging.info("*" * 80)
logging.info("Classifier: {}, Dataset: {}".format(classifier_enum.name, dataset.name))
logging.info("*" * 80)
start = time()
run_classifier_grid_search(clf, classifier_enum, parameters, dataset, final_classification_table_default_parameters, final_classification_table_best_parameters, imdb_multi_class, save_json_with_best_parameters)
end = time() - start
logging.info("It took {} seconds".format(end))
logging.info("*" * 80)
if save_logs_in_file:
print("*" * 80)
print("Classifier: {}, Dataset: {}".format(classifier_enum.name, dataset.name))
print(clf)
print("It took {} seconds".format(end))
print("*" * 80)
logging.info("#" * 80)
if save_logs_in_file:
print("#" * 80)
c_count = c_count + 1
logging.info(
'\n\nCURRENT CLASSIFICATION TABLE: {} DATASET, CLASSIFIER WITH DEFAULT PARAMETERS'.format(dataset.name))
print_final_classification_table(final_classification_table_default_parameters)
logging.info(
'\n\nCURRENT CLASSIFICATION TABLE: {} DATASET, CLASSIFIER WITH BEST PARAMETERS'.format(dataset.name))
print_final_classification_table(final_classification_table_best_parameters)
logging.info('\n\nFINAL CLASSIFICATION TABLE: {} DATASET, CLASSIFIER WITH DEFAULT PARAMETERS'.format(dataset.name))
print_final_classification_table(final_classification_table_default_parameters)
logging.info('\n\nFINAL CLASSIFICATION TABLE: {} DATASET, CLASSIFIER WITH BEST PARAMETERS'.format(dataset.name))
print_final_classification_table(final_classification_table_best_parameters)
def print_final_classification_table(final_classification_table_default_parameters):
logging.info(
'| ID | ML Algorithm | Accuracy Score (%) | K-fold Cross Validation (CV) (k = 5) | CV (Mean +/- Std) | '
'Training time (seconds) | Test time (seconds) |')
logging.info(
'| -- | ------------ | ------------------ | ------------------------------------ | ----------------- | '
' ------------------ | ------------------ |')
for key in sorted(final_classification_table_default_parameters.keys()):
values = final_classification_table_default_parameters[key]
logging.info(
"| {} | {} | {} | {} | {} | {} | {} |".format(key, values[0], values[1], values[2],
values[3], values[4], values[5]))
| [
"ramon.fgrd@gmail.com"
] | ramon.fgrd@gmail.com |
cdd2eecd01704a67982d8dda03dcfaa9b67cffff | 3f9c9a80240b5d059dd48d30859c498c417b3db5 | /visit_column.py | 2f6b2dc3c0dfcd234f404aa02ca94803df8a2f28 | [] | no_license | haiqiang2017/csdn_pageviewers- | 45f858815e712b4240826eb59b6f9d892d520031 | fd5c32a051a49a811b89323fe83f36b628686a33 | refs/heads/master | 2020-05-15T21:50:24.153753 | 2019-05-02T15:40:43 | 2019-05-02T15:40:43 | 182,510,406 | 2 | 0 | null | null | null | null | UTF-8 | Python | false | false | 3,235 | py | #coding:utf-8
import random
import urllib3
import time
from cookie_pool import get_cookie
from UserAgent_pool import get_UserAgent
import requests
"""提取数据访问链接"""
# 禁用urllib3的警告
urllib3.disable_warnings()
class visitSpider(object):
def __init__(self):
self.ua = get_UserAgent().get_random_useragent()
self.cookie = get_cookie().get_random_cookie()
# print self.ua,self.cookie
# 处理 headers
self.headers = {"User-Agent": self.ua, "cookie": self.cookie}
self.urls = []
self.nums = 0
self.currentHour = time.localtime().tm_hour
def readFile(self):
""""""
self.nums += 0
with open("download/column.txt", "r") as f:
for i in f.readlines():
# url = i[:-2]
url = i
if self.checkClone(url,self.urls):
# print("\r读取url:",i)
self.urls.append(i[:-1])
self.nums += 1
def checkClone(self,urls,list):
"""判重"""
flag = 0
if len(list) > 0:
for i in list:
if list == urls:
flag = 1
if flag == 1:
return 0
return 1
def visit(self):
request = urllib3.PoolManager()
# 计算时间
timeNum = 0
# 计算次数
listNum = 0
while True:
# 访问时间 8 12 18 20
#if self.currentHour == 0 or self.currentHour == 1 or self.currentHour == 20 or self.currentHour == 1:
try:
if self.currentHour:
# 得到url值
url = self.urls[random.randint(0,self.nums - 1)]
# 使用urllib3发送请求
# response = request.request('GET', url, headers=self.headers)
response = requests.get(url,headers= self.headers)
# print response.content
# break
# 打印返回信息
print(url,response.status_code,listNum,str(time.localtime().tm_hour) + ":" + str(time.localtime().tm_min))
# 访问一次睡一秒
time.sleep(random.choice(range(8,12)))
timeNum += 1
listNum += 1
# 每访问50次睡30秒
if listNum % 50 == 0:
for i in range(random.choice(range(25,30))):
print("\r剩余休息时间:%d秒"%(30-i))
time.sleep(1)
# 当到达一定时间(1个小时)之后重新读取文档
if timeNum%3600==0:
self.readFile()
else:
print("休息中,当前时间:" + str(time.localtime().tm_hour) + ":" + str(time.localtime().tm_min)
+ ":" + str(time.localtime().tm_sec) + " ...")
time.sleep(1)
except Exception as e:
print str(e)
time.sleep(3600)
def Main():
vi = visitSpider()
vi.readFile()
vi.visit()
if __name__ == "__main__":
Main()
| [
"Venus_haiqiang@163.com"
] | Venus_haiqiang@163.com |
5d74c49cfb918d4a55a35493f41a60b12a5f1a0f | 667c2d8d8a37a3a7719d5aa44586f59f8ce8fd51 | /bakery/__init__.py | 952a95fe79f173ed2e4c6d969e5f5477dc00636a | [
"MIT"
] | permissive | iredelmeier/doughknots | fe2880dff5b32a81e572b088cbe97b1501b2f38a | 487431b189eed8e33d369403100ff3b68d7a4151 | refs/heads/master | 2022-12-23T22:00:33.953861 | 2019-07-17T22:48:59 | 2019-07-17T22:48:59 | 197,473,677 | 3 | 0 | MIT | 2022-12-08T05:54:28 | 2019-07-17T22:49:51 | Python | UTF-8 | Python | false | false | 195 | py | from .bakery import Bakery, NoopBakery
from .httpclient import HttpClient
from .kind import Kind
from .service import Service
__all__ = ["Bakery", "HttpClient", "Kind", "NoopBakery", "Service"]
| [
"iredelmeier@gmail.com"
] | iredelmeier@gmail.com |
383e4c356d475a877c9169bab7cfba4b57eb90d8 | b9e0e10e9014f80ede6ea7c29174697367cf9acb | /src/evaluation_utils.py | 6ad5766e619bc2ac050261f9ead7b8438681df4f | [
"MIT"
] | permissive | talshapira/SASA | 2e9b7353147b5cba8e48b65ebb12b9bd13346a53 | 70db6ba36d7602e46fbb95b6a3cac822c8af2ab9 | refs/heads/main | 2023-08-17T08:52:33.398005 | 2021-10-04T15:36:44 | 2021-10-04T15:36:44 | 407,496,107 | 4 | 0 | null | null | null | null | UTF-8 | Python | false | false | 6,195 | py | from sklearn.metrics import confusion_matrix
import itertools
import matplotlib.pyplot as plt
import numpy as np
from sklearn.metrics import accuracy_score, recall_score, roc_auc_score, roc_curve, auc, precision_score
def get_tpr_specific_fpr(fpr, tpr, s_fpr=0.01):
for i, fp in enumerate(fpr):
if fp > s_fpr:
return fpr[i-1], tpr[i-1]
def print_evaluation_metrics(y_test, y_test_prediction, y_test_prob, model_name):
print("accuracy_score", "for", model_name, accuracy_score(y_test, y_test_prediction))
print("FA", "for", model_name, 1 - recall_score(y_test, y_test_prediction, pos_label=0))
print("Detection rate i.e. recall_score", "for", model_name, recall_score(y_test, y_test_prediction))
print("AUC", "for", model_name, roc_auc_score(y_test, y_test_prob))
fpr, tpr, thresholds = roc_curve(y_test, y_test_prob)
print("TPR@FPR=0.001", "for", model_name, get_tpr_specific_fpr(fpr, tpr, s_fpr=0.001))
print("TPR@FPR=0.01", "for", model_name, get_tpr_specific_fpr(fpr, tpr, s_fpr=0.01))
print("TPR@FPR=0.1", "for", model_name, get_tpr_specific_fpr(fpr, tpr, s_fpr=0.1))
def plot_roc_curve(y_test, y_test_prob, path_prefix, model_name='', max_fp=0.1):
fpr, tpr, thresholds = roc_curve(y_test, y_test_prob)
plt.figure()
lw = 2
plt.plot(fpr, tpr, lw=lw, label=model_name + ' (AUC = %0.3f)' % auc(fpr, tpr))
plt.xlim([0.0, max_fp])
plt.ylim([0.0, 1.05])
plt.xlabel('False Positive Rate')
plt.ylabel('True Positive Rate')
# plt.title('ROC Curves')
plt.legend(loc="lower right")
plt.savefig(path_prefix + "_ROC Curve", bbox_inches='tight')
plt.show()
def plot_roc_curve_multiple(y_test, y_test_prob_list, path_prefix, model_names, max_fp=0.1):
plt.figure()
lw = 2
for i, y_test_prob in enumerate(y_test_prob_list):
fpr, tpr, thresholds = roc_curve(y_test, y_test_prob)
plt.plot(fpr, tpr, lw=lw, label=model_names[i] + ' (AUC = %0.3f)' % auc(fpr, tpr))
plt.xlim([0.0, max_fp])
plt.ylim([0.0, 1.05])
plt.xlabel('False Positive Rate')
plt.ylabel('True Positive Rate')
# plt.title('ROC Curves')
plt.legend(loc="lower right")
plt.savefig(path_prefix + "_ROC Curve", bbox_inches='tight')
plt.show()
def smooth(y, box_pts):
box = np.ones(box_pts)/box_pts
y_smooth = np.convolve(y, box, mode='same')
return y_smooth
def plot_history_accuracy(history, epochs, path_prefix, sm=False, metrics=['acc','val_acc']):
x = np.asarray(range(1, epochs + 1))
# summarize history for accuracy
plt.figure()
if sm:
plt.plot(x, smooth([y*100 for y in history[metrics[0]]],2))
plt.plot(x, smooth([y*100 for y in history[metrics[1]]],2))
else:
plt.plot(x, [y*100 for y in history[metrics[0]]])
plt.plot(x, [y*100 for y in history[metrics[1]]])
plt.ylabel('Accuracy (%)')
plt.xlabel('Epochs')
# plt.ylim(70,100) ###########################
plt.legend(['Training', 'Test'], loc='lower right') #loc='lower right')
plt.grid()
fname = path_prefix + "_accuracy_history"
plt.savefig(fname, bbox_inches='tight')
plt.show()
def plot_history_loss(history, epochs, path_prefix, sm=False, metrics=['loss','val_loss']):
x = np.asarray(range(1, epochs + 1))
# summarize history for accuracy
plt.figure()
if sm:
plt.plot(x, smooth([y*100 for y in history[metrics[0]]],2))
plt.plot(x, smooth([y*100 for y in history[metrics[1]]],2))
else:
plt.plot(x, [y*100 for y in history[metrics[0]]])
plt.plot(x, [y*100 for y in history[metrics[1]]])
plt.ylabel('Loss')
plt.xlabel('Epochs')
# plt.ylim(70,100) ###########################
plt.legend(['Training', 'Test'], loc='upper right') #loc='lower right')
plt.grid()
fname = path_prefix + "_loss_history"
plt.savefig(fname, bbox_inches='tight')
plt.show()
def plot_confusion_matrix(cm, classes,
normalize=False,
fname='Confusion matrix', title=None,
cmap=plt.cm.Blues):
"""
This function prints and plots the confusion matrix.
Normalization can be applied by setting `normalize=True`.
"""
if normalize:
cm = cm.astype('float') / cm.sum(axis=1)[:, np.newaxis]
print("Normalized confusion matrix")
else:
print('Confusion matrix, without normalization')
plt.imshow(cm, interpolation='nearest', cmap=cmap)
if title is not None:
plt.title(title)
cbar = plt.colorbar()
tick_marks = np.arange(len(classes))
plt.xticks(tick_marks, classes, rotation=45)
plt.ylim(-0.5, 1.5)
plt.yticks(tick_marks, classes)
fmt = '.1f' if normalize else 'd'
thresh = cm.max() / 2.
for i, j in itertools.product(range(cm.shape[0]), range(cm.shape[1])):
if normalize:
plt.text(j, i, format(cm[i, j] * 100, fmt) + '%',
horizontalalignment="center",
color="white" if cm[i, j] > thresh else "black")
cbar.set_ticks([0, .2, .4, 0.6, 0.8, 1])
cbar.set_ticklabels(['0%', '20%', '40%', '60%', '80%', '100%'])
else:
plt.text(j, i, format(cm[i, j], fmt),
horizontalalignment="center",
color="white" if cm[i, j] > thresh else "black")
plt.tight_layout()
plt.ylabel('True Label')
plt.xlabel('Predicted Label')
plt.savefig(fname + ".png", bbox_inches='tight')
def compute_confusion_matrix(y_test, y_test_prediction, class_names, path_prefix):
# Compute confusion matrix
cnf_matrix = confusion_matrix(y_test, y_test_prediction)
np.set_printoptions(precision=2)
# Plot non-normalized confusion matrix
plt.figure()
plot_confusion_matrix(cnf_matrix, classes=class_names,
fname=path_prefix + "_" + 'Confusion_matrix_without_normalization')
# Plot normalized confusion matrix
plt.figure()
plot_confusion_matrix(cnf_matrix, classes=class_names, normalize=True,
fname=path_prefix + "_" + 'Normalized_confusion_matrix')
plt.show() | [
"noreply@github.com"
] | noreply@github.com |
fc1a6a670a2055dd23d840641834d8d1a011428d | 18462298cd5636399735339266ece565e7fbd494 | /daily_weather/setup.py | e684da34c60bb9bee45f2fddfa44d55802629abc | [] | no_license | AnkitP7/flask-demo | 4f8897d7563392a398dc10207af900d4afe45115 | 1ce10542669a044cece68c5cf4fb1caaa99003aa | refs/heads/master | 2020-03-29T12:19:44.807918 | 2018-09-22T16:28:29 | 2018-09-22T16:28:29 | 149,893,875 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 192 | py | from setuptools import setup
setup(
name='daily_weather',
packages=['daily_weather'],
include_package_data=True,
install_requires=[
'flask',
'pylint'
],
)
| [
"ankit.patel39@gmail.com"
] | ankit.patel39@gmail.com |
429d42c8fd21f8aeed2ca8697dc6fab586d5a1dd | 1fec393454ffe7f65fce3617c14a2fcedf1da663 | /Searching/Searching I/matrix_median.py | 9cab3f6da7a1a3e9e867bcedf81f9997880f980b | [] | no_license | VarmaSANJAY/InterviewBit-Solution-Python | fbeb1d855a5244a89b40fbd2522640dc596c79b6 | ea26394cc1b9d22a9ab474467621d2b61ef15a31 | refs/heads/master | 2022-11-27T22:46:34.966395 | 2020-08-09T14:10:58 | 2020-08-09T14:10:58 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 864 | py | from bisect import *
class Solution:
def binary_search(self,A, min_el, max_el, cnt_before_mid):
s = min_el
e = max_el
while s < e:
mid = (s+e) // 2
count = 0
for row in A:
count += bisect_right(row, mid)
if count > cnt_before_mid:
e = mid
else:
s = mid + 1
return s
def Solve(self,A):
min_el = float('inf')
max_el = float('-inf')
for i in A:
min_el = min(i[0], min_el)
max_el = max(i[-1], max_el)
m=len(A)
n=len(A[0])
cnt_before_mid = (m*n) // 2
return self.binary_search(A, min_el, max_el,cnt_before_mid)
if __name__ == '__main__':
A = [[1, 3, 5],
[2, 6, 9],
[3, 6, 9]]
B = Solution()
print(B.Solve(A))
| [
"srajsonu02@gmail.com"
] | srajsonu02@gmail.com |
0093acd5c0ab3527f6e25307f4a2f09b05a3eb73 | 27f1be7865eb58d17e5478299b5685fc625a055c | /src/dataset/dataset_mnistm.py | 99eaa5390d2b6840b00a2fc0b30135514eaefe79 | [
"MIT"
] | permissive | mkirchmeyer/adaptation-imputation | e53099f654bf75526e11ed93e9b78e4a26ed0bef | 7ef683f2da08699b3f877467fdb0e00d3b02bccc | refs/heads/main | 2023-08-28T05:31:30.924573 | 2021-10-13T15:03:59 | 2021-10-13T15:09:42 | 351,026,014 | 2 | 1 | null | null | null | null | UTF-8 | Python | false | false | 7,718 | py | """
Dataset setting and data loader for MNIST-M.
Modified from
https://github.com/pytorch/vision/blob/master/torchvision/datasets/mnist.py
CREDIT: https://github.com/corenel
"""
from __future__ import print_function
import errno
import os
import torch
import torch.utils.data as data
from PIL import Image
from src.dataset.sampler import BalancedBatchSampler
class MNISTM(data.Dataset):
"""`MNIST-M Dataset."""
url = "https://github.com/VanushVaswani/keras_mnistm/releases/download/1.0/keras_mnistm.pkl.gz"
raw_folder = 'raw'
processed_folder = 'processed'
training_file = 'mnist_m_train.pt'
test_file = 'mnist_m_test.pt'
def __init__(self,
root, mnist_root="data",
train=True,
transform=None, target_transform=None,
download=False):
"""Init MNIST-M dataset."""
super(MNISTM, self).__init__()
self.root = os.path.expanduser(root)
self.mnist_root = os.path.expanduser(mnist_root)
self.transform = transform
self.target_transform = target_transform
self.train = train # training set or test set
if download:
self.download()
if not self._check_exists():
raise RuntimeError('Dataset not found.' +
' You can use download=True to download it')
if self.train:
self.train_data, self.train_labels = \
torch.load(os.path.join(self.root,
self.processed_folder,
self.training_file))
else:
self.test_data, self.test_labels = \
torch.load(os.path.join(self.root,
self.processed_folder,
self.test_file))
def __getitem__(self, index):
"""Get images and target for data loader.
Args:
index (int): Index
Returns:
tuple: (image, target) where target is index of the target class.
"""
if self.train:
img, target = self.train_data[index], self.train_labels[index]
else:
img, target = self.test_data[index], self.test_labels[index]
# doing this so that it is consistent with all other datasets
# to return a PIL Image
img = Image.fromarray(img.squeeze().numpy(), mode='RGB')
if self.transform is not None:
img = self.transform(img)
if self.target_transform is not None:
target = self.target_transform(target)
return img, target
def __len__(self):
"""Return size of dataset."""
if self.train:
return len(self.train_data)
else:
return len(self.test_data)
def _check_exists(self):
return os.path.exists(os.path.join(self.root,
self.processed_folder,
self.training_file)) and \
os.path.exists(os.path.join(self.root,
self.processed_folder,
self.test_file))
def download(self):
"""Download the MNIST data."""
# import essential packages
from six.moves import urllib
import gzip
import pickle
from torchvision import datasets
# check if dataset already exists
if self._check_exists():
return
# make data dirs
try:
os.makedirs(os.path.join(self.root, self.raw_folder))
os.makedirs(os.path.join(self.root, self.processed_folder))
except OSError as e:
if e.errno == errno.EEXIST:
pass
else:
raise
# download pkl files
print('Downloading ' + self.url)
filename = self.url.rpartition('/')[2]
file_path = os.path.join(self.root, self.raw_folder, filename)
if not os.path.exists(file_path.replace('.gz', '')):
data = urllib.request.urlopen(self.url)
with open(file_path, 'wb') as f:
f.write(data.read())
with open(file_path.replace('.gz', ''), 'wb') as out_f, \
gzip.GzipFile(file_path) as zip_f:
out_f.write(zip_f.read())
os.unlink(file_path)
# process and save as torch files
print('Processing...')
# load MNIST-M images from pkl file
with open(file_path.replace('.gz', ''), "rb") as f:
mnist_m_data = pickle.load(f, encoding='bytes')
mnist_m_train_data = torch.ByteTensor(mnist_m_data[b'train'])
mnist_m_test_data = torch.ByteTensor(mnist_m_data[b'test'])
# get MNIST labels
mnist_train_labels = datasets.MNIST(root=self.mnist_root,
train=True,
download=True).train_labels
mnist_test_labels = datasets.MNIST(root=self.mnist_root,
train=False,
download=True).test_labels
# save MNIST-M dataset
training_set = (mnist_m_train_data, mnist_train_labels)
test_set = (mnist_m_test_data, mnist_test_labels)
with open(os.path.join(self.root,
self.processed_folder,
self.training_file), 'wb') as f:
torch.save(training_set, f)
with open(os.path.join(self.root,
self.processed_folder,
self.test_file), 'wb') as f:
torch.save(test_set, f)
print('Done!')
def get_mnistm(train, transform, path, image_size=28, batch_size=32, in_memory=True, num_channel=1, is_balanced=False,
drop_last=True, download=True):
"""Get MNISTM dataset loader."""
# dataset and data loader
mnistm_dataset = MNISTM(root=f"{path}/data/", train=train, transform=transform, download=download)
if in_memory:
mnistm_data_loader = torch.utils.data.DataLoader(
dataset=mnistm_dataset,
batch_size=1,
shuffle=True,
drop_last=False)
data = torch.zeros((len(mnistm_data_loader), num_channel, image_size, image_size))
label = torch.zeros(len(mnistm_data_loader))
for i, (data_, target) in enumerate(mnistm_data_loader):
# print(i, data_.shape)
data[i] = data_
label[i] = target
full_data = torch.utils.data.TensorDataset(data, label.long())
if is_balanced:
mnistm_data_loader = torch.utils.data.DataLoader(
dataset=full_data,
batch_size=batch_size,
sampler=BalancedBatchSampler(full_data, in_memory=True),
drop_last=drop_last)
else:
mnistm_data_loader = torch.utils.data.DataLoader(
dataset=full_data,
batch_size=batch_size,
shuffle=True,
drop_last=drop_last)
else:
if is_balanced:
mnistm_data_loader = torch.utils.data.DataLoader(
dataset=mnistm_dataset,
batch_size=batch_size,
sampler=BalancedBatchSampler(mnistm_dataset),
drop_last=drop_last)
else:
mnistm_data_loader = torch.utils.data.DataLoader(
dataset=mnistm_dataset,
batch_size=batch_size,
shuffle=True,
drop_last=drop_last)
return mnistm_data_loader
| [
"m.kirchmeyer@criteo.com"
] | m.kirchmeyer@criteo.com |
59f303f4bca380ea680fe9475017fa186b011f1a | 210f3eb02d26831bf4dda1d92973a678d12ca5ce | /session9/p2.1.py | 29053fdcbb1a6b86a20cacc427c08b7dada3a80a | [] | no_license | anh135/C4T-BO1 | 2b3480aa3b30a38eb03bc07ece79708fcce86ea1 | 8c665bf690ab742ea650c1af8f91e7d558bf7343 | refs/heads/master | 2020-06-20T15:51:32.722418 | 2019-07-16T09:59:51 | 2019-07-16T09:59:51 | 197,168,282 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 100 | py | items = ['red', 'blue', 'green', 'orange']
i = int(input("position"))
print("color:", items[i] )
| [
"nquanganh135@gmail.com"
] | nquanganh135@gmail.com |
4b324a9f9ea99b231e13b55494bd0092b1cf52ec | c3ca0bcea4d1b4013a0891f014928922fc81fe7a | /examples/multi_step_training.py | 605e0ac42e4b43a5d9c9b7ba9d1573554d4f6c74 | [
"MIT"
] | permissive | takuseno/d3rlpy | 47894b17fc21fab570eca39fe8e6925a7b5d7d6f | 4ba297fc6cd62201f7cd4edb7759138182e4ce04 | refs/heads/master | 2023-08-23T12:27:45.305758 | 2023-08-14T12:07:03 | 2023-08-14T12:07:03 | 266,369,147 | 1,048 | 222 | MIT | 2023-09-02T08:12:48 | 2020-05-23T15:51:51 | Python | UTF-8 | Python | false | false | 1,483 | py | import argparse
import gym
import d3rlpy
GAMMA = 0.99
def main() -> None:
parser = argparse.ArgumentParser()
parser.add_argument("--env", type=str, default="Pendulum-v1")
parser.add_argument("--seed", type=int, default=1)
parser.add_argument("--n-steps", type=int, default=1)
parser.add_argument("--gpu", action="store_true")
args = parser.parse_args()
env = gym.make(args.env)
eval_env = gym.make(args.env)
# fix seed
d3rlpy.seed(args.seed)
d3rlpy.envs.seed_env(env, args.seed)
d3rlpy.envs.seed_env(eval_env, args.seed)
# setup algorithm
sac = d3rlpy.algos.SACConfig(
batch_size=256,
gamma=GAMMA,
actor_learning_rate=3e-4,
critic_learning_rate=3e-4,
temp_learning_rate=3e-4,
action_scaler=d3rlpy.preprocessing.MinMaxActionScaler(),
).create(device=args.gpu)
# multi-step transition sampling
transition_picker = d3rlpy.dataset.MultiStepTransitionPicker(
n_steps=args.n_steps,
gamma=GAMMA,
)
# replay buffer for experience replay
buffer = d3rlpy.dataset.create_fifo_replay_buffer(
limit=100000,
env=env,
transition_picker=transition_picker,
)
# start training
sac.fit_online(
env,
buffer,
eval_env=eval_env,
n_steps=100000,
n_steps_per_epoch=1000,
update_interval=1,
update_start_step=1000,
)
if __name__ == "__main__":
main()
| [
"takuma.seno@gmail.com"
] | takuma.seno@gmail.com |
350796e30288007d708560cb8c78151b69807870 | 2b2d000525205763a8379621f2413c1c5dae1aa0 | /resize2.py | 869d8b06bdaa1fd07f0269feb4a328531c0c9670 | [] | no_license | DAYARAM99/opencv- | 0c6defbe034c3a237c4093f362dfe65b884e6e3b | af37fd6823cf956fe9b1d13b6727833888e1a443 | refs/heads/master | 2020-08-11T04:31:12.622793 | 2019-10-11T17:23:54 | 2019-10-11T17:23:54 | 214,491,923 | 0 | 0 | null | 2019-10-11T17:16:59 | 2019-10-11T17:16:59 | null | UTF-8 | Python | false | false | 373 | py | # -*- coding: utf-8 -*-
"""
Created on Fri Oct 4 05:44:06 2019
@author: Rajat arya
"""
import cv2
img=cv2.imread("index.png")
resize_img = cv2.resize(img, (int(img.shape[1]/2), int(img.shape[0]/2)))
resized_image = cv2.resize(img, (650,500))
resized_image = cv2.resize(img, (650,500))
cv2.imshow("image",resize_img)
cv2.waitKey(0)
cv2.destroyAllWindows()
| [
"noreply@github.com"
] | noreply@github.com |
75a01a39bc004c6914c4510e6c3287cc71942b9a | 478fe983582eee010b9de9a446383c02e2c3b449 | /utils/merge_fastq.py | 3908ca836dcf3e7df05759cfcc039c4fe0fb7116 | [] | no_license | jrw24/SRI37240 | 79e9d4f1090e3b19fa493cc28e559dfedc917041 | ddb86a12f60abaf593df627b6ed6512097ceb33a | refs/heads/master | 2021-07-23T14:13:02.466810 | 2020-02-13T19:37:13 | 2020-02-13T19:37:13 | 240,345,550 | 0 | 1 | null | null | null | null | UTF-8 | Python | false | false | 1,108 | py | ### Script for merging fastq files from seperate experiments
import sys
import os
import subprocess
import argparse
parser = argparse.ArgumentParser()
parser.add_argument('--inputDir', help= 'directory with fastq files')
parser.add_argument('--outputDir', help = 'directory to send output')
args = parser.parse_args()
inpath = args.inputDir
outpath = args.outputDir
fq1 = [
"1_dmso_A",
"4_g418_A",
"7_sri37240_A"
]
fq2 = [
"2_dmso_B",
"5_g418_B",
"8_sri37240_B"
]
fq3 = [
"3_dmso_C",
"6_g418_C",
"9_sri37240_C"
]
fq_merged = [
"1_dmso",
"2_g418",
"3_sri372340"
]
# if not os.path.exists(FASTQpath): os.makedirs(FASTQpath)
def mergeFastQ(fq1Input, fq2Input, fq3Input, fqOutput):
fq1 = '%s/%s*.fastq.gz' % (inpath, fq1Input)
fq2 = '%s/%s*.fastq.gz' % (inpath, fq2Input)
fg3 = '%s/%s*.fastq.gz' % (inpath, fq3Input)
fqOut = '%s/%s.fastq.gz' % (outpath, fqOutput)
merge_command = 'cat %s %s %s > %s' % (fq1, fq2, fq3, fqOut)
print merge_command
os.system(merge_command)
for sample in range(len(fq_merged)):
mergeFastQ(fq1[sample], fq2[sample], fq3[sample], fq_merged[sample])
| [
"greenlab@greenlabs-pro.win.ad.jhu.edu"
] | greenlab@greenlabs-pro.win.ad.jhu.edu |
2a407ad94d7f94a71b7a2950a9fb841ae5678614 | ca5b57ee732081cf03de08e8f640ba2197b1a11e | /Binary Search/Tree/Longest Tree Sum Path From Root to Leaf.py | cc5a508e70e1b35eaaeddd53b9909ef5fddcf6df | [] | no_license | Ajinkya-Sonawane/Problem-Solving-in-Python | cf0507fca0968f6eef02492656f16b256cc0d07c | 96529af9343d4a831c0f8f4f92df87a49e155854 | refs/heads/main | 2023-06-06T12:46:57.990001 | 2021-06-12T19:38:02 | 2021-06-12T19:38:02 | 369,957,412 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 746 | py | # https://binarysearch.com/problems/Longest-Tree-Sum-Path-From-Root-to-Leaf
# class Tree:
# def __init__(self, val, left=None, right=None):
# self.val = val
# self.left = left
# self.right = right
class Solution:
def solve(self, root):
return self.traverse(root)[0]
def traverse(self,root):
if not root:
return 0,0
l,heightL = self.traverse(root.left)
r,heightR = self.traverse(root.right)
temp = 0
if heightL == heightR:
temp = max(l,r) + root.val
return temp,heightL+1
if heightL > heightR:
temp = l + root.val
return temp,heightL+1
temp = r + root.val
return temp,heightR+1 | [
"sonawaneajinks@gmail.com"
] | sonawaneajinks@gmail.com |
e20a5ee9ecd63a4818ab1e7040a2fe0646911b83 | 6af28264b86db139af2a885a7355be6184e2af7d | /backend/schedule_randomiser.py | de684df001503a678c69bc0c97b7ed87185e89ca | [] | no_license | LieutenantPorky/ember | 8792bc5ce2c48a0c8380b9ccfa08a337ca85308e | 39385d36be49eaad9ffac6c57ff55361d9100e03 | refs/heads/master | 2020-12-14T12:47:21.242585 | 2020-01-19T14:49:52 | 2020-01-19T14:49:52 | 234,749,374 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 1,962 | py | from peewee import *
from playhouse.sqlite_ext import *
import json
import numpy as np
from datetime import date, datetime, time, timedelta
week = [date(day=20 + i,month=1,year=2020) for i in range(0,5)]
#[print(i.isoformat()) for i in week]
randClasses = [
["9:00", "10:00", "Intro to Minecraft"],
["11:00", "13:00", "Applied Numerology"],
["13:00", "14:00", "Physics of Kitties"],
["15:00", "17:00", "Computational Turbodynamics"],
["17:00", "18:00", "Pro Haxxing 101"],
]
def getRand():
randSchedule = {"timetable":{}}
for day in week:
daySchedule = [{"start_time":i[0], "end_time":i[1], "module":{"name":i[2]}} for i in randClasses if np.random.random() > 0.5]
randSchedule["timetable"][day.isoformat()] = daySchedule
return randSchedule
def getZucc():
randSchedule = {"timetable":{}}
for day in week:
daySchedule = []
randSchedule["timetable"][day.isoformat()] = daySchedule
return randSchedule
#print(json.dumps(randSchedule, sort_keys=True, indent=4))
#print(json.dumps(randSchedule, sort_keys=True)
bios = [
"A lonely soul looking for love",
"YeEt",
"Hello world",
"I just want someone to buy me dinner"
]
usersDB = SqliteDatabase("User.db")
class User(Model):
username = CharField(unique=True)
id = AutoField()
schedule = JSONField()
bio = TextField()
class Meta:
database = usersDB
# zucc = User.get(username="Mark the Zucc Zuccson")
# zucc.schedule=json.dumps(getZucc())
# zucc.save()
for i in User.select():
print(i.username, i.bio)
# usersDB.create_tables([User])
# for name in ["Bob", "Bill", "Jeb", "Caroline", "Taylor", "Jim", "Hubert", "Lily", "Timothy", "Jerrington"]:
# newUser = User(username=name,schedule=json.dumps(getRand()), bio = bios[np.random.randint(4)])
# newUser.save()
# zucc = User(username="Mark the Zucc Zuccson", schedule=[], bio = "Single lizard robot looking for cute girl to steal data with")
# zucc.save()
| [
"jacopo@siniscalco.eu"
] | jacopo@siniscalco.eu |
1fff2c19c3e1dc938c736276b810be8d34a5d060 | 7cfbeaa4fe1cdf8e90e4c42fe575a3b76cbf64ab | /setup.py | 193d68ce2b3f00e87fe2c3aacbab22886d5bf1a4 | [
"MIT"
] | permissive | keans/timecache | dac029b51e63fdbb6afa03ee33b8b08806a34993 | 9355e8b155e8071860942303ad364d719742eb78 | refs/heads/master | 2020-03-27T19:29:37.022521 | 2018-09-02T17:25:25 | 2018-09-02T17:25:25 | 146,992,718 | 1 | 0 | null | null | null | null | UTF-8 | Python | false | false | 1,124 | py | from setuptools import setup, find_packages
import codecs
import os
# get current directory
here = os.path.abspath(os.path.dirname(__file__))
def get_long_description():
"""
get long description from README.rst file
"""
with codecs.open(os.path.join(here, "README.rst"), "r", "utf-8") as f:
return f.read()
setup(
name='timecache',
version='0.0.4',
description='Time Cache',
long_description=get_long_description(),
url='https://github.com',
author='Ansgar Kellner',
author_email='keans@gmx.net',
license='MIT',
classifiers=[
'Development Status :: 3 - Alpha',
'Intended Audience :: Developers',
'Topic :: Software Development :: Build Tools',
'License :: OSI Approved :: MIT License',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.5',
'Programming Language :: Python :: 3.6',
'Programming Language :: Python :: 3.7',
],
keywords='python packaging',
packages=find_packages(
exclude=['contrib', 'docs', 'tests']
),
install_requires=[],
)
| [
"keans@gmx.de"
] | keans@gmx.de |
e7277b90bf91d0a0aa2e1eb71e0f8754381554a3 | 12a8b0779182959fe2e48b3df12b093c970da4d8 | /listings/models.py | eacf6c83d0b2667e7b237b3573abd1e5ffa3de36 | [] | no_license | kdogan11/rf_project | 8169e8dd74eeabc1e3d1934239ed35df58e1542b | ec245e19b80ced07a1fbe8468479d18e46d34ac0 | refs/heads/master | 2020-04-10T14:24:47.281703 | 2018-12-09T21:28:10 | 2018-12-09T21:28:10 | 161,076,644 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 1,474 | py | from django.db import models
from datetime import datetime
from realtors.models import Realtor
class Listing(models.Model):
realtor = models.ForeignKey(Realtor, on_delete=models.DO_NOTHING)
title = models.CharField(max_length = 200)
address = models.CharField(max_length = 200)
city = models.CharField(max_length = 100)
state = models.CharField(max_length = 100)
zipcode = models.CharField(max_length = 20)
description = models.TextField(blank = True)
price = models.IntegerField()
bedrooms = models.IntegerField()
bathrooms = models.DecimalField(max_digits = 2, decimal_places = 1)
garage = models.IntegerField(default=0)
sqft = models.IntegerField()
lot_size = models.DecimalField(max_digits=5, decimal_places=1)
photo_main = models.ImageField(upload_to = 'photos/%Y%m/%d/')
photo_1 = models.ImageField(upload_to = 'photos/%Y%m/%d/', blank = True)
photo_2 = models.ImageField(upload_to = 'photos/%Y%m/%d/', blank = True)
photo_3 = models.ImageField(upload_to = 'photos/%Y%m/%d/', blank = True)
photo_4 = models.ImageField(upload_to = 'photos/%Y%m/%d/', blank = True)
photo_5 = models.ImageField(upload_to = 'photos/%Y%m/%d/', blank = True)
photo_6 = models.ImageField(upload_to = 'photos/%Y%m/%d/', blank = True)
is_published = models.BooleanField(default=True)
list_date = models.DateTimeField(default = datetime.now, blank = True)
def __str__(self):
return self.title
| [
"kdogan11@gmail.com"
] | kdogan11@gmail.com |
b1487c88e0ba1b6b72123718eba565f36d6903b3 | 8b0dcbc828284e273e1f2065b8d4870521681455 | /app.py | 0a9d7432681831d781202487642fbe84c4b276b5 | [] | no_license | BCStudentSoftwareDevTeam/Scrolling-Font-Changer | d46a06b9b750e2bee364f5d9c14fd36553980f4c | b9717eac8e207b6c2d6e3b6eb07a94bd7a0d5f96 | refs/heads/master | 2020-08-01T23:22:59.518056 | 2019-09-30T18:37:53 | 2019-09-30T18:37:53 | 211,156,250 | 0 | 0 | null | 2019-09-26T18:29:04 | 2019-09-26T18:29:03 | null | UTF-8 | Python | false | false | 2,889 | py | from flask import Flask, render_template, request, redirect, url_for
import threading, time
app = Flask(__name__)
app.jinja_env.trim_blocks = True
app.jinja_env.lstrip_blocks = True
@app.route('/young')
def young():
return render_template("main.html", age = "young")
@app.route('/old')
def old():
return render_template("main.html", age = "old")
@app.route('/editDisplay')
def editDisplay():
f = open('currentFont.txt', 'r')
font = f.readline()
f.close()
return render_template("editDisplay.html", font = font)
@app.route('/getWords')
def getWords():
f = open('currentFont.txt', 'r')
font = f.readline()
# print(font)
f.close()
f = open('words.txt', 'r')
words = f.read()
f.close()
f = open('muteDisplay.txt', 'r')
muter = f.read()
f.close()
return muter + "||" + font + "||" + words
@app.route('/vetWords')
def vetWords():
f = open('pendingWords.txt', 'r')
words = f.readlines()
print(words)
f.close()
f = open('words.txt', 'r')
vettedWords = f.readlines()
return render_template("vetWords.html", words = words, vettedWords = vettedWords)
@app.route('/approve/<word>')
def approve(word):
f = open('words.txt', 'a')
f.write(word + "\n")
f.close()
with open('pendingWords.txt', 'r') as f:
lines = f.readlines()
with open('pendingWords.txt', 'w') as f:
# Remove word from pendingWords that was approved
for line in lines:
print("Line: ", line.strip())
print("word: ", word.strip())
print("Evaluated: ", line.strip() != word.strip())
if line.strip() != word.strip():
print("adding", word, line.strip("\n"))
f.write(line)
else:
print("Skipping: ", line)
f = open('pendingWords.txt', 'r')
words = f.read()
f.close()
if len(words) > 0:
return words
else:
return ""
@app.route('/removeWord', methods = ["POST"])
def removeWord():
word = request.form.get("word")
f = open('pendingWords.txt', 'r')
words = f.read()
words = words.replace(word, " ")
f = open('pendingWords.txt', 'w')
f.write(words)
f.close()
return redirect(url_for('vetWords'))
@app.route('/sendWord/<age>/<word>')
def sendWord(age, word):
f = open('pendingWords.txt', 'a')
f.write(age + ": " + word.strip() + ":|: \n")
f.close()
return word
@app.route('/sendFont/<font>')
def sendFont(font):
f = open('currentFont.txt', 'w')
f.write(font)
f.close()
return font
@app.route('/getFont')
def getFont():
f = open('currentFont.txt', 'r')
font = f.read()
print(font)
f.close()
return font
def updateMuteState():
states = {"true": "false",
"false": "true",
"": "true"}
f = open('muteDisplay.txt', 'r')
currentState = f.read()
f = open('muteDisplay.txt', 'w')
f.write(states[currentState])
f.close()
@app.route('/muteDisplay')
def muteDisplay():
updateMuteState()
threading.Timer(20.0, updateMuteState).start()
f = open('muteDisplay.txt', 'r')
currentState = f.read()
f.close()
return currentState
| [
"heggens@berea.edu"
] | heggens@berea.edu |
90e17c94f2e06a546e4b52d1a8e0b9da901b50ce | 991666b692f8017c2d753b0267eb9189413f54eb | /bin/Debug/models/sum.py | 32fb9016bc4f31059e4160cc0aae8e9682a9deed | [
"Apache-2.0"
] | permissive | DaniilVdovin/InformaticalPasCompilCore | 8a0b4e532e439b2497ec35028d9bade6c37b1b98 | 6d3f6fd3bef63a355c4fd39aabccf3a768daaa7c | refs/heads/main | 2023-04-20T12:47:09.101144 | 2021-05-02T09:00:35 | 2021-05-02T09:00:35 | 363,605,177 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 101 | py | import sys
if(sys.argv[1]=="-r"):
print("r:0:100:2")
else:
print(sum(map(int,sys.argv[1:])))
| [
"stels1040533@gmail.com"
] | stels1040533@gmail.com |
3edc6166c5ab9e995f874c861d02d67b0d48ae21 | 3d671fcdd27ae90698c29d0e066c662dcd4e5ee9 | /myproject/myroot/mysite/settings.py | e0fa391dbcf7d7712f450517d82f4800b9950be3 | [] | no_license | GeethuEipe/Django | 63212ac6e4bcaefa8cc5c05ea8641e72ddd04518 | 799c1157b0d7fdab07b02ca491a722ce60219ac0 | refs/heads/main | 2023-03-26T18:32:24.119589 | 2021-03-27T12:29:41 | 2021-03-27T12:29:41 | 352,067,335 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 3,157 | py | """
Django settings for mysite project.
Generated by 'django-admin startproject' using Django 3.1.7.
For more information on this file, see
https://docs.djangoproject.com/en/3.1/topics/settings/
For the full list of settings and their values, see
https://docs.djangoproject.com/en/3.1/ref/settings/
"""
from pathlib import Path
# Build paths inside the project like this: BASE_DIR / 'subdir'.
BASE_DIR = Path(__file__).resolve().parent.parent
# Quick-start development settings - unsuitable for production
# See https://docs.djangoproject.com/en/3.1/howto/deployment/checklist/
# SECURITY WARNING: keep the secret key used in production secret!
SECRET_KEY = 'ljwyn^1i_($byl&$m#xn!34+#)6i^%l*uurs8)6dukbq!-*n%q'
# SECURITY WARNING: don't run with debug turned on in production!
DEBUG = True
ALLOWED_HOSTS = []
# Application definition
INSTALLED_APPS = [
'django.contrib.admin',
'django.contrib.auth',
'django.contrib.contenttypes',
'django.contrib.sessions',
'django.contrib.messages',
'django.contrib.staticfiles',
'event'
]
MIDDLEWARE = [
'django.middleware.security.SecurityMiddleware',
'django.contrib.sessions.middleware.SessionMiddleware',
'django.middleware.common.CommonMiddleware',
'django.middleware.csrf.CsrfViewMiddleware',
'django.contrib.auth.middleware.AuthenticationMiddleware',
'django.contrib.messages.middleware.MessageMiddleware',
'django.middleware.clickjacking.XFrameOptionsMiddleware',
]
ROOT_URLCONF = 'mysite.urls'
TEMPLATES = [
{
'BACKEND': 'django.template.backends.django.DjangoTemplates',
'DIRS': [BASE_DIR / 'mysite/templates'],
'APP_DIRS': True,
'OPTIONS': {
'context_processors': [
'django.template.context_processors.debug',
'django.template.context_processors.request',
'django.contrib.auth.context_processors.auth',
'django.contrib.messages.context_processors.messages',
],
},
},
]
WSGI_APPLICATION = 'mysite.wsgi.application'
# Database
# https://docs.djangoproject.com/en/3.1/ref/settings/#databases
DATABASES = {
'default': {
'ENGINE': 'django.db.backends.sqlite3',
'NAME': BASE_DIR / 'db.sqlite3',
}
}
# Password validation
# https://docs.djangoproject.com/en/3.1/ref/settings/#auth-password-validators
AUTH_PASSWORD_VALIDATORS = [
{
'NAME': 'django.contrib.auth.password_validation.UserAttributeSimilarityValidator',
},
{
'NAME': 'django.contrib.auth.password_validation.MinimumLengthValidator',
},
{
'NAME': 'django.contrib.auth.password_validation.CommonPasswordValidator',
},
{
'NAME': 'django.contrib.auth.password_validation.NumericPasswordValidator',
},
]
# Internationalization
# https://docs.djangoproject.com/en/3.1/topics/i18n/
LANGUAGE_CODE = 'en-us'
TIME_ZONE = 'UTC'
USE_I18N = True
USE_L10N = True
USE_TZ = True
# Static files (CSS, JavaScript, Images)
# https://docs.djangoproject.com/en/3.1/howto/static-files/
STATIC_URL = '/static/'
STATICFILES_DIRS = [BASE_DIR / 'mysite/static']
| [
"geethueipe97@gmail.com"
] | geethueipe97@gmail.com |
f73ad632a0b644f358bb65369b1122bfe0e5dda5 | 3143d971afa307c824c76cb3b6fba27b47b53ff8 | /showTest/grade_crawle/src/example.py | 869587302567e21d03fd1ee54a9abc628ccaef67 | [] | no_license | InnerAc/GradeQuery | 62838fef31af17e012198bfdc754c83abaca5869 | 53769a50e27b9e4aed146c30a8d30e05466c6c04 | refs/heads/master | 2020-05-29T17:55:52.086130 | 2016-02-27T03:06:42 | 2016-02-27T03:06:42 | 42,023,260 | 2 | 0 | null | null | null | null | UTF-8 | Python | false | false | 821 | py | from segmentation import NormalSegmenter
from feature_extraction import SimpleFeatureExtractor
from analyzer import KNNAnalyzer
import random
import urllib
def getImage(url, file_path):
u = urllib.urlopen(url)
data = u.read()
f = open(file_path, 'wb')
f.write(data)
f.close()
segmenter = NormalSegmenter()
extractor = SimpleFeatureExtractor( feature_size=20, stretch=False )
analyzer = KNNAnalyzer( segmenter, extractor)
analyzer.train('../data/features.jpg')
for i in range(1):
rand = random.random()
url = "http://202.119.113.135/validateCodeAction.do?random=" + str(rand);
#print url
file_path = "../train/crawler.jpg"
getImage(url,file_path)
result = analyzer.analyze('../train/crawler.jpg')
print result
#analyzer.display()
#analyzer.display_binary()
| [
"anjicun@live.com"
] | anjicun@live.com |
c5e7d15d5d15185d551a3d9dcfd54449f26ac850 | 986a9c6a9463e029e33dfa564aeaabde1dc573c8 | /traffic_sign.py | af907a13df2fc8d470aefae12f80a555275d95a0 | [] | no_license | rushikeshkorde/Traffic-Signs-Classification | 15c1eda0261771cc8d5932305773bf9a1f6f1172 | 226deb010b5973c3b4d82fa09c69510a75542241 | refs/heads/master | 2022-08-16T01:54:48.752862 | 2020-05-31T14:13:22 | 2020-05-31T14:13:22 | 267,787,807 | 2 | 0 | null | null | null | null | UTF-8 | Python | false | false | 3,218 | py | import numpy as np
import pandas as pd
import matplotlib.pyplot as plt
import cv2
import tensorflow as tf
from PIL import Image
import os
from sklearn.model_selection import train_test_split
from keras.utils import to_categorical
from keras.models import Sequential, load_model
from keras.layers import Conv2D, MaxPool2D, Dense, Flatten, Dropout
data = []
labels = []
classes = 43
cur_path = os.getcwd()
#Retrieving the images and their labels
for i in range(classes):
path = os.path.join(cur_path,'train',str(i))
images = os.listdir(path)
for a in images:
try:
image = Image.open(path + '\\'+ a)
image = image.resize((30,30))
image = np.array(image)
#sim = Image.fromarray(image)
data.append(image)
labels.append(i)
except:
print("Error loading image")
#Converting lists into numpy arrays
data = np.array(data)
labels = np.array(labels)
print(data.shape, labels.shape)
#Splitting training and testing dataset
X_train, X_test, y_train, y_test = train_test_split(data, labels, test_size=0.2, random_state=42)
print(X_train.shape, X_test.shape, y_train.shape, y_test.shape)
#Converting the labels into one hot encoding
y_train = to_categorical(y_train, 43)
y_test = to_categorical(y_test, 43)
#Building the model
model = Sequential()
model.add(Conv2D(filters=32, kernel_size=(5,5), activation='relu', input_shape=X_train.shape[1:]))
model.add(Conv2D(filters=32, kernel_size=(5,5), activation='relu'))
model.add(MaxPool2D(pool_size=(2, 2)))
model.add(Dropout(rate=0.25))
model.add(Conv2D(filters=64, kernel_size=(3, 3), activation='relu'))
model.add(Conv2D(filters=64, kernel_size=(3, 3), activation='relu'))
model.add(MaxPool2D(pool_size=(2, 2)))
model.add(Dropout(rate=0.25))
model.add(Flatten())
model.add(Dense(256, activation='relu'))
model.add(Dropout(rate=0.5))
model.add(Dense(43, activation='softmax'))
#Compilation of the model
model.compile(loss='categorical_crossentropy', optimizer='adam', metrics=['accuracy'])
epochs = 15
history = model.fit(X_train, y_train, batch_size=32, epochs=epochs, validation_data=(X_test, y_test))
model.save("my_model.h5")
#plotting graphs for accuracy
plt.figure(0)
plt.plot(history.history['accuracy'], label='training accuracy')
plt.plot(history.history['val_accuracy'], label='val accuracy')
plt.title('Accuracy')
plt.xlabel('epochs')
plt.ylabel('accuracy')
plt.legend()
plt.show()
plt.figure(1)
plt.plot(history.history['loss'], label='training loss')
plt.plot(history.history['val_loss'], label='val loss')
plt.title('Loss')
plt.xlabel('epochs')
plt.ylabel('loss')
plt.legend()
plt.show()
#testing accuracy on test dataset
from sklearn.metrics import accuracy_score
y_test = pd.read_csv('Test.csv')
labels = y_test["ClassId"].values
imgs = y_test["Path"].values
data=[]
for img in imgs:
image = Image.open(img)
image = image.resize((30,30))
data.append(np.array(image))
X_test=np.array(data)
pred = model.predict_classes(X_test)
#Accuracy with the test data
from sklearn.metrics import accuracy_score
print(accuracy_score(labels, pred))
| [
"noreply@github.com"
] | noreply@github.com |
9d72434ff4c42cd9934c292efbbb2cdcf75e5a58 | f719ec76a8417fc05a2d46ada2501052e2bf9469 | /exp_runners/traffic/cent_traffic_runner.py | 2179e0136d393da470ab919a3989f6ab9e970282 | [] | no_license | yang-xy20/DICG | cc31064a3e4a3dd01414161e42b228c2c09bfea7 | c64ba9dbbe0f2b745cd04ce516aa1fed4c2cffc7 | refs/heads/master | 2023-07-04T18:25:18.461196 | 2021-08-19T21:34:06 | 2021-08-19T21:34:06 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 9,351 | py | import sys
import os
current_file_path = os.path.dirname(os.path.abspath(__file__))
sys.path.append(current_file_path + '/../../')
import socket
import collections
import numpy as np
import argparse
import joblib
import time
import matplotlib.pyplot as plt
from types import SimpleNamespace
import torch
from torch.nn import functional as F
import akro
import garage
from garage import wrap_experiment
from garage.envs import GarageEnv
from garage.experiment.deterministic import set_seed
from envs import TrafficJunctionWrapper
from dicg.torch.baselines import GaussianMLPBaseline
from dicg.torch.algos import CentralizedMAPPO
from dicg.torch.policies import CentralizedCategoricalMLPPolicy
from dicg.experiment.local_runner_wrapper import LocalRunnerWrapper
from dicg.sampler import CentralizedMAOnPolicyVectorizedSampler
def run(args):
# device = torch.device('cuda' if torch.cuda.is_available() else 'cpu')
# garage.torch.utils.set_gpu_mode(mode=torch.cuda.is_available())
# print(garage.torch.utils.global_device())
if args.exp_name is None:
exp_layout = collections.OrderedDict([
('cent{}_ppo', ''),
('entcoeff={}', args.ent),
('dim={}', args.dim),
('nagents={}', args.n_agents),
('difficulty={}', args.difficulty),
('curr={}', bool(args.curriculum)),
('steps={}', args.max_env_steps),
('nenvs={}', args.n_envs),
('bs={:0.0e}', args.bs),
('splits={}', args.opt_n_minibatches),
('miniepoch={}', args.opt_mini_epochs),
('seed={}', args.seed)
])
exp_name = '_'.join(
[key.format(val) for key, val in exp_layout.items()]
)
else:
exp_name = args.exp_name
prefix = 'traffic'
id_suffix = ('_' + str(args.run_id)) if args.run_id != 0 else ''
unseeded_exp_dir = './data/' + args.loc +'/' + exp_name[:-7]
exp_dir = './data/' + args.loc +'/' + exp_name + id_suffix
# Enforce
args.center_adv = False if args.entropy_method == 'max' else args.center_adv
if args.mode == 'train':
# making sequential log dir if name already exists
@wrap_experiment(name=exp_name,
prefix=prefix,
log_dir=exp_dir,
snapshot_mode='last',
snapshot_gap=1)
def train_traffic(ctxt=None, args_dict=vars(args)):
args = SimpleNamespace(**args_dict)
set_seed(args.seed)
if args.curriculum:
curr_start = int(0.125 * args.n_epochs)
curr_end = int(0.625 * args.n_epochs)
else:
curr_start = 0
curr_end = 0
args.add_rate_min = args.add_rate_max
env = TrafficJunctionWrapper(
centralized=True, # centralized training and critic
dim=args.dim,
vision=1,
add_rate_min=args.add_rate_min,
add_rate_max=args.add_rate_max,
curr_start=curr_start,
curr_end=curr_end,
difficulty=args.difficulty,
n_agents=args.n_agents,
max_steps=args.max_env_steps
)
env = GarageEnv(env)
runner = LocalRunnerWrapper(
ctxt,
eval=args.eval_during_training,
n_eval_episodes=args.n_eval_episodes,
eval_greedy=args.eval_greedy,
eval_epoch_freq=args.eval_epoch_freq,
save_env=env.pickleable
)
hidden_nonlinearity = F.relu if args.hidden_nonlinearity == 'relu' \
else torch.tanh
policy = CentralizedCategoricalMLPPolicy(
env.spec,
env.n_agents,
hidden_nonlinearity=hidden_nonlinearity,
hidden_sizes=args.hidden_sizes,
name='dec_categorical_mlp_policy'
)
baseline = GaussianMLPBaseline(env_spec=env.spec,
hidden_sizes=(64, 64, 64))
# Set max_path_length <= max_steps
# If max_path_length > max_steps, algo will pad obs
# obs.shape = torch.Size([n_paths, algo.max_path_length, feat_dim])
algo = CentralizedMAPPO(
env_spec=env.spec,
policy=policy,
baseline=baseline,
max_path_length=args.max_env_steps, # Notice
discount=args.discount,
center_adv=bool(args.center_adv),
positive_adv=bool(args.positive_adv),
gae_lambda=args.gae_lambda,
policy_ent_coeff=args.ent,
entropy_method=args.entropy_method,
stop_entropy_gradient=True \
if args.entropy_method == 'max' else False,
clip_grad_norm=args.clip_grad_norm,
optimization_n_minibatches=args.opt_n_minibatches,
optimization_mini_epochs=args.opt_mini_epochs,
)
runner.setup(algo, env,
sampler_cls=CentralizedMAOnPolicyVectorizedSampler,
sampler_args={'n_envs': args.n_envs})
runner.train(n_epochs=args.n_epochs,
batch_size=args.bs)
train_traffic(args_dict=vars(args))
elif args.mode in ['restore', 'eval']:
data = joblib.load(exp_dir + '/params.pkl')
env = data['env']
algo = data['algo']
if args.mode == 'restore':
from dicg.experiment.runner_utils import restore_training
restore_training(exp_dir, exp_name, args,
env_saved=env.pickleable, env=env)
elif args.mode == 'eval':
env.eval(algo.policy, n_episodes=args.n_eval_episodes, greedy=args.eval_greedy,
load_from_file=True, max_steps=args.max_env_steps, render=args.render)
if __name__ == '__main__':
parser = argparse.ArgumentParser()
# Meta
parser.add_argument('--mode', '-m', type=str, default='train')
parser.add_argument('--loc', type=str, default='local')
parser.add_argument('--exp_name', type=str, default=None)
# Train
parser.add_argument('--seed', '-s', type=int, default=1)
parser.add_argument('--n_epochs', type=int, default=1000)
parser.add_argument('--bs', type=int, default=60000)
parser.add_argument('--n_envs', type=int, default=1)
# Eval
parser.add_argument('--run_id', type=int, default=0) # sequential naming
parser.add_argument('--n_eval_episodes', type=int, default=100)
parser.add_argument('--render', type=int, default=0)
parser.add_argument('--inspect_steps', type=int, default=0)
parser.add_argument('--eval_during_training', type=int, default=1)
parser.add_argument('--eval_greedy', type=int, default=1)
parser.add_argument('--eval_epoch_freq', type=int, default=5)
# Env
parser.add_argument('--max_env_steps', type=int, default=20)
parser.add_argument('--dim', type=int, default=8)
parser.add_argument('--n_agents', '-n', type=int, default=5)
parser.add_argument('--difficulty', type=str, default='easy')
parser.add_argument('--add_rate_max', type=float, default=0.3)
parser.add_argument('--add_rate_min', type=float, default=0.1)
parser.add_argument('--curriculum', type=int, default=0)
# Algo
# parser.add_argument('--max_algo_path_length', type=int, default=n_steps)
parser.add_argument('--hidden_nonlinearity', type=str, default='tanh')
parser.add_argument('--discount', type=float, default=0.99)
parser.add_argument('--center_adv', type=int, default=1)
parser.add_argument('--positive_adv', type=int, default=0)
parser.add_argument('--gae_lambda', type=float, default=0.97)
parser.add_argument('--ent', type=float, default=0.02) # 0.01 is too small
parser.add_argument('--entropy_method', type=str, default='regularized')
parser.add_argument('--clip_grad_norm', type=float, default=7)
parser.add_argument('--opt_n_minibatches', type=int, default=4,
help='The number of splits of a batch of trajectories for optimization.')
parser.add_argument('--opt_mini_epochs', type=int, default=10,
help='The number of epochs the optimizer runs for each batch of trajectories.')
# Policy
# Example: --encoder_hidden_sizes 12 123 1234
parser.add_argument('--hidden_sizes', nargs='+', type=int)
args = parser.parse_args()
# Enforce values
if args.difficulty == 'hard':
args.max_env_steps = 60
args.dim = 18
args.n_agents = 20
args.add_rate_min = 0.02
args.add_rate_max = 0.05
elif args.difficulty == 'medium':
args.max_env_steps = 40
args.dim = 14
args.n_agents = 10
args.add_rate_min = 0.05
args.add_rate_max = 0.2
elif args.difficulty == 'easy':
args.max_env_steps = 20
args.dim = 8
args.n_agents = 5
args.add_rate_min = 0.1
args.add_rate_max = 0.3
if args.hidden_sizes is None:
args.hidden_sizes = [265, 128, 64]
run(args)
| [
"lisheng@stanford.edu"
] | lisheng@stanford.edu |
c02da091fdeacb53d6ce13fd9ec1162d84589d2e | 058a0b8ca26624c74edf260e19ead70548f66e25 | /UserInterface/Admin_Mode_1.py | 39e252dc85979c0e3bc38a874e3642c1b9389258 | [] | no_license | chenhuik029/TimeLog_py | 35a9eb291e81c5b4f28de84f0ae2f0a677ebddd8 | a47d174dd189cb8221b7bd8545b4e91a0d0ba4ab | refs/heads/master | 2023-06-08T02:40:20.856109 | 2021-07-04T13:54:55 | 2021-07-04T13:54:55 | 330,116,703 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 20,101 | py | # -*- coding: utf-8 -*-
# Form implementation generated from reading ui file '04_Admin_Mode_01.ui'
#
# Created by: PyQt5 UI code generator 5.15.1
#
# WARNING: Any manual changes made to this file will be lost when pyuic5 is
# run again. Do not edit this file unless you know what you are doing.
from PyQt5 import QtCore, QtGui, QtWidgets
class Ui_Admin_Mode_1(object):
def setupUi(self, Admin_Mode_1):
Admin_Mode_1.setObjectName("Admin_Mode_1")
Admin_Mode_1.resize(1022, 835)
palette = QtGui.QPalette()
brush = QtGui.QBrush(QtGui.QColor(255, 255, 255))
brush.setStyle(QtCore.Qt.SolidPattern)
palette.setBrush(QtGui.QPalette.Active, QtGui.QPalette.Base, brush)
brush = QtGui.QBrush(QtGui.QColor(255, 233, 190))
brush.setStyle(QtCore.Qt.SolidPattern)
palette.setBrush(QtGui.QPalette.Active, QtGui.QPalette.Window, brush)
brush = QtGui.QBrush(QtGui.QColor(255, 255, 255))
brush.setStyle(QtCore.Qt.SolidPattern)
palette.setBrush(QtGui.QPalette.Inactive, QtGui.QPalette.Base, brush)
brush = QtGui.QBrush(QtGui.QColor(255, 233, 190))
brush.setStyle(QtCore.Qt.SolidPattern)
palette.setBrush(QtGui.QPalette.Inactive, QtGui.QPalette.Window, brush)
brush = QtGui.QBrush(QtGui.QColor(255, 233, 190))
brush.setStyle(QtCore.Qt.SolidPattern)
palette.setBrush(QtGui.QPalette.Disabled, QtGui.QPalette.Base, brush)
brush = QtGui.QBrush(QtGui.QColor(255, 233, 190))
brush.setStyle(QtCore.Qt.SolidPattern)
palette.setBrush(QtGui.QPalette.Disabled, QtGui.QPalette.Window, brush)
Admin_Mode_1.setPalette(palette)
self.centralwidget = QtWidgets.QWidget(Admin_Mode_1)
self.centralwidget.setObjectName("centralwidget")
self.verticalLayout = QtWidgets.QVBoxLayout(self.centralwidget)
self.verticalLayout.setObjectName("verticalLayout")
self.frame_title = QtWidgets.QFrame(self.centralwidget)
self.frame_title.setMaximumSize(QtCore.QSize(16777215, 50))
self.frame_title.setFrameShape(QtWidgets.QFrame.NoFrame)
self.frame_title.setFrameShadow(QtWidgets.QFrame.Raised)
self.frame_title.setObjectName("frame_title")
self.verticalLayout_2 = QtWidgets.QVBoxLayout(self.frame_title)
self.verticalLayout_2.setObjectName("verticalLayout_2")
self.label_title = QtWidgets.QLabel(self.frame_title)
palette = QtGui.QPalette()
brush = QtGui.QBrush(QtGui.QColor(0, 0, 255))
brush.setStyle(QtCore.Qt.SolidPattern)
palette.setBrush(QtGui.QPalette.Active, QtGui.QPalette.WindowText, brush)
brush = QtGui.QBrush(QtGui.QColor(0, 0, 255))
brush.setStyle(QtCore.Qt.SolidPattern)
palette.setBrush(QtGui.QPalette.Inactive, QtGui.QPalette.WindowText, brush)
brush = QtGui.QBrush(QtGui.QColor(120, 120, 120))
brush.setStyle(QtCore.Qt.SolidPattern)
palette.setBrush(QtGui.QPalette.Disabled, QtGui.QPalette.WindowText, brush)
self.label_title.setPalette(palette)
font = QtGui.QFont()
font.setPointSize(18)
font.setBold(True)
font.setWeight(75)
self.label_title.setFont(font)
self.label_title.setObjectName("label_title")
self.verticalLayout_2.addWidget(self.label_title)
self.verticalLayout.addWidget(self.frame_title)
self.frame_instruction = QtWidgets.QFrame(self.centralwidget)
self.frame_instruction.setMaximumSize(QtCore.QSize(16777215, 50))
self.frame_instruction.setFrameShape(QtWidgets.QFrame.NoFrame)
self.frame_instruction.setFrameShadow(QtWidgets.QFrame.Raised)
self.frame_instruction.setObjectName("frame_instruction")
self.verticalLayout_3 = QtWidgets.QVBoxLayout(self.frame_instruction)
self.verticalLayout_3.setObjectName("verticalLayout_3")
self.label_instruction = QtWidgets.QLabel(self.frame_instruction)
font = QtGui.QFont()
font.setFamily("Arial")
font.setPointSize(14)
font.setBold(True)
font.setUnderline(True)
font.setWeight(75)
self.label_instruction.setFont(font)
self.label_instruction.setObjectName("label_instruction")
self.verticalLayout_3.addWidget(self.label_instruction)
self.verticalLayout.addWidget(self.frame_instruction)
self.frame_input = QtWidgets.QFrame(self.centralwidget)
palette = QtGui.QPalette()
brush = QtGui.QBrush(QtGui.QColor(255, 255, 255))
brush.setStyle(QtCore.Qt.SolidPattern)
palette.setBrush(QtGui.QPalette.Active, QtGui.QPalette.Base, brush)
brush = QtGui.QBrush(QtGui.QColor(255, 233, 190))
brush.setStyle(QtCore.Qt.SolidPattern)
palette.setBrush(QtGui.QPalette.Active, QtGui.QPalette.Window, brush)
brush = QtGui.QBrush(QtGui.QColor(255, 255, 255))
brush.setStyle(QtCore.Qt.SolidPattern)
palette.setBrush(QtGui.QPalette.Inactive, QtGui.QPalette.Base, brush)
brush = QtGui.QBrush(QtGui.QColor(255, 233, 190))
brush.setStyle(QtCore.Qt.SolidPattern)
palette.setBrush(QtGui.QPalette.Inactive, QtGui.QPalette.Window, brush)
brush = QtGui.QBrush(QtGui.QColor(255, 233, 190))
brush.setStyle(QtCore.Qt.SolidPattern)
palette.setBrush(QtGui.QPalette.Disabled, QtGui.QPalette.Base, brush)
brush = QtGui.QBrush(QtGui.QColor(255, 233, 190))
brush.setStyle(QtCore.Qt.SolidPattern)
palette.setBrush(QtGui.QPalette.Disabled, QtGui.QPalette.Window, brush)
self.frame_input.setPalette(palette)
self.frame_input.setStyleSheet("b")
self.frame_input.setFrameShape(QtWidgets.QFrame.NoFrame)
self.frame_input.setFrameShadow(QtWidgets.QFrame.Raised)
self.frame_input.setObjectName("frame_input")
self.horizontalLayout = QtWidgets.QHBoxLayout(self.frame_input)
self.horizontalLayout.setContentsMargins(20, -1, 20, -1)
self.horizontalLayout.setObjectName("horizontalLayout")
self.frame = QtWidgets.QFrame(self.frame_input)
self.frame.setMinimumSize(QtCore.QSize(350, 0))
self.frame.setMaximumSize(QtCore.QSize(350, 16777215))
self.frame.setFrameShape(QtWidgets.QFrame.NoFrame)
self.frame.setFrameShadow(QtWidgets.QFrame.Raised)
self.frame.setObjectName("frame")
self.verticalLayout_4 = QtWidgets.QVBoxLayout(self.frame)
self.verticalLayout_4.setSpacing(30)
self.verticalLayout_4.setObjectName("verticalLayout_4")
self.label_emp_name = QtWidgets.QLabel(self.frame)
font = QtGui.QFont()
font.setPointSize(11)
self.label_emp_name.setFont(font)
self.label_emp_name.setObjectName("label_emp_name")
self.verticalLayout_4.addWidget(self.label_emp_name)
self.label_emp_id = QtWidgets.QLabel(self.frame)
font = QtGui.QFont()
font.setPointSize(11)
self.label_emp_id.setFont(font)
self.label_emp_id.setObjectName("label_emp_id")
self.verticalLayout_4.addWidget(self.label_emp_id)
self.label_card_id = QtWidgets.QLabel(self.frame)
font = QtGui.QFont()
font.setPointSize(11)
self.label_card_id.setFont(font)
self.label_card_id.setObjectName("label_card_id")
self.verticalLayout_4.addWidget(self.label_card_id)
self.label_emp_sal = QtWidgets.QLabel(self.frame)
font = QtGui.QFont()
font.setPointSize(11)
self.label_emp_sal.setFont(font)
self.label_emp_sal.setObjectName("label_emp_sal")
self.verticalLayout_4.addWidget(self.label_emp_sal)
self.label_2 = QtWidgets.QLabel(self.frame)
font = QtGui.QFont()
font.setPointSize(11)
self.label_2.setFont(font)
self.label_2.setObjectName("label_2")
self.verticalLayout_4.addWidget(self.label_2)
self.horizontalLayout.addWidget(self.frame)
self.frame_2 = QtWidgets.QFrame(self.frame_input)
self.frame_2.setFrameShape(QtWidgets.QFrame.NoFrame)
self.frame_2.setFrameShadow(QtWidgets.QFrame.Raised)
self.frame_2.setObjectName("frame_2")
self.verticalLayout_5 = QtWidgets.QVBoxLayout(self.frame_2)
self.verticalLayout_5.setSpacing(30)
self.verticalLayout_5.setObjectName("verticalLayout_5")
self.lineEdit_emp_name = QtWidgets.QLineEdit(self.frame_2)
self.lineEdit_emp_name.setMinimumSize(QtCore.QSize(200, 40))
self.lineEdit_emp_name.setMaximumSize(QtCore.QSize(16777215, 40))
font = QtGui.QFont()
font.setPointSize(10)
self.lineEdit_emp_name.setFont(font)
self.lineEdit_emp_name.setAutoFillBackground(False)
self.lineEdit_emp_name.setObjectName("lineEdit_emp_name")
self.verticalLayout_5.addWidget(self.lineEdit_emp_name)
self.lineEdit_emp_id = QtWidgets.QLineEdit(self.frame_2)
self.lineEdit_emp_id.setMinimumSize(QtCore.QSize(200, 40))
self.lineEdit_emp_id.setMaximumSize(QtCore.QSize(16777215, 40))
font = QtGui.QFont()
font.setPointSize(10)
self.lineEdit_emp_id.setFont(font)
self.lineEdit_emp_id.setObjectName("lineEdit_emp_id")
self.verticalLayout_5.addWidget(self.lineEdit_emp_id)
self.lineEdit_card_id = QtWidgets.QLineEdit(self.frame_2)
self.lineEdit_card_id.setMinimumSize(QtCore.QSize(200, 40))
self.lineEdit_card_id.setMaximumSize(QtCore.QSize(16777215, 40))
font = QtGui.QFont()
font.setPointSize(10)
self.lineEdit_card_id.setFont(font)
self.lineEdit_card_id.setObjectName("lineEdit_card_id")
self.verticalLayout_5.addWidget(self.lineEdit_card_id)
self.lineEdit_emp_sal = QtWidgets.QLineEdit(self.frame_2)
self.lineEdit_emp_sal.setMinimumSize(QtCore.QSize(200, 40))
self.lineEdit_emp_sal.setMaximumSize(QtCore.QSize(16777215, 40))
font = QtGui.QFont()
font.setPointSize(10)
self.lineEdit_emp_sal.setFont(font)
self.lineEdit_emp_sal.setObjectName("lineEdit_emp_sal")
self.verticalLayout_5.addWidget(self.lineEdit_emp_sal)
self.comboBox_emp_stat = QtWidgets.QComboBox(self.frame_2)
self.comboBox_emp_stat.setMinimumSize(QtCore.QSize(200, 30))
self.comboBox_emp_stat.setObjectName("comboBox_emp_stat")
self.comboBox_emp_stat.addItem("")
self.comboBox_emp_stat.addItem("")
self.verticalLayout_5.addWidget(self.comboBox_emp_stat)
self.horizontalLayout.addWidget(self.frame_2)
self.verticalLayout.addWidget(self.frame_input)
self.frame_button = QtWidgets.QFrame(self.centralwidget)
self.frame_button.setFrameShape(QtWidgets.QFrame.NoFrame)
self.frame_button.setFrameShadow(QtWidgets.QFrame.Raised)
self.frame_button.setObjectName("frame_button")
self.horizontalLayout_2 = QtWidgets.QHBoxLayout(self.frame_button)
self.horizontalLayout_2.setObjectName("horizontalLayout_2")
spacerItem = QtWidgets.QSpacerItem(40, 20, QtWidgets.QSizePolicy.Expanding, QtWidgets.QSizePolicy.Minimum)
self.horizontalLayout_2.addItem(spacerItem)
self.pushButton_cancel = QtWidgets.QPushButton(self.frame_button)
self.pushButton_cancel.setMinimumSize(QtCore.QSize(100, 40))
self.pushButton_cancel.setMaximumSize(QtCore.QSize(100, 40))
font = QtGui.QFont()
font.setPointSize(10)
font.setBold(False)
font.setWeight(50)
self.pushButton_cancel.setFont(font)
self.pushButton_cancel.setAutoFillBackground(False)
self.pushButton_cancel.setStyleSheet("background: #f0f0f0")
self.pushButton_cancel.setObjectName("pushButton_cancel")
self.horizontalLayout_2.addWidget(self.pushButton_cancel)
self.pushButton_2 = QtWidgets.QPushButton(self.frame_button)
self.pushButton_2.setMinimumSize(QtCore.QSize(100, 40))
self.pushButton_2.setMaximumSize(QtCore.QSize(100, 40))
font = QtGui.QFont()
font.setPointSize(10)
self.pushButton_2.setFont(font)
self.pushButton_2.setAutoFillBackground(False)
self.pushButton_2.setStyleSheet("background: #f0f0f0")
self.pushButton_2.setObjectName("pushButton_2")
self.horizontalLayout_2.addWidget(self.pushButton_2)
self.verticalLayout.addWidget(self.frame_button)
self.label = QtWidgets.QLabel(self.centralwidget)
font = QtGui.QFont()
font.setPointSize(12)
font.setBold(True)
font.setWeight(75)
self.label.setFont(font)
self.label.setObjectName("label")
self.verticalLayout.addWidget(self.label)
self.EmployeeDataBase_Table = QtWidgets.QTableWidget(self.centralwidget)
sizePolicy = QtWidgets.QSizePolicy(QtWidgets.QSizePolicy.Expanding, QtWidgets.QSizePolicy.Expanding)
sizePolicy.setHorizontalStretch(0)
sizePolicy.setVerticalStretch(0)
sizePolicy.setHeightForWidth(self.EmployeeDataBase_Table.sizePolicy().hasHeightForWidth())
self.EmployeeDataBase_Table.setSizePolicy(sizePolicy)
self.EmployeeDataBase_Table.setLayoutDirection(QtCore.Qt.LeftToRight)
self.EmployeeDataBase_Table.setFrameShape(QtWidgets.QFrame.StyledPanel)
self.EmployeeDataBase_Table.setLineWidth(1)
self.EmployeeDataBase_Table.setSizeAdjustPolicy(QtWidgets.QAbstractScrollArea.AdjustIgnored)
self.EmployeeDataBase_Table.setEditTriggers(QtWidgets.QAbstractItemView.NoEditTriggers)
self.EmployeeDataBase_Table.setObjectName("EmployeeDataBase_Table")
self.EmployeeDataBase_Table.setColumnCount(7)
self.EmployeeDataBase_Table.setRowCount(0)
item = QtWidgets.QTableWidgetItem()
self.EmployeeDataBase_Table.setHorizontalHeaderItem(0, item)
self.EmployeeDataBase_Table.setColumnWidth(0, 30)
item = QtWidgets.QTableWidgetItem()
self.EmployeeDataBase_Table.setHorizontalHeaderItem(1, item)
self.EmployeeDataBase_Table.setColumnWidth(1, 270)
item = QtWidgets.QTableWidgetItem()
self.EmployeeDataBase_Table.setHorizontalHeaderItem(2, item)
self.EmployeeDataBase_Table.setColumnWidth(2, 200)
item = QtWidgets.QTableWidgetItem()
self.EmployeeDataBase_Table.setHorizontalHeaderItem(3, item)
self.EmployeeDataBase_Table.setColumnWidth(3, 100)
item = QtWidgets.QTableWidgetItem()
self.EmployeeDataBase_Table.setHorizontalHeaderItem(4, item)
self.EmployeeDataBase_Table.setColumnWidth(4, 100)
item = QtWidgets.QTableWidgetItem()
self.EmployeeDataBase_Table.setHorizontalHeaderItem(5, item)
self.EmployeeDataBase_Table.setColumnWidth(5, 150)
item = QtWidgets.QTableWidgetItem()
self.EmployeeDataBase_Table.setHorizontalHeaderItem(6, item)
self.EmployeeDataBase_Table.setColumnWidth(6, 100)
self.EmployeeDataBase_Table.horizontalHeader().setCascadingSectionResizes(False)
self.EmployeeDataBase_Table.horizontalHeader().setMinimumSectionSize(39)
self.EmployeeDataBase_Table.horizontalHeader().setSortIndicatorShown(True)
self.EmployeeDataBase_Table.horizontalHeader().setStretchLastSection(False)
self.EmployeeDataBase_Table.verticalHeader().setVisible(False)
self.EmployeeDataBase_Table.verticalHeader().setStretchLastSection(False)
self.verticalLayout.addWidget(self.EmployeeDataBase_Table)
spacerItem1 = QtWidgets.QSpacerItem(20, 30, QtWidgets.QSizePolicy.Minimum, QtWidgets.QSizePolicy.Maximum)
self.verticalLayout.addItem(spacerItem1)
Admin_Mode_1.setCentralWidget(self.centralwidget)
self.menubar = QtWidgets.QMenuBar(Admin_Mode_1)
self.menubar.setGeometry(QtCore.QRect(0, 0, 1022, 20))
self.menubar.setObjectName("menubar")
self.menuFile = QtWidgets.QMenu(self.menubar)
self.menuFile.setObjectName("menuFile")
self.menuNavigate = QtWidgets.QMenu(self.menubar)
self.menuNavigate.setObjectName("menuNavigate")
Admin_Mode_1.setMenuBar(self.menubar)
self.statusbar = QtWidgets.QStatusBar(Admin_Mode_1)
self.statusbar.setObjectName("statusbar")
Admin_Mode_1.setStatusBar(self.statusbar)
self.actionExit = QtWidgets.QAction(Admin_Mode_1)
self.actionExit.setObjectName("actionExit")
self.actionBack = QtWidgets.QAction(Admin_Mode_1)
self.actionBack.setObjectName("actionBack")
self.menuFile.addAction(self.actionExit)
self.menuNavigate.addAction(self.actionBack)
self.menubar.addAction(self.menuFile.menuAction())
self.menubar.addAction(self.menuNavigate.menuAction())
self.retranslateUi(Admin_Mode_1)
QtCore.QMetaObject.connectSlotsByName(Admin_Mode_1)
def retranslateUi(self, Admin_Mode_1):
_translate = QtCore.QCoreApplication.translate
Admin_Mode_1.setWindowTitle(_translate("Admin_Mode_1", "Attendance Recorder System - Admin Mode (Add Employee ID)"))
self.label_title.setText(_translate("Admin_Mode_1", "Administrator Mode"))
self.label_instruction.setText(_translate("Admin_Mode_1", "Add Employee ID"))
self.label_emp_name.setText(_translate("Admin_Mode_1", "Employee Name:\n"
"(Last Name, First Name)"))
self.label_emp_id.setText(_translate("Admin_Mode_1", "Employee ID: \n"
"(For Manual Entry)"))
self.label_card_id.setText(_translate("Admin_Mode_1", "Card ID: \n"
"(Please Tap Designated Card at Card Reader\n"
" to retrieve Card ID)"))
self.label_emp_sal.setText(_translate("Admin_Mode_1", "Employee Salary:\n"
" (For Salary Disbursement Usage)"))
self.label_2.setText(_translate("Admin_Mode_1", "Employment Status:"))
self.lineEdit_emp_name.setPlaceholderText(_translate("Admin_Mode_1", "Employee Name..."))
self.lineEdit_emp_id.setPlaceholderText(_translate("Admin_Mode_1", "Employee ID..."))
self.lineEdit_card_id.setPlaceholderText(_translate("Admin_Mode_1", "Please tap RFID Card on Card Reader for Card ID..."))
self.lineEdit_emp_sal.setPlaceholderText(_translate("Admin_Mode_1", "Employee salary..."))
self.comboBox_emp_stat.setItemText(0, _translate("Admin_Mode_1", "Active"))
self.comboBox_emp_stat.setItemText(1, _translate("Admin_Mode_1", "Inactive"))
self.pushButton_cancel.setText(_translate("Admin_Mode_1", "Back"))
self.pushButton_2.setText(_translate("Admin_Mode_1", "Apply"))
self.label.setText(_translate("Admin_Mode_1", "Employee Database"))
self.EmployeeDataBase_Table.setSortingEnabled(True)
item = self.EmployeeDataBase_Table.horizontalHeaderItem(0)
item.setText(_translate("Admin_Mode_1", "ID"))
item = self.EmployeeDataBase_Table.horizontalHeaderItem(1)
item.setText(_translate("Admin_Mode_1", "Employee Name"))
item = self.EmployeeDataBase_Table.horizontalHeaderItem(2)
item.setText(_translate("Admin_Mode_1", "Employee ID"))
item = self.EmployeeDataBase_Table.horizontalHeaderItem(3)
item.setText(_translate("Admin_Mode_1", "Card ID"))
item = self.EmployeeDataBase_Table.horizontalHeaderItem(4)
item.setText(_translate("Admin_Mode_1", "Salary"))
item = self.EmployeeDataBase_Table.horizontalHeaderItem(5)
item.setText(_translate("Admin_Mode_1", "Employee Status"))
item = self.EmployeeDataBase_Table.horizontalHeaderItem(6)
item.setText(_translate("Admin_Mode_1", "Date Joined"))
self.menuFile.setTitle(_translate("Admin_Mode_1", "File"))
self.menuNavigate.setTitle(_translate("Admin_Mode_1", "Navigate"))
self.actionExit.setText(_translate("Admin_Mode_1", "Exit"))
self.actionBack.setText(_translate("Admin_Mode_1", "Back"))
if __name__ == "__main__":
import sys
app = QtWidgets.QApplication(sys.argv)
Admin_Mode_1 = QtWidgets.QMainWindow()
ui = Ui_Admin_Mode_1()
ui.setupUi(Admin_Mode_1)
Admin_Mode_1.show()
sys.exit(app.exec_())
| [
"chenhui_k029@hotmail.com"
] | chenhui_k029@hotmail.com |
01685b4a849a3156658fa0dbdaad10650ff9d148 | b14802e3892a661fa62d9d0772f72becc0abd612 | /evaluation/get_top_socored.py | 0bd0d8919ad1d0eed44022b6a57cbb69617117bb | [] | no_license | gombru/HateSpeech | e4c4b7993354ce2cb49334b814f929364fdcb446 | 7891c7e2835f17ed2a9985abd285e19788685c66 | refs/heads/master | 2022-02-23T08:57:34.909778 | 2022-02-10T12:54:41 | 2022-02-10T12:54:41 | 138,057,409 | 6 | 2 | null | null | null | null | UTF-8 | Python | false | false | 1,326 | py | import numpy as np
import operator
import shutil
import os
model_name = 'MMHS_classification_CNNinit_SCM_ALL_epoch_10_ValAcc_62'
out_folder_name = 'top_MMHS_classification_CNNinit_SCM_ALL_epoch_10_ValAcc_62'
out_file = open('../../../datasets/HateSPic/MMHS/top_scored/' + out_folder_name + '.txt','w')
if not os.path.exists('../../../datasets/HateSPic/MMHS/top_scored/' + out_folder_name):
os.makedirs('../../../datasets/HateSPic/MMHS/top_scored/' + out_folder_name)
results = {}
with open('../../../datasets/HateSPic/MMHS/results/' + model_name + '/test.txt') as f:
for line in f:
data = line.split(',')
id = int(data[0])
label = int(data[1])
hate_score = float(data[3])
notHate_score = float(data[2])
softmax_hate_score = np.exp(hate_score) / (np.exp(hate_score) + np.exp(notHate_score))
results[id] = softmax_hate_score
results = sorted(results.items(), key=operator.itemgetter(1))
results = list(reversed(results))
for i,r in enumerate(results):
if i == 50: break
print r[1]
shutil.copyfile('../../../datasets/HateSPic/MMHS/img_resized/' + str(str(r[0])) + '.jpg', '../../../datasets/HateSPic/MMHS/top_scored/' + out_folder_name + '/' + str(i) + '-' + str(r[0]) + '.jpg')
out_file.write(str(r[0]) + '\n')
out_file.close()
print("Done") | [
"raulgombru@gmail.com"
] | raulgombru@gmail.com |
7c4cb0d388dfd9e306500f3f0b0cc9ceb415e596 | e15575c838c4f656e751a0d544f6cf1e49580305 | /User_Accounts/migrations/0001_initial.py | 326403f3fb22f7c06dd73251b6fda46741b4ffed | [] | no_license | manojakumarpanda/Blog_post_and-comment | bd764676b9f17ab95eb163bbe17c7dc4f1550a66 | 84540228d33d8e6337a12eb834ffbc2ccbdd5fc2 | refs/heads/master | 2022-07-03T21:33:10.230573 | 2020-05-14T17:39:20 | 2020-05-14T17:39:20 | 263,977,200 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 5,041 | py | # Generated by Django 2.1.7 on 2020-05-07 15:27
import django.contrib.auth.models
from django.db import migrations, models
import django.db.models.deletion
import django.utils.timezone
import uuid
class Migration(migrations.Migration):
initial = True
dependencies = [
('auth', '0009_alter_user_last_name_max_length'),
]
operations = [
migrations.CreateModel(
name='Cities',
fields=[
('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')),
('city_name', models.CharField(blank=True, default='Berhampur', max_length=50, null=True)),
],
options={
'db_table': 'cities',
'ordering': ['city_name'],
},
),
migrations.CreateModel(
name='Countrey',
fields=[
('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')),
('countrey_name', models.CharField(default='india', max_length=30, verbose_name='countrey')),
],
options={
'db_table': 'countrey',
'ordering': ['countrey_name'],
},
),
migrations.CreateModel(
name='Districts',
fields=[
('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')),
('district_name', models.CharField(blank=True, default='Ganjam', max_length=30, null=True)),
],
options={
'db_table': 'districts',
'ordering': ['district_name'],
},
),
migrations.CreateModel(
name='States',
fields=[
('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')),
('state_name', models.CharField(blank=True, default='Odisha', max_length=30, null=True)),
('count', models.ForeignKey(default=1, on_delete=django.db.models.deletion.CASCADE, related_name='Districts', to='User_Accounts.Countrey')),
],
options={
'db_table': 'states',
'ordering': ['state_name'],
},
),
migrations.CreateModel(
name='Users',
fields=[
('password', models.CharField(max_length=128, verbose_name='password')),
('last_login', models.DateTimeField(blank=True, null=True, verbose_name='last login')),
('date_joined', models.DateTimeField(default=django.utils.timezone.now, verbose_name='date joined')),
('id', models.UUIDField(default=uuid.UUID('77bcd7ff-2296-46b4-a4b1-723b9bee9955'), primary_key=True, serialize=False)),
('username', models.CharField(blank=True, max_length=30, null=True, unique=True, verbose_name='username')),
('first_name', models.CharField(max_length=30)),
('last_name', models.CharField(max_length=30)),
('full_name', models.CharField(blank=True, max_length=60, null=True, verbose_name='fullname')),
('email', models.EmailField(max_length=254, unique=True, verbose_name='email address')),
('house_num', models.CharField(default='4/1', max_length=7, verbose_name='House Numebr/Flat Number')),
('address', models.CharField(max_length=300, verbose_name='Address')),
('pin_code', models.CharField(default='760008', max_length=6)),
('is_active', models.BooleanField(default=True)),
('is_staff', models.BooleanField(default=False)),
('is_superuser', models.BooleanField(default=False)),
('updated_at', models.DateTimeField(auto_now=True)),
('groups', models.ManyToManyField(blank=True, help_text='The groups this user belongs to. A user will get all permissions granted to each of their groups.', related_name='user_set', related_query_name='user', to='auth.Group', verbose_name='groups')),
('user_permissions', models.ManyToManyField(blank=True, help_text='Specific permissions for this user.', related_name='user_set', related_query_name='user', to='auth.Permission', verbose_name='user permissions')),
],
options={
'db_table': 'Accounts',
},
managers=[
('objects', django.contrib.auth.models.UserManager()),
],
),
migrations.AddField(
model_name='districts',
name='state',
field=models.ForeignKey(default=1, on_delete=django.db.models.deletion.CASCADE, to='User_Accounts.States'),
),
migrations.AddField(
model_name='cities',
name='dist',
field=models.ForeignKey(default=1, on_delete=django.db.models.deletion.CASCADE, to='User_Accounts.Districts'),
),
]
| [
"kumarpandamanoja@gmail.com"
] | kumarpandamanoja@gmail.com |
86673876860a16e73baeba13cf15a5f5f9a6b8f6 | ca9bf5b1d53ff7d755c53e45486238cf4c2fec43 | /src/accounts/forms.py | 3fce5be865495125379ea09990f5b3f0c5dbfb1e | [] | no_license | pagnn/Geolocator | 6aa2e75aab8395f0f22048dd844b7d69b2ac3ee7 | b45d9500667e3552acc8c851e7f8f383d3d86c9f | refs/heads/master | 2021-09-01T09:13:55.726030 | 2017-12-26T06:13:11 | 2017-12-26T06:13:11 | 110,927,555 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 262 | py | from django import forms
from django.contrib.auth.forms import AuthenticationForm
class LoginForm(AuthenticationForm):
def confirm_login_allowed(self,user):
if not user.is_active:
raise forms.ValidationError('This account is inactive',code='inactive')
| [
"sylviawei19950920@gmail.com"
] | sylviawei19950920@gmail.com |
025eb96b12404c357e6dfb6cd2f21162f9492c7b | bf014a59c19683bf6c6e0c2ec5b89a18c4305b3c | /backends/__init__.py | 6ab5708fda5ed59498f8977037d2d3a11eef6adf | [
"MIT"
] | permissive | l3robot/pytorch-a3c | 5974e2c894aff121223fdb80c77eea7bbc36313e | 0bc46ac67346d77c5bd01cff8fd98ea617efc708 | refs/heads/master | 2020-03-18T10:40:55.832062 | 2018-05-23T21:30:06 | 2018-05-23T21:30:06 | 133,974,170 | 0 | 0 | null | 2018-05-18T15:43:46 | 2018-05-18T15:43:45 | null | UTF-8 | Python | false | false | 75 | py | from .atari import create_atari_env
from .unity3d import create_unity3d_env | [
"louis-emile.robitaille@elementai.com"
] | louis-emile.robitaille@elementai.com |
27f391a18eeeb3d62288d25e1dbfff1a600a7beb | bcebaeb318059cbf8b7d2bb0e991253440be3518 | /importingEXCEL.py | 2a856236e021f8cc6988304a8248026866c52057 | [] | no_license | JadfJamal98/FEC_final_version | add4dc70658f4f427295ec54bdf021f1c4027fce | 1ac3b9c00038ea21610c372efeedb85a5a2a49b7 | refs/heads/main | 2023-08-05T16:13:45.711104 | 2021-10-11T21:22:28 | 2021-10-11T21:22:28 | 370,786,734 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 6,375 | py | from openpyxl import load_workbook
def searchcatsub(sr,D):
""" This function return a list of subcategory's name for a category or the categor's name for the topic
input:
sr: Excel Search Engine for a specific Sheet
D: the Name of the column from which we want to extract the names
Output:
list containing the list of lists containing the names of sub categories
"""
result = [] # initializing
for j in range(1,len(sr[D])+1): # looping through all rows
k = D + str(j) # creating the cell's ID
if sr[k].value is None: # if the va
continue # skips
else: # if not null
result.append(sr[k].value.lower()) # adopting a norm of lower values
return result
def clean(arr):
"""
clean array from empty '' caused by the splitting
"""
cleaned = []
for i in arr:
if len(i) == 0: # such strings have length 0
continue # so we skip them
else:
cleaned.append(i) # we append them otherwise
return cleaned
def searchvalues(sr,p,n):
"""
this function has a job of collecting the words tokenize them and add them to list, which also will be categorised
in a list according the the category/subcategory
Input:
sr: Excel Search engine given a sheet
p: the tag of the Column where there is the sub-category/category
n: the tag of the column where there is the actual words corresponding
Output:
results: list of lists, the lists inside are divided to accomodate for the number of subsequent categories/sub
"""
result , sub = [] , []
for j in range(1,len(sr[p])+1):
prev = p + str(j) # creating the cell's ID of the previous column ie. the sub/categories
act = n + str(j) # creating the cell's ID of the cell containing the values
if sr[act].value is None: # if the cell of words is empty, skip this cell
continue
if sr[prev].value is not None: # if the prev cell is not empty means we have a new sub/cat
result.append(clean(sub)) # append the previously collected , cleaned
sub = [] # reinitiate the same list
if "," in sr[act].value: # if the cell contains many words
# lower case, eliminate spaces and split at ',' and add this list to sub
sub+=sr[act].value.lower().replace(" ","").split(",")
else:
sub.append(sr[act].value.lower()) # otherwise it appends the lower case value
result.append(clean(sub)) # appending the last result as its was not appended
return result[1:] # the first list initialted is added directly so it was taken out
def importing(sr):
"""
this function collects the data collected in the previous functions and returns a dictionary with multiple layers
with keys equal to categories sub categories and if the the category has a some words its added under key : 'self'
Input:
sr: Excel Engine related to a specific excel sheet
Output:
Result: dict, containing all the words under their correct distribution in the excel file
"""
# first getting the Data
Topic_value = searchvalues(sr,'A','B') # the words associated directly with the topic
Categories_name = searchcatsub(sr,'C') # the list of categories under the topic
Categories_values = searchvalues(sr,'C','D') # the words assigned to each of these categories
Subcateg_name = searchcatsub(sr,'E') # the list of sub category under each cateogry
Subcateg_values = searchvalues(sr,'E','F') # list of words for each of these sub categories
inhertence = searchvalues(sr,'C','E') # the list of sub cateogies with respect to each category
# backward induction to build the dictionary
last_layer = {} # initializing the last layer of our dictionary
for i in range(len(Subcateg_name)): # loopoing through each subcategory name
# appending to the last layer keys being the names of subcategory and the
# value being the list of corresponding words
last_layer.update({Subcateg_name[i]:Subcateg_values[i]})
second_layer ={} #initializing the second to last layer
for j in range(len(Categories_name)): # looping through all categories name
second_layerh = {} # initalizing the hidden layer i.e. the dictionary in the dictionary
second_layerh.update({'self' : Categories_values[j]}) # appending first the own words of the category
for k in inhertence[j]: # then looping through all its inheritance ie. the subs corresponding the each category
# for each inheritant sub category, adding the name as key and the value
# as the previous layer with the same key
second_layerh.update({k : last_layer[k]})
second_layer.update({Categories_name[j]:second_layerh}) # then adding all this hidden dictionary in the second layer one
Result = {} # this is the return dictionary containint all the words neatly oraganized by category and subs
Result.update({'self':Topic_value[0]}) # the topic has its words, adding them under the key self
for l in Categories_name: # looping through all categories
# appending the dictionary with key as the cateogry and the value as the dictionary of the category
Result.update({l:second_layer[l]})
return Result
wb = load_workbook(filename="LoughranMcDonald_SentimentWordLists_2018.xlsx") # initialzing for the first Sentiment T
sheetn = wb.sheetnames[1:] # the first sheet is just information we don't need
Sentinents = {} # initiliazing the dictionar
for i in range(len(sheetn)): # looping through all sheets
listword=[] # in each sheet we redifine a new list
sr=wb[sheetn[i]] # we set the engine to work in the specific sheet
for j in range(1,sr.max_row+1): # we loop till the last row
k = 'A' + str(j) # creating the the ID of the cell
listword.append(sr[k].value.lower()) # appending its lower case value
Sentinents.update({sheetn[i]:listword}) # we append to the main dictionary the list collected, with a key equal to the name of the sheet
| [
"71495618+JadfJamal98@users.noreply.github.com"
] | 71495618+JadfJamal98@users.noreply.github.com |
d30aefdbac47c69b1651c631e5c0d110bdced301 | abdc30ddc3e2aa874afe85f3b3cf914c55ef98a4 | /RecipeForDisaster/views.py | 681a2096fc19587138817e22f8a920a6b05f2c06 | [] | no_license | dann4520/RecipeSite | 65fd1b8afbb9ef0414dd4431cc918d3eeaf34b94 | 851b49fac08d74aabe5dde88d4c56d375edf705d | refs/heads/master | 2020-04-19T12:37:50.444276 | 2019-01-29T18:22:21 | 2019-01-29T18:22:21 | 168,088,066 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 334 | py | from django.shortcuts import render
from django.utils import timezone
from .models import Recipe
# Create your views here.
def recipe_list(request):
recipes = Recipe.objects.filter(created_date__lte=timezone.now()).order_by('created_date')
return render(request, 'RecipeForDisaster/recipe_list.html', {'recipes': recipes})
| [
"stabdan@gmail.com"
] | stabdan@gmail.com |
576cdc82df4f4c2c6b43fd86bf0387f7bd3e2b14 | 54ff26f132923cf5cf6c19e56156e8b99b65c6aa | /jobsapp/migrations/0009_auto_20201108_0009.py | 1db9f61d18ca784a57804d5c939fdde783fc738a | [
"MIT"
] | permissive | sokogfb/Job-Portal | e19d2a80b82b24dfc2e305346db9fbddfe06d81d | 15c9065853920ddfe6e43a37062cc7fd32fb7f8e | refs/heads/main | 2023-01-21T06:50:48.925252 | 2020-11-09T19:25:29 | 2020-11-09T19:25:29 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 457 | py | # Generated by Django 3.0.7 on 2020-11-07 18:39
from django.db import migrations, models
class Migration(migrations.Migration):
dependencies = [
('jobsapp', '0008_auto_20200810_1925'),
]
operations = [
migrations.AlterField(
model_name='job',
name='category',
field=models.CharField(choices=[('1', 'Full time'), ('2', 'Part time'), ('3', 'Internship')], max_length=100),
),
]
| [
"54090909+rajpateln1995@users.noreply.github.com"
] | 54090909+rajpateln1995@users.noreply.github.com |
412e272a77c61d7b9f5a0c1f4eeb2a6cdd56efbc | 09f175c759b0c798c1f5605b1720f9571fb5d4aa | /app/main.py | 451951e6016c9e0931d8a05daedf417eb3c471de | [] | no_license | dparker2/internet-trends | 5fe13172fd9bdb5b7d557435e4f699245b47d7df | 1e778c4d52140d1d0f92523ef2608eb623f75bc9 | refs/heads/master | 2022-02-25T00:34:40.988473 | 2019-10-14T06:47:01 | 2019-10-14T06:47:01 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 181 | py | import falcon
from app.resources.html import HTMLResource
print("PRINTED")
app = falcon.API()
HTML_resource = HTMLResource()
app.add_route("/", HTML_resource, suffix="index")
| [
"crazdave@gmail.com"
] | crazdave@gmail.com |
0fe975cae36d9d31d2e48d6d1cdbcfdd27ce3810 | 25687385836b292ee8d92f855782f2b98cc6b500 | /operator_app/consumers.py | e9f2f2bff1b32712d7dd117bf59c2cf75f270559 | [] | no_license | Konbini-shubham/Konbini | 3b9e2f11ef198aaad3347265f3c317766d2a51f6 | 85d3c2cc75e3bc6294adf57bd77b2589bb58cf81 | refs/heads/master | 2020-05-30T08:41:07.990166 | 2016-06-01T04:56:40 | 2016-06-01T04:56:40 | 59,843,148 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 792 | py | from channels.sessions import channel_session
from channels import Group
from urllib.parse import urlparse, parse_qs
import pprint
pp = pprint.PrettyPrinter(indent=4)
@channel_session
def ws_connect(message):
query_parameters = parse_qs(message.content['query_string'])
machine_id = query_parameters['id'][0]
Group(machine_id).add(message.reply_channel)
message.channel_session['id'] = machine_id
message.reply_channel.send({'text': 'In ws_connect'})
@channel_session
def ws_receive(message):
print("In ws_receive")
group = Group(message.channel_session['id'])
message.reply_channel.send({'text': 'In ws_receive'})
group.send({
"text": message.content['text'],
})
def ws_disconnect(message):
print("In ws_disconnect")
message.reply_channel.send({'text': 'In ws_disconnect'}) | [
"shubhamjigoyal@gmail.com"
] | shubhamjigoyal@gmail.com |
fef8db32f61c0a08006394e2202189ef6d30a1d7 | bba12c5af82ea9d1f0321231bd4d33e835212128 | /redisPubsub.py | 565f0398530d4426cbc67a325ae307265ba30a0b | [] | no_license | stock-ed/material-study | c151598b2b22f34ee8fb906ac86689b264094769 | 04b5607881b44faa9d4568b412ed5249a6f170b1 | refs/heads/main | 2023-08-23T22:24:16.810727 | 2021-10-18T07:22:08 | 2021-10-18T07:22:08 | 404,207,971 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 1,777 | py | import threading
import redis
import json
from redisTSBars import RealTimeBars
from redisUtil import KeyName, RedisAccess
class RedisSubscriber(threading.Thread):
def __init__(self, channels, r=None, callback=None):
threading.Thread.__init__(self)
self.redis = RedisAccess.connection(r)
self.pubsub = self.redis.pubsub()
self.pubsub.subscribe(channels)
self.callback = callback
def get_redis(self):
return self.redis
def work(self, package):
if (self.callback == None):
print(package['channel'], ":", package['data'])
else:
data = json.loads(package['data'])
self.callback(data)
def run(self):
for package in self.pubsub.listen():
if package['data'] == "KILL":
self.pubsub.unsubscribe()
print("unsubscribed and finished")
break
elif package['type'] == 'message':
self.work(package)
else:
pass
class RedisPublisher:
def __init__(self, channels, r=None):
self.redis = RedisAccess.connection(r)
self.channels = channels
def publish(self, data):
package = json.dumps(data)
self.redis.publish(self.channels[0], package)
def killme(self):
self.redis.publish(self.channels[0], 'KILL')
class StreamBarsSubscriber(RedisSubscriber):
def __init__(self):
self.rtb = RealTimeBars()
RedisSubscriber.__init__(self,
KeyName.EVENT_BAR2DB, callback=self.rtb.redisAdd1Min)
class StreamBarsPublisher(RedisPublisher):
def __init__(self):
RedisPublisher.__init__(self, KeyName.EVENT_BAR2DB)
if __name__ == "__main__":
pass
| [
"kyoungd@hotmail.com"
] | kyoungd@hotmail.com |
11e111e1dce4624067f7d5607b2f5bc263d234b6 | 152b31f0da5899569c1e30cec9c901ff9ef0a231 | /pythonHelloWord.py | 2233b8c7bdddf33dd1af56c9defa2c7bc478b431 | [] | no_license | lindaTest01/withIgnorefile | acf80a5d91c054847220455d4888012544595980 | cd2c75065ad6cc9417db9be300ec0fd83ee524cf | refs/heads/master | 2022-11-11T07:18:20.756362 | 2020-07-06T08:51:07 | 2020-07-06T08:51:07 | 276,827,968 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 134 | py | # -*- coding: UTF-8 -*-
# Filename : helloworld.py
# author by : www.runoob.com
# 该实例输出 Hello World!
print('Hello World!')
| [
"noreply@github.com"
] | noreply@github.com |
551d0de7166a4b76fbeb42575292be18fcab560f | ba6923aa77c6abeb4428f071528c5a36a7000732 | /ChangeTheWorld/AndAgainAndAgainAndAgainAndAgainAndAgainAndAgain.py | 4b68c7de0db1b0c9188d8f77a3399c7851033194 | [] | no_license | page2me/IT-Xtoberfest2021 | 8522c81a0159dd1de377a8825c0b2f27b84d0be1 | 7b39b819d4815a4840b62a4782b3672472cf9a0a | refs/heads/main | 2023-08-27T13:50:21.420977 | 2021-10-30T18:50:06 | 2021-10-30T18:50:06 | 420,277,403 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 436 | py | """Func"""
def func(text):
"""AndAgainAndAgainAndAgainAndAgainAndAgainAndAgain"""
answer = []
for i in text:
counter = 0
counter += i.count("a") + i.count("e") + i.count("i") + i.count("o") + i.count("u")
if counter >= 2:
answer.append(i)
answer.sort()
if len(answer) == 0:
print("Nope")
else:
print(*answer, sep="\n")
func(input().replace(".", "").split())
| [
"earth_killerdark@hotmail.com"
] | earth_killerdark@hotmail.com |
1701c7e4aa7e3cded6d76cdf36ae3df50910147a | 7cbcfc334d0c99b7c2a4740de26bebce42907362 | /1.6. Input print:Hour and minutes.py | a0647d328fb61a693e8d1e043a43ec06b77e6910 | [] | no_license | YukPathy/100-days-of-code | 55603bb80a22fcfdbc1751e1bc7ee5884faa0ddc | 959ed06064c5955fb15d9ceed68c449881c59062 | refs/heads/master | 2020-07-01T14:04:34.627510 | 2019-08-06T15:29:10 | 2019-08-06T15:29:10 | 201,191,395 | 0 | 0 | null | 2019-08-08T06:20:11 | 2019-08-08T06:20:11 | null | UTF-8 | Python | false | false | 107 | py | # Read an integer:
a = int(input())
#Print a value:
h=a//3600;
m = (a//60)
# print(a)
print(str(h),str(m))
| [
"noreply@github.com"
] | noreply@github.com |
8d9e1c544ea7627b9c50bc0bde5e458f4f1713df | 4293afea6240e2216eb5ed79579de1bc999b25ff | /TextViewer/icon.py | 985e10764f34b730212cdbb9f4360ee18a5b471d | [] | no_license | zoemurmure/TextViewer | 3df1737ed3b836ff89e1d5b018dd722e097ea4fc | f9ee29cfc8d15d440a74d2a6211db37d87738ba7 | refs/heads/master | 2022-11-23T14:11:38.247111 | 2020-07-31T08:01:46 | 2020-07-31T08:01:46 | 272,883,299 | 1 | 0 | null | null | null | null | UTF-8 | Python | false | false | 22,619 | py | img=b'AAABAAEAQEAAAAEAIAAoQgAAFgAAACgAAABAAAAAgAAAAAEAIAAAAAAAAEAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAPb29iL29vaP9vb20/X19fX29vX/9vb1//b29f/29vX/9vb1//b29f/29vX/9vb1//b29f/29vX/9vb1//b29f/29vX/9vb1//b29f/29vX/9vb1//b29f/29vX/9vb1//b29f/29vX/9vb1//b29f/29vX/9vb1//b29f/29vX/9vb1//b29f/29vX/9vb1//b29f/29vX/9vb1//b29f/29vX/9vb1//b29f/29vX/9vb1//b29f/29vX/9vb1//b29f/29vX/9vb1//b29f/29vX/9vb1//b29f/29vX/9vb29fb29tP29vaP9vb2IgAAAAAAAAAAAAAAAPb29lD29vbz9fb2//Tz8//p5N//6OLe/+ji3v/o4t7/6OLe/+ji3v/o4t7/6OLe/+ji3v/o4t7/6OLe/+ji3v/o4t7/6OLe/+ji3v/o4t7/6OLe/+ji3v/o4t7/6OLe/+ji3v/o4t7/6OLe/+ji3v/o4t7/6OLe/+ji3v/o4t7/6OLe/+ji3v/o4t7/6OLe/+ji3v/o4t7/6OLe/+ji3v/o4t7/6OLe/+ji3v/o4t7/6OLe/+ji3v/o4t7/6OLe/+ji3v/o4t7/6OLe/+ji3v/o4t7/6OLe/+ji3v/o4t7/5+Le/+jk3//09PP/9fb2//b19vP29vZQAAAAAPb29iL29vbz9vb2/+/v7P+Vd1v/fFIu/31RLf99US3/fVEt/31RLf99US3/fVEt/31RLf99US3/fVEt/31RLf99US3/fVEt/31RLf99US3/fVEt/31RLf99US3/fVEt/31RLf99US3/fVEt/31RLf99US3/fVEt/31RLf99US3/fVEt/31RLf99US3/fVEt/31RLf99US3/fVEt/31RLf99US3/fVEt/31RLf99US3/fVEt/31RLf99US3/fVEt/31RLf99US3/fVEt/31RLf99US3/fVEt/31RLf99US3/fVEt/3xRLf98Ui3/l3dc//Du7P/29fb/9vb28/b29iL29vaP9vb2//b29v/LvrH/e1Eu/3tRLf98Ui3/fFIt/3xSLf98Ui3/fFIt/3xSLf98Ui3/fFIt/3xSLf98Ui3/fFIt/3xSLf98Ui3/fFIt/3xSLf98Ui3/fFIt/3xSLf98Ui3/fFIt/3xSLf98Ui3/fFIt/3xSLf98Ui3/fFIt/3xSLf98Ui3/fFIt/3xSLf98Ui3/fFIt/3xSLf98Ui3/fFIt/3xSLf98Ui3/fFIt/3xSLf98Ui3/fFIt/3xSLf98Ui3/fFIt/3xSLf98Ui3/fFIt/3xSLf98Ui3/fFIt/3xSLf98Ui3/fFEs/3xSLf/MwLH/9vb2//b29v/29vaP9vb20/b29v/29vb/29PM/3xTL/97US3/e1Et/3tRLf97US3/e1Et/3tRLf97US3/e1Et/3tRLf97US3/e1Et/3tRLf97US3/e1Et/3tRLf97US3/e1Et/3tRLf97US3/e1Et/3tRLf97US3/e1Et/3tRLf97Uiz/fFEt/3xSLf98Ui3/fFIt/3tRLP98Ui3/e1Et/3tRLf97US3/e1Et/3tRLf97US3/e1Et/3tRLf97US3/e1Et/3tRLf97US3/e1Et/3tRLf97US3/e1Et/3tRLf97US3/e1Et/3tRLf97US3/fFIt/3xRLf99Uy7/3NPL//b29v/29vb/9vb20/b29vX29vb/9vb2//X19f/OwbT/qY54/6eLdf+ninX/p4p1/6eKdf+ninX/p4p1/6eKdf+ninX/p4p1/6eKdf+ninX/p4p1/6eKdf+ninX/p4p1/6eKdf+ninX/p4p1/6eKdf+ninX/p4p1/6eKdf+ni3T/pIhu/3xSLf98Ui3/fFIt/3xSLf+kh27/p4t0/6eKdf+ninX/p4p1/6eKdf+ninX/p4p1/6eKdf+ninX/p4p1/6eKdf+ninX/p4p1/6eKdf+ninX/p4p1/6eKdf+ninX/p4p1/6eKdf+ninX/p4p1/6eLdf+ojnj/z8G2//T19f/19fX/9vb2//X19fX29vb/9vb2//b29v/29vb/9vb2//X19f/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2/8m7rf98Ui3/fFIt/3xSLf98US3/yLut//X29f/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/19fX/9fX1//X19f/19fX/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b19f/19fX/9fb2/+jj3f+IY0P/e1Et/3xSLf98Ui3/fFIt/4hjQ//m4t3/9vX1//X19f/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//X19f/29vX/9fX1//X19f/19fX/9fX1//X19f/19fX/9fX1//X19f/19fX/9fX1//X19f/19fX/9fX1//X19f/19fX/9fX1//X19f/19fX/9fX1//X19f/19fX/9fX1//X19f/19fX/8O/t/8q9rv+HY0H/fFIt/3xSLf98Ui3/fFIt/3tRLf97US7/iGJB/8i7rf/x7+7/9fX1//X19f/19fX/9fX1//X19f/19fX/9fX1//X19f/19fX/9fX1//X19f/19fX/9fX1//X19f/19fX/9fX1//X19f/19fX/9fX1//X19f/19fX/9fX1//X19f/19fX/9fX1//X19f/Syr7/lHJU/41lRf+NZUb/jWVG/41lRv+NZUb/jWVG/41lRv+NZUb/jWVG/41lRv+NZUb/jWVG/41lRv+NZUb/jWVG/41lRv+NZUb/jWVG/41lRv+NZUb/jWVG/41lRv+MZUX/imRD/39YNf96US3/fFIu/3xSLv98Ui3/fFIt/3xSLf98Ui3/e1Et/3tRLf97US3/f1g2/4tkQ/+NZUX/jWVG/41lRv+NZUb/jWVG/41lRv+NZUb/jWVG/41lRv+NZUb/jWVG/41lRv+NZUb/jWVG/41lRv+NZUb/jWVG/41lRv+NZUb/jWVG/41lRv+NZUb/i2VF/5RyVP/Vyr7/h2JA/3tSLP98US7/fFEt/3xRLf98US3/fFEt/3xRLf98US3/fFEt/3xRLf98US3/fFEt/3xRLf98US3/fFEt/3xRLf98US3/fFEt/3xRLf98US3/fFEt/3xRLf98US3/fFEt/3xSLv98Ui3/fFIt/3xSLf98US3/e1Es/3xSLf98US3/e1Is/3xSLf98Ui3/fFIt/3xSLv98Ui3/fFEu/3xRLf98US3/fFEt/3xRLf98US3/fFEt/3xRLf98US3/fFEt/3xRLf98US3/fFEt/3xRLf98US3/fFEt/3xRLf98US3/fFEt/3xRLf98US3/fFEt/3xRLf98US3/h2JB/3tRLP97USz/fFIt/3tRLf98Ui3/fFIs/3xSLP98Uiz/fFIs/3xSLP98Uiz/fFIs/3xSLP98Uiz/fFIs/3xSLP98Uiz/fFIs/3xSLP98Uiz/fFIs/3xSLP98Uiz/fFIs/3xSLP98Ui3/fFEt/3tSLP98Ui3/kGpI/4FZNf98Ui3/fFEt/4BaNf+ObU7/fFIs/3xSLf97US3/fFIt/3xSLf98Ui3/fFIt/3xSLf98Ui3/fFIt/3xSLf98Ui3/fFIt/3xSLf98Ui3/fFIt/3xSLf98Ui3/fFIt/3xSLf98Ui3/fFIt/3xSLf98Ui3/fFIt/3xSLf97US3/fFIt/3tRLf98Ui3/fFIt/3xSLf+AVjL/iGI//4dhP/+HYT//h2E//4dhP/+HYT//h2E//4dhP/+HYT//h2E//4dhP/+HYT//h2E//4dhP/+HYT//h2E//4dhP/+HYT//h2E//4dhP/+IYT//iWJA/5NuS/+piGj/0bic//Tkyf+JYj//fFIt/3tSLf+IZEP/8fHw/83Btv+ojXb/kXBR/4hlQ/+IY0H/iGNB/4hjQf+IY0H/iGNB/4hjQf+IY0H/iGNB/4hjQf+IY0H/iGNB/4hjQf+IY0H/iGNB/4hjQf+IY0H/iGNB/4hjQf+IY0H/iGNB/4hjQv9/VTL/e1Et/3xSLf98Ui3/fFIt/3xSLf98Ui7/mHRS//jnzv/4587/+OfO//jnzv/4587/+OfO//jnzv/4587/+OfO//jnzv/4587/+OfO//jnzv/4587/+OfO//jnzv/4587/+OfO//jnzv/4587/+efO//nnzv/5587/+ubO//vm0P/55s7/iWJA/3xSLf97Ui3/iGVC//T19P/29vb/9fX1//b29f/19fX/9fX1//X19f/19fX/9fX1//X19f/19fX/9fX1//X19f/19fX/9fX1//X19f/19fX/9fX1//X19f/19fX/9fX1//X19f/19fX/9fX1//X19f/19fb/mXhc/3tRLP98Ui3/fFIt/3xSLf98Ui3/fFIu/5h0U//65s//++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/+ebO/4liP/98Ui3/e1It/4hlQv/19fX/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9PX2/5l6XP98Uiz/fFIt/3xSLf98Ui3/fFIt/3xSLv+YdFP/+ubP//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//nmzv+JYj//fFIt/3tSLf+IZUL/9fX1//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//T19v+Zelz/fFIs/3xSLf98Ui3/fFIt/3xSLf98Ui7/mHRT//rmz//75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/55s7/iWI//3xSLf97Ui3/iGVC//X19f/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/09fb/mXpc/3xSLP98Ui3/fFIt/3xSLf98Ui3/fFIu/5h0U//65s//++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/+ebO/4liP/98Ui3/e1It/4hlQv/19fX/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9PX2/5l6XP98Uiz/fFIt/3xSLf98Ui3/fFIt/3xSLv+YdFP/+ubP//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//nmzv+JYj//fFIt/3tSLf+IZUL/9fX1//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//T19v+Zelz/fFIs/3tRLf98US3/fFIt/3xSLf98Ui7/mHRT//rmz//75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/55s7/iWI//3xSLf97Ui3/iGVC//X19f/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29fb/nYBm/3xSLP96Uiz/glg2/3xSLf98Ui3/fFIu/5h0U//65s//++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/+ebO/4liP/98Ui3/e1It/4hlQv/19fX/9vb2//b29v/29vb/9vb2//X19f/19fX/9fX1//X19f/19fX/9fX1//X19f/19fX/9fX1//X19f/19fX/9fX1//X19f/19fX/9fX1//X19f/29vb/9vb2//b29v/29vb/9vb2/9LGvf+BWjb/flMw/7qmlv98Ui3/fFIt/3xSLv+YdFP/+ubP//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//nmzv+JYj//fFIt/3tSLf+IZUL/9fX1//b29v/29vb/9vb2//b29v/19fX/9fX1//X19f/19fX/9fX1//X19f/19fX/9fX1//X19f/19fX/9fX1//X19f/19fX/9fX1//X19f/19fX/9vb2//b29v/29vb/9vb2//b29v/29vb/8O3q/+vo5P/19fX/fFIt/3xSLf98Ui7/mHRT//rmz//75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/55s7/iWI//3xSLf97Ui3/iGVC//X19f/29vb/9vb2//b29v/z8vH/sZyI/5JwU/+Rb1H/kW9R/5FvUf+Rb1H/kW9R/5FvUf+Rb1H/kW9R/5FvUf+Rb1H/kW9R/5FvUf+Rb1L/rJJ8//Du7P/19fb/9vb2//b29v/29vb/9vb2//b29f/19vX/9fX1/3xSLf98Ui3/fFIu/5h0U//65s//++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/+ebO/4liP/98Ui3/e1It/4hlQv/19fX/9vb2//b29v/19fb/y7uu/3tSLP98Uiz/fFIs/3xSLP98Uiz/fFIs/3xSLP98Uiz/fFIs/3xSLP98Uiz/fFIs/3xSLP98Uiz/fFIt/3tRLf+6qJb/9fX2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v98Ui3/fFIt/3xSLv+YdFP/+ubP//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//nmzv+JYj//fFIt/3tSLf+IZUL/9fX1//b29v/29vb/9fX2/8O2pf98US7/e1Et/3tRLv97US7/e1Eu/3tRLv97US7/e1Eu/3tRLv97US7/e1Eu/3tRLv97US7/e1Eu/3tRLf97US7/t6KP//X29f/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/fFIt/3xSLf98Ui7/mHRT//rmz//75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/55s7/iWI//3xSLf97Ui3/iGVC//X19f/29vb/9vb2//b19v/x7+3/o4lx/4NdOv+CXDj/gls4/4JbOP+CWzj/gls4/4JbOP+CWzj/gls4/4JbOP+CWzj/gls4/4FbOP+DXDr/nH1i/+vo5f/29vb/9vb2//b29v/29vb/9vb2//X19f/29vb/9vb2/3xSLf98Ui3/fFIu/5h0U//65s//++bQ//vm0P/75tD/+ObQ//fm0P/65tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//rlz//75tD/++bQ//vm0P/75tD/+ebO/4liP/98Ui3/e1It/4hlQv/19fX/9vb2//b29v/19fX/9vX2//b29v/29fX/9fX0//X19P/19fT/9fX0//X19P/19fT/9fX0//X19P/19fT/9fX0//X19P/19fT/9fX0//b29f/29vX/9vb2//b29v/29vb/9vb2//X19v/r5uP/5N/a//X19v98Ui3/fFIt/3xSLv+YdFP/+ubP//vm0P/75tD/++bQ//Xly/+8ooL/pYNi/6aCYv+mgmL/poJi/6aCYv+mgmL/poJi/6aCYv+mgmL/poJi/6aCYv+mgmL/poJi/6eEY//Gq43/+ObN//jmz//75tD/++bQ//nmzv+JYj//fFIt/3tSLf+IZUL/9fX1//b29v/29vb/9vb2//b29v/19fX/9fX1//X19f/19fX/9fX1//X19f/19fX/9fX1//X19f/19fX/9fX1//X19f/19fX/9fX1//X19f/29vb/9vb2//b29v/29vb/9vb2//X19f/PwrP/f1Y0/3xTLf+0nor/fFIt/3xSLf98Ui7/mHRT//rmz//75tD/++bQ//jmz//DqYv/fFEt/3tRLP97US3/e1It/3tSLf97Ui3/e1It/3tSLf97Ui3/e1It/3tSLf97Ui3/e1It/3tSLf97USz/fFEt/9G6nv/4587/++bQ//vm0P/55s7/iWI//3xSLf97Ui3/iGVC//X19f/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/19fb/nX5j/3xSLf98Ui3/glc1/3xSLf98Ui3/fFIu/5h0U//65s//++bQ//vm0P/55tD/tZZ4/3tSLv98Ui3/fFIt/3xSLf98Ui3/fFIt/3xSLf98Ui3/fFIt/3xSLf98Ui3/fFIt/3xSLf98Ui3/e1Es/3xSLP/DqYz/+ObP//vm0P/75tD/+ebO/4liP/98Ui3/e1It/4hlQv/19fX/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9PX2/5l5XP98Uiz/fFIt/3xRLP98Ui3/fFIt/3xSLv+YdFP/+ubP//vm0P/75tD/+ubQ/+bSuP+MZUP/e1Mt/3tSLf97Ui3/e1It/3tSLf97Ui3/e1It/3tSLf97Ui3/e1It/3tSLf97Ui3/e1It/3xSLv+Sb03/7t3B//jnzv/75tD/++bQ//nmzv+JYj//fFIt/3tSLf+IZUL/9fX1//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//T19v+Zelz/fFIs/3xSLf98Ui3/fFIt/3xSLf98Ui7/mHRT//rmz//75tD/++bQ//vm0P/65s//9ubL/+zbvv/r2r3/69q9/+vavf/r2r3/69q9/+vavf/r2r3/69q9/+vavf/r2r3/69q9/+vavv/t28D/9+fN//rmz//65s//++bQ//vm0P/55s7/iWI//3xSLf97Ui3/iGVC//X19f/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/09fb/mXpc/3xSLP98Ui3/fFIt/3xSLf98Ui3/fFIu/5h0U//65s//++bQ//vm0P/75tD/++bQ//rmz//55s//+ebP//nm0P/55tD/+ebQ//nm0P/55tD/+ebQ//nm0P/55tD/+ebQ//nm0P/65s//+ubP//rmz//75tD/++bQ//vm0P/75tD/+ebO/4liP/98Ui3/e1It/4hlQv/19fX/9vb2//b29v/29vb/9vb2//b29v/29vb/9fX1//X19f/19fX/9fX1//X19f/19fX/9fX1//X19f/19fX/9fX1//X19f/19fX/9vb2//X19f/29vb/9vb2//b29v/29vb/9PX2/5l6XP98Uiz/fFIt/3xSLf98Ui3/fFIt/3xSLv+YdFP/+ubP//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//nmzv+JYj//fFIt/3tSLf+IZUL/9fX1//b29v/29vb/9vb2//b19f/Pwrb/sJiC/6+Xgf+vl4H/r5eB/6+Xgf+vl4H/r5eB/6+Xgf+vl4H/r5eB/6+Xgf+vl4H/r5eB/6+Xgv/Ht6r/9fX1//b29v/29vb/9vb2//T19v+Zelz/fFIs/3xSLf98Ui3/fFIt/3xSLf98Ui7/mHRT//rmz//75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/55s7/iWI//3xSLf97Ui3/iGVC//X19f/29vb/9vb2//b19f/Syr//fFIv/3xRLf98Ui3/fFIt/3xSLf98Ui3/fFIt/3xSLf98Ui3/fFIt/3xSLf98Ui3/fFIt/3xRLf98Uiz/e1Et/8e4qv/19fb/9vb2//b29v/09fb/mXpc/3xSLP98Ui3/fFIt/3xSLf98Ui3/fFIu/5h0U//65s//++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/+ebO/4liP/98Ui3/e1It/4hlQv/19fX/9vb2//b29v/29vX/v66d/3xRLv98Ui3/fFIt/3xSLf98Ui3/fFIt/3xSLf98Ui3/fFIt/3xSLf98Ui3/fFIt/3xSLf98Ui3/fFIt/3xSLf+xm4b/9fX2//b29v/29vb/9PX2/5l6XP98Uiz/fFIt/3xSLf98Ui3/fFIt/3xSLv+YdFP/+ubP//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//nmzv+JYj//fFIt/3tSLf+IZUL/9fX1//b29v/29vb/9vX1/+jk4f+KZkj/e1Is/3xSLf97Uiz/e1Is/3tSLP97Uiz/e1Is/3tSLP97Uiz/e1Is/3tSLP97Uiz/fFIt/3xSLf+GXj3/4NjS//X29f/29vb/9vb2//T19v+Zelz/fFIs/3xSLf98Ui3/fFIt/3xSLf98Ui7/mHRT//rmz//75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/55s7/iWI//3xSLf97Ui3/iGVC//X19f/29vb/9vb2//b29v/19fX/8vLw/+Hb1f/f2NL/39jR/9/Y0f/f2NH/39jR/9/Y0f/f2NH/39jR/9/Y0f/f2NH/39jR/9/Y0v/f2tP/8O/s//X19f/19vX/9vb2//b29v/09fb/mXpc/3xSLP98Ui3/fFIt/3xSLf98Ui3/fFIu/5h0U//65s//++bQ//vm0P/75tD/+ubO//vmz//6587/+ufO//rnzv/6587/+ufO//rnzv/6587/+ufO//rnzv/6587/+ufO//rnzv/6587/+efN//nnzv/75s7/++bQ//vm0P/75tD/+ebO/4liP/98Ui3/e1It/4hlQv/19fX/9vb2//b29v/29vb/9vb2//X19f/19vb/9fb2//X29v/19vb/9fb2//X29v/19vb/9fb2//X29v/19vb/9fb2//X29v/19vb/9vb2//X19f/29vb/9vb2//b29v/29vb/9PX2/5l6XP98Uiz/fFIt/3xSLf98Ui3/fFIt/3xSLv+YdFP/+ubP//vm0P/75tD/++bQ//nmzf/Tu5//up5+/7mefv+5nX7/uZ1+/7mdfv+5nX7/uZ1+/7mdfv+5nX7/uZ1+/7mdfv+5nX7/uZ19/7qff//axab/+ufO//vm0P/75tD/++bQ//nmzv+JYj//fFIt/3tSLf+IZUL/9fX1//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//T19v+Zelz/fFIs/3xSLf98Ui3/fFIt/3xSLf98Ui7/mHRT//rmz//75tD/++bQ//vm0P/NtJn/e1It/3tSLf97Ui3/e1It/3tSLf97Ui3/e1It/3tSLf97Ui3/e1It/3tSLf97Ui3/e1It/3tSLf98Ui3/fVMv/9vGqP/55s7/++bQ//vm0P/55s7/iWI//3xSLf97Ui3/iGVC//X19f/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/09fb/mXpc/3xSLP98Ui3/fFIt/3xSLf98Ui3/fFIu/5h0U//65s//++bQ//vm0P/65c//spJ0/3tRLf97USz/fFIt/3xSLf98Ui3/fFIt/3xSLf98Ui3/fFIt/3xSLf98Ui3/fFIt/3xSLf98Ui3/fFIt/3tSLf/Bpoj/+OfO//vm0P/75tD/+ebO/4liP/98Ui3/e1It/4hlQv/19fX/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9PX2/5l6XP98Uiz/fFIt/3xSLf98Ui3/fFIt/3xSLv+YdFP/+ubP//vm0P/75tD/++bO/93Hqv+CWDP/fFEt/31SLf99Ui3/fVIt/31SLf99Ui3/fVIt/31SLf99Ui3/fVIt/31SLf99Ui3/fVIt/3xSLf+FXjz/59K5//nm0P/75tD/++bQ//nmzv+JYj//fFIt/3tSLf+IZUL/9fX1//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//T19v+Zelz/fFIs/3xSLf98Ui3/fFIt/3xSLf98Ui7/mHRT//rmz//75tD/++bQ//vnzv/55s7/7Ni//9jApf/YwKT/2L+k/9i/pP/Yv6T/2L+k/9i/pP/Yv6T/2L+k/9i/pP/Yv6T/2L+k/9jApP/ZwaT/8N3C//nnzv/75tD/++bQ//vm0P/55s7/iWI//3xSLf97Ui3/iGVC//X19f/29vb/9vb2//b29v/19fX/9fX1//b29v/19fX/9fX1//X19f/19fX/9fX1//X19f/19fX/9fX1//X19f/19fX/9fX1//b29v/19fX/9fb1//b29v/29vb/9vb2//b29v/09fb/mXpc/3xSLP98Ui3/fFIt/3xSLf98Ui3/fFIu/5h0U//65s//++bQ//vm0P/75tD/++bQ//jmz//6587/++fO//vnzv/7587/++fO//vnzv/7587/++fO//vnzv/7587/++fO//vnzv/7587/++fO//rnzv/55s//++bQ//vm0P/75tD/+ebO/4liP/98Ui3/e1It/4hlQv/19fX/9vb2//b29v/19fb/9PX2/+Tg2v/KvLD/yruu/8q7rv/Ku67/yruu/8q7rv/Ku67/yruu/8q7rv/Ku67/yruu/8q7rv/Lu6//yruv/+Lb1P/29vX/9fX1//b29v/29vb/9PX2/5l6XP98Uiz/fFIt/3xSLf98Ui3/fFIt/3xSLv+YdFP/+ubP//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//nmzv+JYj//fFIt/3tSLf+IZUL/9fX1//b29v/29vb/9vb1/+DZ0v9/WTf/fFIt/3xSLP98Uiz/fFIs/3xSLP98Uiz/fFIs/3xSLP98Uiz/fFIs/3xSLP98Uiz/fFIt/3tSLf9/VTH/08rA//b29v/29vb/9vb2//T19v+Zelz/fFIs/3xSLf98Ui3/fFIt/3xSLf98Ui7/mHRT//rmz//75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/55s7/iWI//3xSLf97Ui3/iGVC//X19f/29vb/9vb2//X29f++rZ3/e1Eu/3xSLf98Ui3/fFIt/3xSLf98Ui3/fFIt/3xSLf98Ui3/fFIt/3xSLf98Ui3/fFIt/3xSLf98Ui3/fFEt/6+YhP/29vb/9vb2//b29v/09fb/mXpc/3xSLP98Ui3/fFIt/3xSLf98Ui3/fFIu/5h0U//65s//++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/+ebO/4liP/98Ui3/e1It/4hlQv/19fX/9vb2//b29v/19fb/3tbO/39XNf98US3/fFEs/3xRLP98USz/fFEs/3xRLP98USz/fFEs/3xRLP98USz/fFEs/3xRLP98USz/fFIs/31TLv/Qxbv/9vb2//b29v/29vb/9PX2/5l6XP98Uiz/fFIt/3xSLf98Ui3/fFIt/3xSLv+YdFP/+ubP//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//nmzv+JYj//fFIt/3tSLf+IZUL/9fX1//b29v/29vb/9vb2//X19f/g2tP/xbSk/8SypP/EsqP/xLKj/8Syo//EsqP/xLKj/8Syo//EsqP/xLKj/8Syo//EsqP/xLKj/8SypP/c08v/9vb2//b29v/29vb/9vb2//T19v+Zelz/fFIs/3xSLf98Ui3/fFIt/3xSLf98Ui7/mHRT//rmz//75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/55s7/iWI//3xSLf97Ui3/iGVC//X19f/29vb/9vb2//b29v/19fX/9fX1//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/09fb/mXpc/3xSLP98Ui3/fFIt/3xSLf98Ui3/fFIu/5h0U//65s//++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/+ebO/4liP/98Ui3/e1It/4hlQv/19fX/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9PX2/5l6XP98Uiz/fFIt/3xSLf98Ui3/fFIt/3xSLv+YdFP/+ubP//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//nmzv+JYj//fFIt/3tSLf+IZUL/9fX1//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//T19v+Zelz/fFIs/3xSLf98Ui3/fFIt/3xSLf98Ui7/mHRT//rmz//75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/55s7/iWI//3xSLf97Ui3/iGVD//X19f/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/09fb/mXpc/3xSLP98Ui3/fFIt/3xSLf98Ui3/fFIu/5h0U//65s//++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/+ubO/4liP/98Ui3/e1It/4djQ//09fX/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9PX2/5l6XP98Uiz/fFIt/3xSLf98Ui3/fFIt/3xSLv+YdFP/+ubP//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/+ubQ//bjy/+DWTb/fFIt/3xSLP+BWjj/8fHw//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//T19v+Zelz/fFIs/3xSLf98Ui3/fFIt/3xSLf98Ui7/mHRT//rmz//75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/+uXP//nmz//VwKX/e1Is/3xSLf98Ui3/e1Eu/9TJwP/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/09fb/mXpc/3xSLP98Ui3/fFIt/3xSLf98Ui3/fFIu/5h0U//65s//++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/++bQ//vm0P/75tD/+ubQ//nm0P/y4cj/k3BO/3xSLf98Ui3/fFIt/3xSLP+Tc1X/8e/t//b19f/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9PX2/5l6XP98Uiz/fFIt/3xSLf98Ui3/fFIt/3xSLv+YdFP/+ubP//rlz//65c//+uXP//rlz//65c//+uXP//rlz//65c//+uXP//rlz//65c//+uXP//rlz//65c//+uXP//rlz//65c//+uXP//rlz//65c//+uXP//nmz//l0rf/mHZV/3tSLf98Ui3/fFIt/3xSLf98Ui3/e1It/5p5Xv/j3tj/9fb1//b19f/19vX/9fb1//X29f/19vX/9fb1//X29f/19vX/9fb1//X29f/19vX/9fb1//X29f/19vX/9fb1//X29f/19vX/9fb1//X29f/19vX/9fb1//X19v+Zelz/fFIt/3xSLf98Ui3/fFIs9XxSLf98Ui3/hl86/6iHZv+oh2b/qIdm/6iHZv+oh2b/qIdm/6iHZv+oh2b/qIdm/6iHZv+oh2b/qIdm/6iHZv+oh2b/qIdm/6iHZv+oh2b/qIdm/6iHZv+oh2b/qIdm/6WFY/+XclD/fVQw/3tRLv98Ui3/fFIt/3tRLf97Ui3/fFIt/3xSLf98USz/fVQw/5V1Wf+minH/qItz/6eLc/+ni3P/p4tz/6eLc/+ni3P/p4tz/6eLc/+ni3P/p4tz/6eLc/+ni3P/p4tz/6eLc/+ni3P/p4tz/6eLc/+ni3P/p4tz/6eLc/+ni3T/hl8//3tRLf98Ui3/fFIt9XxSLNN7US3/fFIt/3xSLf97US7/fFIu/3xSLv98Ui7/fFIu/3xSLv98Ui7/fFIu/3xSLv98Ui7/fFIu/3xSLv98Ui7/fFIu/3xSLv98Ui7/fFIu/3xSLv98Ui7/fFIu/3xSLv97US7/fFIu/3xSLv97Ui3/fFIt/3tSLf+xmYT/sJmF/3tRLP98Ui3/fFIt/3tRLP97USz/fFIt/3xSLf98Ui3/fFIt/3xSLf98Ui3/fFIt/3xSLf98Ui3/fFIt/3xSLf98Ui3/fFIt/3xSLf98Ui3/fFIt/3xSLf98Ui3/fFIt/3xSLf98Ui3/fFIt/3xSLf98Ui3/e1It/3xSLNN8UiyPfFIt/3tRLf97US3/e1Et/3tRLf97US3/e1Et/3tRLf97US3/e1Et/3tRLf97US3/e1Et/3tRLf97US3/e1Et/3tRLf97US3/e1Et/3tRLf97US3/e1Et/3tRLf97US3/e1Et/3xRLf98USz/e1Iu/39YNf++rJr/9PX0//b19f++rJv/f1g1/3tSLv98USz/e1Es/3xSLf97US3/e1Et/3tRLf97US3/e1Et/3tRLf97US3/e1Et/3tRLf97US3/e1Et/3tRLf97US3/e1Et/3tRLf97US3/e1Et/3tRLf97US3/e1Et/3tRLf97US3/e1Et/3tSLf9+UyuPfVAoInxTMPN7US7/e1Eu/3tRLv97US7/e1Eu/3tRLv97US7/e1Eu/3tRLv97US7/e1Eu/3tRLv97US7/e1Eu/3tRLv97US7/e1Eu/3tRLv97US7/e1Eu/3tRLv97US7/e1Eu/3xSLv99Uy//jWdI/7Kahv/k39v/9fX1//b29v/19fX/9fX2/+Tg2v+xmYb/jWdI/3xTL/97US//e1Eu/3tRLv97US7/e1Eu/3tRLv97US7/e1Eu/3tRLv97US7/e1Eu/3tRLv97US7/e1Eu/3tRLv97US7/e1Eu/3tRLv97US7/e1Eu/3tRLv97US7/e1Eu/3xRLv99Uy7zfVIrIgAAAADc0sxQ5uHd8+fi3v/n4t7/5+Le/+fi3v/n4t7/5+Le/+fi3v/n4t7/5+Le/+fi3v/n4t7/5+Le/+fi3v/n4t7/5+Le/+fi3v/n4t7/5+Le/+fi3v/n4t7/5+Le/+fi3v/o497/7+zq//X19f/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//X19f/u6+r/6OPe/+fi3v/n4t7/5+Le/+fi3v/n4t7/5+Le/+fi3v/n4t7/5+Le/+fi3v/n4t7/5+Le/+fi3v/n4t7/5+Le/+fi3v/n4t7/5+Le/+fi3v/n4t7/5+Le/+fi3v/n4d3z29TLUAAAAAAAAAAAAAAAAPb29iL29vaP9vb20/b29vX29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb2//b29v/29vb/9vb29fb29tP29vaP9vb2IgAAAAAAAAAA4AAAAAAAAAfAAAAAAAAAA4AAAAAAAAABAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAACAAAAAAAAAAcAAAAAAAAAD4AAAAAAAAAc=' | [
"zoemurmure@gmail.com"
] | zoemurmure@gmail.com |
231208107ccc0ef9cd28b169c0c9c14596a48d32 | f93f0936ce11e0e4f7fbda66c6921a3fdc481e10 | /scrapy_aishanghai/aishanghai/aishanghai/middlewares.py | 8ad154919e9f8ccbfd798039d80d90046a13a393 | [] | no_license | leiyanhui/leiyh_projects | 11c827ea4b68040ba73022ade6b47da08bd7161a | 8e33be618f5078d08dec92475bf70ba1fb94ff67 | refs/heads/master | 2020-03-23T22:19:53.547259 | 2018-09-11T12:32:39 | 2018-09-11T12:32:39 | 142,169,163 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 3,605 | py | # -*- coding: utf-8 -*-
# Define here the models for your spider middleware
#
# See documentation in:
# https://doc.scrapy.org/en/latest/topics/spider-middleware.html
from scrapy import signals
class AishanghaiSpiderMiddleware(object):
# Not all methods need to be defined. If a method is not defined,
# scrapy acts as if the spider middleware does not modify the
# passed objects.
@classmethod
def from_crawler(cls, crawler):
# This method is used by Scrapy to create your spiders.
s = cls()
crawler.signals.connect(s.spider_opened, signal=signals.spider_opened)
return s
def process_spider_input(self, response, spider):
# Called for each response that goes through the spider
# middleware and into the spider.
# Should return None or raise an exception.
return None
def process_spider_output(self, response, result, spider):
# Called with the results returned from the Spider, after
# it has processed the response.
# Must return an iterable of Request, dict or Item objects.
for i in result:
yield i
def process_spider_exception(self, response, exception, spider):
# Called when a spider or process_spider_input() method
# (from other spider middleware) raises an exception.
# Should return either None or an iterable of Response, dict
# or Item objects.
pass
def process_start_requests(self, start_requests, spider):
# Called with the start requests of the spider, and works
# similarly to the process_spider_output() method, except
# that it doesn’t have a response associated.
# Must return only requests (not items).
for r in start_requests:
yield r
def spider_opened(self, spider):
spider.logger.info('Spider opened: %s' % spider.name)
class AishanghaiDownloaderMiddleware(object):
# Not all methods need to be defined. If a method is not defined,
# scrapy acts as if the downloader middleware does not modify the
# passed objects.
@classmethod
def from_crawler(cls, crawler):
# This method is used by Scrapy to create your spiders.
s = cls()
crawler.signals.connect(s.spider_opened, signal=signals.spider_opened)
return s
def process_request(self, request, spider):
# Called for each request that goes through the downloader
# middleware.
# Must either:
# - return None: continue processing this request
# - or return a Response object
# - or return a Request object
# - or raise IgnoreRequest: process_exception() methods of
# installed downloader middleware will be called
return None
def process_response(self, request, response, spider):
# Called with the response returned from the downloader.
# Must either;
# - return a Response object
# - return a Request object
# - or raise IgnoreRequest
return response
def process_exception(self, request, exception, spider):
# Called when a download handler or a process_request()
# (from other downloader middleware) raises an exception.
# Must either:
# - return None: continue processing this exception
# - return a Response object: stops process_exception() chain
# - return a Request object: stops process_exception() chain
pass
def spider_opened(self, spider):
spider.logger.info('Spider opened: %s' % spider.name)
| [
"18817380161@163.com"
] | 18817380161@163.com |
04bb08d4b13fe38a056228962344ffdfb9bf975a | 92e09d003c43662f8452f8445fdc793b60406670 | /Python/python 爬虫/Maoyantop100/spider.py | 634c40a7994be216c33dcb259f6c60942d2ac138 | [] | no_license | WangJian1314/python_spider | 63f9bd98c8f6618aeb3b6f51e6a49563cb3e6066 | 6d0394b560a556df8dca1388dfa6182475ef885b | refs/heads/master | 2020-06-02T22:50:36.577067 | 2019-04-20T09:02:51 | 2019-04-20T09:02:51 | 191,334,030 | 1 | 0 | null | 2019-06-11T09:03:54 | 2019-06-11T09:03:53 | null | UTF-8 | Python | false | false | 1,578 | py | import json
from multiprocessing import Pool
import requests
import re
from requests.exceptions import RequestException
def get_one_page(url):
try:
headers = {
'User-Agent': 'Mozilla/5.0 (Macintosh; Intel Mac OS X 10_13_3) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/65.0.3325.162 Safari/537.36'
}
response = requests.get(url, headers=headers)
if response.status_code == 200:
return response.text
return None
except RequestException:
return None
def parse_one_page(html):
pattern = re.compile('<dd>.*?board-index.*?>(\d+)</i>.*?data-src="(.*?)".*?name"><a'
+ '.*?>(.*?)</a>.*?star">(.*?)</p>.*?releasetime">(.*?)</p>'
+ '.*?integer">(.*?)</i>.*?fraction">(.*?)</i>.*?</dd>', re.S)
items = re.findall(pattern, html)
for items in items:
yield {
'index': items[0],
'image': items[1],
'title': items[2],
'actor': items[3].strip()[3:],
'time': items[4].strip()[5:],
'score': items[5]+items[6]
}
def write_to_file(content):
with open('result.txt', 'a', encoding='utf-8') as f:
f.write(json.dumps(content, ensure_ascii=False) + '\n')
f.close()
def main(offset):
url = 'http://maoyan.com/board/4?offset=' + str(offset)
html = get_one_page(url)
for item in parse_one_page(html):
print(item)
write_to_file(item)
if __name__ == '__main__':
pool = Pool()
pool.map(main, [i*10 for i in range(10)])
| [
"303061411@qq.com"
] | 303061411@qq.com |
4e58c678d8ebf19f2a3b0ab83528aa49d5f251d1 | 4d0472bcb230cf060e31e0d5320f3de2f5e9e55e | /train_generator.py | fd3098f5830fbb894befa01558bb81b18581e701 | [] | no_license | sharadmv/CharacterGAN | 7b9d37d8515978655084ea815b714596e3d5adc7 | 4a0253102f23313a5ee72a09f938c64a220fe6b8 | refs/heads/master | 2021-01-15T21:14:26.016372 | 2016-03-01T02:06:08 | 2016-03-01T02:06:08 | 52,777,523 | 0 | 0 | null | 2016-02-29T08:57:04 | 2016-02-29T08:57:03 | null | UTF-8 | Python | false | false | 5,040 | py | import numpy as np
import cPickle as pickle
import theano
import sys
import csv
import logging
import random
from dataset import *
from deepx.nn import *
from deepx.rnn import *
from deepx.loss import *
from deepx.optimize import *
from argparse import ArgumentParser
theano.config.on_unused_input = 'ignore'
logging.basicConfig(level=logging.DEBUG)
def parse_args():
argparser = ArgumentParser()
argparser.add_argument("reviews")
argparser.add_argument("--log", default="loss/generator_loss_current.txt")
return argparser.parse_args()
class WindowedBatcher(object):
def __init__(self, sequences, encodings, batch_size=100, sequence_length=50):
self.sequences = sequences
self.pre_vector_sizes = [c.seq[0].shape[0] for c in self.sequences]
self.pre_vector_size = sum(self.pre_vector_sizes)
self.encodings = encodings
self.vocab_sizes = [c.index for c in self.encodings]
self.vocab_size = sum(self.vocab_sizes)
self.batch_index = 0
self.batches = []
self.batch_size = batch_size
self.sequence_length = sequence_length + 1
self.length = len(self.sequences[0])
self.batch_index = 0
self.X = np.zeros((self.length, self.pre_vector_size))
self.X = np.hstack([c.seq for c in self.sequences])
N, D = self.X.shape
assert N > self.batch_size * self.sequence_length, "File has to be at least %u characters" % (self.batch_size * self.sequence_length)
self.X = self.X[:N - N % (self.batch_size * self.sequence_length)]
self.N, self.D = self.X.shape
self.X = self.X.reshape((self.N / self.sequence_length, self.sequence_length, self.D))
self.N, self.S, self.D = self.X.shape
self.num_sequences = self.N / self.sequence_length
self.num_batches = self.N / self.batch_size
self.batch_cache = {}
def next_batch(self):
idx = (self.batch_index * self.batch_size)
if self.batch_index >= self.num_batches:
self.batch_index = 0
idx = 0
if self.batch_index in self.batch_cache:
batch = self.batch_cache[self.batch_index]
self.batch_index += 1
return batch
X = self.X[idx:idx + self.batch_size]
y = np.zeros((X.shape[0], self.sequence_length, self.vocab_size))
for i in xrange(self.batch_size):
for c in xrange(self.sequence_length):
seq_splits = np.split(X[i, c], np.cumsum(self.pre_vector_sizes))
vec = np.concatenate([e.convert_representation(split) for
e, split in zip(self.encodings, seq_splits)])
y[i, c] = vec
X = y[:, :-1, :]
y = y[:, 1:, :self.vocab_sizes[0]]
X = np.swapaxes(X, 0, 1)
y = np.swapaxes(y, 0, 1)
# self.batch_cache[self.batch_index] = X, y
self.batch_index += 1
return X, y
def generate_sample(length):
'''Generate a sample from the current version of the generator'''
characters = [np.array([0])]
generator2.reset_states()
for i in xrange(length):
output = generator2.predict(np.eye(len(text_encoding))[None, characters[-1]])
sample = np.random.choice(xrange(len(text_encoding)), p=output[0, 0])
characters.append(np.array([sample]))
characters = np.array(characters).ravel()
num_seq = NumberSequence(characters[1:])
return num_seq.decode(text_encoding)
if __name__ == '__main__':
args = parse_args()
logging.debug('Reading file...')
with open(args.reviews, 'r') as f:
reviews = [r[3:] for r in f.read().strip().split('\n')]
reviews = [r.replace('\x05', '') for r in reviews]
reviews = [r.replace('<STR>', '') for r in reviews]
logging.debug('Retrieving text encoding...')
with open('data/charnet-encoding.pkl', 'rb') as fp:
text_encoding = pickle.load(fp)
# Create reviews and targets
logging.debug('Converting to one-hot...')
review_sequences = [CharacterSequence.from_string(r) for r in reviews]
num_sequences = [c.encode(text_encoding) for c in review_sequences]
final_sequences = NumberSequence(np.concatenate([c.seq.astype(np.int32) for c in num_sequences]))
# Construct the batcher
batcher = WindowedBatcher([final_sequences], [text_encoding], sequence_length=200, batch_size=100)
generator = Sequence(Vector(len(text_encoding), batch_size=100)) >> Repeat(LSTM(1024, stateful=True), 2) >> Softmax(len(text_encoding))
generator2 = Sequence(Vector(len(text_encoding), batch_size=1)) >> Repeat(LSTM(1024, stateful=True), 2) >> Softmax(len(text_encoding))
logging.debug('Loading prior model...')
with open('models/generative/generative-model-0.0.renamed.pkl', 'rb') as fp:
generator.set_state(pickle.load(fp))
with open('models/generative/generative-model-0.0.renamed.pkl', 'rb') as fp:
generator2.set_state(pickle.load(fp))
# Optimization procedure
rmsprop = RMSProp(generator, CrossEntropy())
def train_generator(iterations, step_size):
with open(args.log, 'w') as f:
for _ in xrange(iterations):
X, y = batcher.next_batch()
loss = rmsprop.train(X, y, step_size)
print >> f, 'Loss[%u]: %f' % (_, loss)
print 'Loss[%u]: %f' % (_, loss)
f.flush()
with open('models/generative/generative-model-current.pkl', 'wb') as g:
pickle.dump(generator.get_state(), g)
generator2.set_state(generator.get_state())
| [
"liam.fedus@gmail.com"
] | liam.fedus@gmail.com |
92bce89e8eb6178bc209f07bc8b16f2f963e3867 | 20b47d5a51508cbcb8cd57f950e1d0c6f679b71c | /education.py | 0a28a7aef18e8dec4b013e26cccebcc1862c4daf | [] | no_license | amycbaker/Thinkful_Assignments | 91722e60d08022387009548e0849d69b6fa04327 | 46d14e50c25f01bfe857f4620771253a559f2f6f | refs/heads/master | 2021-01-19T23:42:30.374995 | 2017-06-13T04:05:00 | 2017-06-13T04:05:00 | 89,014,116 | 0 | 0 | null | 2017-05-17T16:30:58 | 2017-04-21T18:48:20 | Python | UTF-8 | Python | false | false | 2,696 | py | from bs4 import BeautifulSoup
import requests
import pandas as pd
import csv
import sqlite3 as lite
import statsmodels.formula.api as smf
import math
# store url for school years
url = "http://web.archive.org/web/20110514112442/http://unstats.un.org/unsd/demographic/products/socind/education.htm"
# get the html
r = requests.get(url)
# parse the html content with bs
soup = BeautifulSoup(r.content)
mylist = soup.findAll('tr', attrs=('class', 'tcont'))
mylist = mylist[:93]
#country_name, year, school years, male , female
countries = []
for item in mylist:
countries.append([item.contents[1].string,
item.contents[3].string,
item.contents[9].string,
item.contents[15].string,
item.contents[21].string])
# convert data to pandas dataframe and define column names
df = pd.DataFrame(countries)
df.columns = ['Country', 'DataYear', 'TotalYears', 'MaleYears', 'FemaleYears']
# convert school years to integers
df['TotalYears'] = df['MaleYears'].map(lambda x: int(x))
df['MaleYears'] = df['MaleYears'].map(lambda x: int(x))
df['FemaleYears'] = df['FemaleYears'].map(lambda x: int(x))
print("The city mean years are:")
mean = df.mean()
print mean
print("The city mean years are:")
median = df.median()
print median
max = df.max()
print("The maximum years is")
print max
min = df.min()
print("The minimum years is")
print min
con = lite.connect('education.db')
with con:
cur = con.cursor()
df.to_sql("education_years", con, if_exists="replace")
cur.execute("DROP TABLE IF EXISTS gdp")
cur.execute('CREATE TABLE gdp (country_name text, _1999 integer, _2000 integer, _2001 integer, _2002 integer, _2003 integer, _2004 integer, _2005 integer, _2006 integer, _2007 integer, _2008 integer, _2009 integer, _2010 integer)')
with open('API_NY.GDP.MKTP.CD_DS2_en_csv_v2.csv','rU') as inputFile:
next(inputFile)
next(inputFile)
next(inputFile)
next(inputFile)
header = next(inputFile)
inputReader = csv.reader(inputFile)
for line in inputReader:
cur.execute('INSERT INTO gdp (country_name, _1999, _2000, _2001, _2002, _2003, _2004, _2005, _2006, _2007, _2008, _2009, _2010) VALUES ("' + line[0] + '","' + '","'.join(line[42:-8]) + '");')
cur.execute("SELECT country_name, TotalYears, _2000, _2005, _2010 FROM education_years INNER JOIN gdp ON Country = country_name")
rows = cur.fetchall()
cols = [desc[0] for desc in cur.description]
gdp_df = pd.DataFrame(rows, columns=cols)
est = smf.ols(formula='TotalYears ~ _2010', data=gdp_df).fit()
print(est.summary())
| [
"amybaker@gmail.com"
] | amybaker@gmail.com |
af38ecfee37543b5b56fdeb4ae4fe169f1676baa | ea4b8ad32345a94ec1c566c30efb4dfc9fd46b8e | /GeoGossip/webapps/geogossip/tests.py | 7b788d76bdb245e32e179aac540d6504c253de3b | [] | no_license | yyi1/GeoGossip | 76e09b585c0a765485547560def4b2f9aa407777 | 99fa1d06c4f26ad1f0ab8b1c007ab2d54d3cc56f | refs/heads/master | 2020-04-06T04:21:46.476790 | 2017-02-25T04:45:26 | 2017-02-25T04:45:26 | 82,977,730 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 5,040 | py | from django.test import TestCase
from django.test import Client
from django.contrib.auth import authenticate
from models import User
# Create your tests here.
class EndToEndTest(TestCase):
def setUp(self):
super(EndToEndTest, self).setUp()
self.client = Client()
self.user = User.objects.create_user(username='stonebai', first_name='Shi', last_name='Bai',
email='shib@andrew.cmu.edu', password='123')
new_user = authenticate(username=self.user.username, password='123')
self.assertIsNotNone(new_user)
self.client.login(username=self.user.username, password='123')
pass
#
# def test_home(self):
# response = self.client.get('/')
# self.assertEqual(response.status_code, 200)
# pass
def test_profile(self):
response = self.client.get('/geogossip/profile/' + str(self.user.id))
self.assertEqual(response.status_code, 200)
pass
def test_get_group_with_get_method(self):
response = self.client.get('/geogossip/get-groups')
self.assertEqual(response.status_code, 404)
pass
def test_get_group_success(self):
response = self.client.post('/geogossip/get-groups', data={
'lat': 0.0,
'lon': 0.0
})
self.assertEqual(response.status_code, 200)
pass
def test_get_group_with_invalid_lat(self):
response = self.client.post('/geogossip/get-groups', data={
'lat': 91.0,
'lon': 0.0
})
self.assertEqual(response.status_code, 400)
pass
def test_get_group_with_invalid_lon(self):
response = self.client.post('/geogossip/get-groups', data={
'lat': 0.0,
'lon': 181.0
})
self.assertEqual(response.status_code, 400)
pass
def test_get_business_with_get_method(self):
response = self.client.get('/geogossip/get-businesses')
self.assertEqual(response.status_code, 404)
pass
def test_get_business_success(self):
response = self.client.post('/geogossip/get-businesses', data={
'lat': 0.0,
'lon': 0.0
})
self.assertEqual(response.status_code, 200)
pass
def test_get_business_with_invalid_lat(self):
response = self.client.post('/geogossip/get-businesses', data={
'lat': 91.0,
'lon': 0.0
})
self.assertEqual(response.status_code, 400)
pass
def test_get_business_with_invalid_lon(self):
response = self.client.post('/geogossip/get-businesses', data={
'lat': 0.0,
'lon': 181.0
})
self.assertEqual(response.status_code, 400)
pass
def test_non_exists_group_chat(self):
response = self.client.get('/geogossip/group-chat/1')
self.assertEqual(response.status_code, 404)
pass
def test_non_exists_avatar(self):
response = self.client.get('/geogossip/avatar/1')
self.assertEqual(response.status_code, 404)
pass
# def test_profile(self):
# response = self.client.get('/geogossip/profile/7')
# self.assertEqual(response.status_code, 200)
# pass
# test user_id = 30(invalid uid), redirect to home page
def test_get_profileWithInvalidID_session(self):
response = self.client.get('/geogossip/profile/30')
self.assertEqual(response.status_code, 404)
pass
#############################################################
# Test @login_required #
#############################################################
def test_home_session(self):
client = Client()
response = client.get('/')
self.assertEqual(response.status_code, 302)
pass
def test_logout_session(self):
client = Client()
response = client.get('/geogossip/logout')
self.assertEqual(response.status_code, 302)
pass
def test_get_group_session(self):
client = Client()
response = client.get('/geogossip/get-groups')
self.assertEqual(response.status_code, 302)
pass
def test_get_getBusinesses_session(self):
client = Client()
response = client.get('/geogossip/get-businesses')
self.assertEqual(response.status_code, 302)
pass
def test_get_createGroup_session(self):
client = Client()
response = client.get('/geogossip/create-group')
self.assertEqual(response.status_code, 302)
pass
# test user_id = 7
def test_get_profile_session(self):
client = Client()
response = client.get('/geogossip/profile/7')
self.assertEqual(response.status_code, 302)
pass
def test_get_profileWithoutID_session(self):
client = Client()
response = client.get('/geogossip/profile')
self.assertEqual(response.status_code, 404)
pass
pass
| [
"yyi1@andrew.cmu.edu"
] | yyi1@andrew.cmu.edu |
c3bbb5738b81da3295cb82f51894e74b8553f71b | 7765c093fbfaebc3328f8500db2e462977ac42a5 | /sqlite/sample.py | f4dc2f38f85c48f038a9b6f853da204c4bf0df63 | [] | no_license | iamkamleshrangi/datascience | e118e41591850f24438aa344100a07737490fd29 | 7add9501c3ac75323e94df5351e2baf6cadb73ae | refs/heads/master | 2022-02-02T20:19:20.986813 | 2018-07-23T13:26:37 | 2018-07-23T13:26:37 | 128,158,552 | 0 | 0 | null | 2022-01-21T04:26:26 | 2018-04-05T04:22:15 | Python | UTF-8 | Python | false | false | 358 | py | # Create engine: engine
engine = create_engine('sqlite:///Chinook.sqlite')
# Open engine in context manager
with engine.connect() as con:
rs = con.execute('select * from Employee order by BirthDate asc')
df = pd.DataFrame(rs.fetchall())
# Set the DataFrame's column names
df.columns = rs.keys()
# Print head of DataFrame
print(df.head())
| [
"iamkamleshrangi@gmail.com"
] | iamkamleshrangi@gmail.com |
28afd10dd4bf86cc9fc12239cac8891a7b46c5df | a9243f735f6bb113b18aa939898a97725c358a6d | /0.12/_downloads/plot_time_frequency_mixed_norm_inverse.py | 65ac593e852afd7ae0cd4471a6c573000a16b131 | [] | permissive | massich/mne-tools.github.io | 9eaf5edccb4c35831400b03278bb8c2321774ef2 | 95650593ba0eca4ff8257ebcbdf05731038d8d4e | refs/heads/master | 2020-04-07T08:55:46.850530 | 2019-09-24T12:26:02 | 2019-09-24T12:26:02 | 158,233,630 | 0 | 0 | BSD-3-Clause | 2018-11-19T14:06:16 | 2018-11-19T14:06:16 | null | UTF-8 | Python | false | false | 4,959 | py | """
=============================================
Compute MxNE with time-frequency sparse prior
=============================================
The TF-MxNE solver is a distributed inverse method (like dSPM or sLORETA)
that promotes focal (sparse) sources (such as dipole fitting techniques).
The benefit of this approach is that:
- it is spatio-temporal without assuming stationarity (sources properties
can vary over time)
- activations are localized in space, time and frequency in one step.
- with a built-in filtering process based on a short time Fourier
transform (STFT), data does not need to be low passed (just high pass
to make the signals zero mean).
- the solver solves a convex optimization problem, hence cannot be
trapped in local minima.
References:
A. Gramfort, D. Strohmeier, J. Haueisen, M. Hamalainen, M. Kowalski
Time-Frequency Mixed-Norm Estimates: Sparse M/EEG imaging with
non-stationary source activations
Neuroimage, Volume 70, 15 April 2013, Pages 410-422, ISSN 1053-8119,
DOI: 10.1016/j.neuroimage.2012.12.051.
A. Gramfort, D. Strohmeier, J. Haueisen, M. Hamalainen, M. Kowalski
Functional Brain Imaging with M/EEG Using Structured Sparsity in
Time-Frequency Dictionaries
Proceedings Information Processing in Medical Imaging
Lecture Notes in Computer Science, 2011, Volume 6801/2011,
600-611, DOI: 10.1007/978-3-642-22092-0_49
https://doi.org/10.1007/978-3-642-22092-0_49
"""
# Author: Alexandre Gramfort <alexandre.gramfort@telecom-paristech.fr>
#
# License: BSD (3-clause)
import mne
from mne.datasets import sample
from mne.minimum_norm import make_inverse_operator, apply_inverse
from mne.inverse_sparse import tf_mixed_norm
from mne.viz import plot_sparse_source_estimates
print(__doc__)
data_path = sample.data_path()
subjects_dir = data_path + '/subjects'
fwd_fname = data_path + '/MEG/sample/sample_audvis-meg-eeg-oct-6-fwd.fif'
ave_fname = data_path + '/MEG/sample/sample_audvis-no-filter-ave.fif'
cov_fname = data_path + '/MEG/sample/sample_audvis-shrunk-cov.fif'
# Read noise covariance matrix
cov = mne.read_cov(cov_fname)
# Handling average file
condition = 'Left visual'
evoked = mne.read_evokeds(ave_fname, condition=condition, baseline=(None, 0))
evoked = mne.pick_channels_evoked(evoked)
# We make the window slightly larger than what you'll eventually be interested
# in ([-0.05, 0.3]) to avoid edge effects.
evoked.crop(tmin=-0.1, tmax=0.4)
# Handling forward solution
forward = mne.read_forward_solution(fwd_fname, force_fixed=False,
surf_ori=True)
###############################################################################
# Run solver
# alpha_space regularization parameter is between 0 and 100 (100 is high)
alpha_space = 50. # spatial regularization parameter
# alpha_time parameter promotes temporal smoothness
# (0 means no temporal regularization)
alpha_time = 1. # temporal regularization parameter
loose, depth = 0.2, 0.9 # loose orientation & depth weighting
# Compute dSPM solution to be used as weights in MxNE
inverse_operator = make_inverse_operator(evoked.info, forward, cov,
loose=loose, depth=depth)
stc_dspm = apply_inverse(evoked, inverse_operator, lambda2=1. / 9.,
method='dSPM')
# Compute TF-MxNE inverse solution
stc, residual = tf_mixed_norm(evoked, forward, cov, alpha_space, alpha_time,
loose=loose, depth=depth, maxit=200, tol=1e-4,
weights=stc_dspm, weights_min=8., debias=True,
wsize=16, tstep=4, window=0.05,
return_residual=True)
# Crop to remove edges
stc.crop(tmin=-0.05, tmax=0.3)
evoked.crop(tmin=-0.05, tmax=0.3)
residual.crop(tmin=-0.05, tmax=0.3)
# Show the evoked response and the residual for gradiometers
ylim = dict(grad=[-120, 120])
evoked.pick_types(meg='grad', exclude='bads')
evoked.plot(titles=dict(grad='Evoked Response: Gradiometers'), ylim=ylim,
proj=True)
residual.pick_types(meg='grad', exclude='bads')
residual.plot(titles=dict(grad='Residuals: Gradiometers'), ylim=ylim,
proj=True)
###############################################################################
# View in 2D and 3D ("glass" brain like 3D plot)
plot_sparse_source_estimates(forward['src'], stc, bgcolor=(1, 1, 1),
opacity=0.1, fig_name="TF-MxNE (cond %s)"
% condition, modes=['sphere'], scale_factors=[1.])
time_label = 'TF-MxNE time=%0.2f ms'
clim = dict(kind='value', lims=[10e-9, 15e-9, 20e-9])
brain = stc.plot('sample', 'inflated', 'rh', clim=clim, time_label=time_label,
smoothing_steps=5, subjects_dir=subjects_dir)
brain.show_view('medial')
brain.set_data_time_index(120)
brain.add_label("V1", color="yellow", scalar_thresh=.5, borders=True)
brain.add_label("V2", color="red", scalar_thresh=.5, borders=True)
| [
"larson.eric.d@gmail.com"
] | larson.eric.d@gmail.com |
dfee4727ac0ae042e9312e84a9fcd32b98b17fc0 | 81c0dcb009cd30e12e6948b90a0a2ff71fa88d98 | /word/index.py | 9474bdb8bd374a2cdc94a2b0ad6775e1490590af | [] | no_license | function2-llx/IR-system | c4bc3b5a84f693c2c62b03979adea39601cb09c9 | 8b40e8c9637e1d5d3665df48670a9095f4027665 | refs/heads/master | 2023-02-09T03:25:45.723609 | 2021-01-01T16:14:36 | 2021-01-01T16:14:36 | 311,276,530 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 845 | py | import json
from tqdm import tqdm
from elasticsearch import Elasticsearch
from elasticsearch.helpers import bulk
client = Elasticsearch()
inner_sep = '\t'
token_sep = ' '
if __name__ == '__main__':
corpus = json.load(open('sample.json'))
chunk_size = 1000
for i in tqdm(range(0, len(corpus), chunk_size)):
batch = corpus[i:i + chunk_size]
actions = []
for j, doc in enumerate(corpus[i:i + chunk_size]):
tokens = doc['tokens']
content = token_sep.join([inner_sep.join((token, pos, tokens[head - 1] if head else '', rel)) for token, pos, (_, head, rel) in zip(tokens, doc['pos'], doc['dep'])])
actions.append({
'_id': i * chunk_size + j,
'_source': {'content': content}
})
results = bulk(client, actions, index='docs')
| [
"function2@qq.com"
] | function2@qq.com |
2f4b33789431edee4986b09ab5ab538d8efb35ac | 10914bad0901b0e9d0233418f850050b23fe4da2 | /models/pretrained_mobilenet.py | 34aa6ef8eeb074987ccbbc850e52d60e7e50dc2f | [] | no_license | wQuole/image_classifier | 3ced4f48bdef555eaf07858a56c56930ca3f39b3 | 7ebd1794381902ddb316161d4778cb84a3cc0b13 | refs/heads/master | 2022-09-09T08:22:41.818211 | 2020-05-31T17:50:43 | 2020-05-31T17:50:43 | 267,560,900 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 1,289 | py | from tensorflow.keras.applications import MobileNetV2
from tensorflow.keras.models import Sequential
from tensorflow.keras.optimizers import RMSprop
from tensorflow.keras.losses import BinaryCrossentropy
from tensorflow.keras.layers import Conv2D, MaxPooling2D, Dropout, Flatten, Dense, Activation, GlobalAveragePooling2D
class PretrainedMobileNetV2:
def __init__(self, IMAGE_SIZE):
self.image_size = IMAGE_SIZE
self.model = Sequential()
self.fine_tune()
def fine_tune(self):
mobilenetv2_model = MobileNetV2(input_shape=(self.image_size),
include_top=False,
pooling='avg',
weights="imagenet")
for layer in mobilenetv2_model.layers:
# Freeze layers that should not be re-trained
layer.trainable = False
# Add all layers from basemodel, trainable and non-trainable to our model
self.model.add(mobilenetv2_model)
# Add classification block
self.model.add(Dense(2, activation='softmax'))
self.model.compile(optimizer=RMSprop(lr=1e-4),
loss=BinaryCrossentropy(from_logits=True),
metrics=['accuracy'])
| [
"wgkvaale@gmail.com"
] | wgkvaale@gmail.com |
bde1909ef256bb98db99f9655bb6c8af594fcd1c | f3dc7c8ab38e3affaacbc05179aa30d14589adf1 | /SOLA.py | e51c3c4172777cf5b63909b718772f4a7582392f | [] | no_license | jishnub/moccasin | fa2c6ffd7f42f99aacb9f4519a58017f37715fab | 3495cc82c70d8a62f2fda551b2d269b17f4efeef | refs/heads/master | 2020-03-18T07:34:39.901461 | 2019-01-21T11:02:12 | 2019-01-21T11:02:12 | 134,461,587 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 6,599 | py | from astropy.io import fits
import numpy as np; import array
import matplotlib.pyplot as plt
from matplotlib import cm
import scipy
import math
from numpy.linalg import inv
import numpy.matlib
def dfac(n):
a = np.zeros_like(n)
for j in range(0,n.size):
a[j] = np.prod(np.arange(n[j],1,-2))
return a
def fac(n):
a = np.zeros_like(n)
for j in range(0,n.size):
a[j] = math.factorial(n[j])
return a
elmin = 1
elmax = 10
dellmin = 1
dellmax =1
instrument = "HMI"
track = 430
smin =0
smax = 7
perform_inversion = True
sigma_min = 1 # micro-Hertz
sigma_max = 200 # micro-Hertz
nyears = 6 # each "year" is 360 days
startyear = 1
r0 = 0.15
deltar = 0.05
reg = 1e6
dnu = 1e6/(nyears*360.0*86400.0) # in mu Hz
freqfloor = int(np.floor(sigma_min/dnu) +1)
r=np.squeeze(fits.open("radius.fits")[0].data)
trackch =str("{:03d}".format(track));
lminst =str("{:03d}".format(elmin));
lmaxst =str("{:03d}".format(elmax));
direct0 = '/scratch/shravan/HMI'
direct = direct0 + '/tracking' + trackch
ns = (smax+1)**2
dellst =str("{:01d}".format(dellmin));
kern = np.squeeze(fits.open(instrument+"_kernels_"+lminst +"_to_"+lmaxst+"_dell_"+dellst+".fits")[0].data)
indices = np.loadtxt(instrument+"_indices_"+lminst +"_to_"+lmaxst+"_dell_"+dellst)
for dell in range(dellmin+1, dellmax+1):
dellst =str("{:01d}".format(dell));
kerntemp = np.squeeze(fits.open(instrument+"_kernels_"+lminst +"_to_"+lmaxst+"_dell_"+dellst+".fits")[0].data)
g = np.loadtxt(instrument+"_indices_"+lminst +"_to_"+lmaxst+"_dell_"+dellst)
indices = np.r_[indices, g]
kern = np.r_[kern, kerntemp]
h = kern.shape
nkerns = h[0]
dr = np.zeros_like(r)
nr = r.shape
nr = nr[0]
dr[0:nr-2] = r[1:nr-1] - r[0:nr-2]
dr[nr-1] = dr[nr-2]
target = np.exp(-(r-r0)**2.0/(2.0*deltar**2))
target = target/np.sum(target*dr)
A = np.zeros((nkerns,nkerns))
rhs = np.zeros((nkerns,1))
for j in range(0,nkerns):
temp = dr*kern[j,:]
rhs[j] = np.sum(temp*target)
for i in range(j,nkerns):
A[i,j] = np.sum(temp*kern[i,:])
A[j,i] = A[i,j]
parity = np.zeros((nkerns,1))
coeffstor = np.zeros((nkerns,smax+1))
coeffspol = np.zeros((nkerns,smax+1))
elldiff = indices[:,2] - indices[:,0]
for s in range(smin,smax+1):
parity = np.mod(elldiff + s,2)
hh = np.where(abs(elldiff) <= s)[0]
if (hh.size == 0):
continue
tind = hh[np.where(parity[hh] == 1)[0]]
#tind = tind[hh]
pind = hh[np.where(parity[hh] == 0)[0]]
#pind = np.where(parity == 0 and abs(elldiff) <= s)
sumdiff1 = s + elldiff
sumdiff2 = s - elldiff
if (tind.size >0):
factor = (1-2*np.mod((sumdiff1[tind] - 1)/2,2)) * dfac(sumdiff1[tind]) * dfac(sumdiff2[tind])/(fac(sumdiff1[tind])*fac(sumdiff2[tind]))**0.5
rhstor = rhs[tind,0] * factor
Ator = A[np.ix_(tind, tind)]
coeffstor[tind,s] = np.matmul(inv(Ator + reg * np.eye(tind.size)), rhstor)
# if (s==2):
# func = np.matmul(np.squeeze(coeffstor[tind,s]),kern[tind,:])
# plt.plot(r,target/target.max()); plt.plot(r,func/func.max()); plt.show()
# stop
if (pind.size >0):
facpol = (1-2*np.mod(sumdiff1[pind]/2,2)) * elldiff[pind] * dfac(sumdiff1[pind]-1) * dfac(sumdiff2[pind]-1)/(fac(sumdiff1[pind])*fac(sumdiff2[pind]))**0.5
rhspol = rhs[pind,0] * facpol
Apol = A[np.ix_(pind, pind)]
coeffspol[pind,s] = np.matmul(inv(Apol + reg * np.eye(pind.size)), rhspol)
if (perform_inversion):
nfreq = int(np.floor((sigma_max - sigma_min)/dnu)) + 2
a = np.zeros([nfreq,60,ns],'complex'); powpos = 0; powneg = 0;
nus = (np.arange(nfreq) + freqfloor)*dnu
noitoroidal = np.zeros([nfreq,ns])
noipoloidal = np.zeros([nfreq,ns])
toroidal = np.zeros([nfreq,ns], dtype = complex)
poloidal = np.zeros([nfreq,ns], dtype = complex)
nors = np.zeros([60]);nord = np.zeros([60]); nordp = np.zeros([60])
stry1 = str("{:02d}".format(5*(startyear-1)+1));
stry2 = str("{:02d}".format(5*(startyear+nyears-1)));
stryear = '_year_'+stry1+'_'+stry2
for dell in range(dellmin, dellmax+1):
for ell in range(elmin, elmax+1-dell):
print "ell:", ell
ellc =str("{:03d}".format(ell))
ellp = ell + dell
ellpc =str("{:03d}".format(ellp))
allind = np.where(indices[:,0] == ell)[0]
allind = allind[np.where(indices[allind,2] == ellp)]
te = fits.open(direct+'/bcoef_l_'+ellc +'_lp_'+ellpc+stryear+'.fits')[0].data
noit = fits.open(direct+'/noise_l_'+ellc +'_lp_'+ellpc+stryear+'.fits')[0].data
nfrequ = te.shape[1]
te = te[0,:,:,:]+1j*te[1,:,:,:]
f = open(direct+'/frequency_metadata_l_'+ellc +'_lp_'+ellpc+stryear, 'r')
j=-1
k= -1
for line in f:
j = j+1
if (j==0):
line = line.strip()
columns = line.split()
dnu = float(columns[0])
if (j <= 3):
continue
k = k+1
line = line.strip()
columns = line.split()
freqdiff = np.float(columns[5]) - np.float(columns[2])
if (np.absolute(freqdiff) < sigma_min or np.absolute(freqdiff) > sigma_max):
continue
fst = int(np.floor(np.absolute(freqdiff)/dnu)) - freqfloor
fend = np.minimum(fst + nfrequ, nfreq)
nord = int(columns[1])
nordp = int(columns[4])
nind = np.where(indices[allind,1] == nord)[0]
freql = nfrequ + (fend-fst - nfrequ)
#print fst,fend,nus[fst],nus[fend-1],fend-fst,nfrequ,freql,nus.shape
# if (fend > nfreq-1):
# print "Frequency range too high, skipping, ell =", ell, "dell =", dell, "n = ", nord, "n' = ", nordp, "freq. difference = ", freqdiff
# continue
coefind = allind[np.where(indices[allind[nind],3] == nordp)][0]
for s in range(smin,smax+1):
for t in range(-s,s+1):
ind = s**2 + s + t
poloidal[fst:fend,ind] = poloidal[fst:fend,ind] + te[0:freql,k,ind]*coeffspol[coefind,s]*(1.0 + 0.0j)
noipoloidal[fst:fend,ind] = noipoloidal[fst:fend,ind] + noit[0:freql,k,ind]*coeffspol[coefind,s]
toroidal[fst:fend,ind] = toroidal[fst:fend,ind] + te[0:freql,k,ind]*coeffstor[coefind,s]*(1.0 + 0.0j)
noitoroidal[fst:fend,ind] = noitoroidal[fst:fend,ind] + noit[0:freql,k,ind]*coeffstor[coefind,s]
f.close()
| [
"hanasoge@gmail.com"
] | hanasoge@gmail.com |
ae5a6526f090a8363d28b2f1de374b3d972022b0 | 5b256252d35410f7c3239b13d8aed801dc3ad0e8 | /More Exercises/city.py | 90fd5eda5f36e8d55273d3847aae6fc6ab996092 | [] | no_license | Nmazil-Dev/PythonCrashCoursePractice | d52d3f4585fbce18011720cc3cef83f095add526 | 7bf55dc8298e70ee4b5b5345e1171c188ea8bbda | refs/heads/master | 2020-04-27T09:56:22.937795 | 2019-03-06T23:15:05 | 2019-03-06T23:15:05 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 310 | py | def city_name(city, country, population=''):
if population == '':
name = city.title() + ', ' + country.title() + population
elif population != '':
name = city.title() + ', ' + country.title() + ' - population ' + str(population)
print (name)
city_name('brookfield', 'usa')
| [
"nmazil68@gmail.com"
] | nmazil68@gmail.com |
ea9db589734e38f7ee6202aef62fb859d876b357 | 05f7f004ccd926c1611dc03473e0778d8c332e14 | /lcmf_projects/bond_click.py | c44ffa7db70e91bd17ff2cafe6c8693d6261a9f2 | [] | no_license | tongch8819/lcmf_projects | 06cd875e5b001a871cc11cdc8cf45a32f7faa105 | a7243aee94da9bbf9651e1365351c5b4ef364b80 | refs/heads/master | 2022-04-04T22:15:26.726761 | 2020-02-24T03:51:23 | 2020-02-24T03:51:23 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 1,807 | py | import numpy as np
import click
from scipy import optimize
@click.command()
@click.option('--coefficients', default=(1,0,2), help='tuple for representation of polynomial')
def ployvalTest(coefficients: tuple =(1,0,2) ) -> str:
print('Input: ', coefficients)
res = np.polyval(coefficients, 1)
if res == 3:
return 'descending'
else:
return 'ascending'
@click.command()
@click.option('--value', help='value of bond')
@click.option('--price', help='price of bond')
@click.option('--couponRate', help='couponRate of bond')
@click.option('--period', help='period of bond')
# 跟 wind 实盘 最新YTM 对不上 !!!!!!!!!!
def normalBondYTM(value: float, price: float, couponRate: float, period: int):
coupon = value * couponRate
poly = np.array([-price] + [coupon] * (period-1) + [coupon+value])
roots = np.roots(poly)
for root in roots:
if root.imag == 0.0:
return root.real - 1
# 跟 wind 实盘 久期 对不上 !!!!!!!!!!
def normalBondDuration(value: float, price: float, couponRate: float, period: int):
YTM = normalBondYTM(value, price, couponRate, period)
vec1 = np.array([i*np.exp(-YTM * i) for i in range(period+1)[1:]])
coupon = value * couponRate
vec2 = np.array([coupon] * (period -1) + [coupon+value])
if price != 0:
continuousDuration = vec1.dot(vec2) / price
modifiedDuration = continuousDuration/(1+YTM)
return modifiedDuration
else:
print('price is zero')
return
if __name__ == '__main__':
# print(ployvalTest())
# print(normalBondYTM(100, 90.68, 0.0375, 5))
# print(normalBondDuration(100, 90.68, 0.0375, 5))
sen = ployvalTest()
print(sen) | [
"tong.cheng.8819@outlook.com"
] | tong.cheng.8819@outlook.com |
9b9bf9b3e9aaeb11ed05526758f3d5fb48fab438 | 09b41a5db055eccee47aba8b305c25515d303793 | /title_test.py | f84ee707d3e011be6b95904389fbf36cf84834ed | [
"Apache-2.0"
] | permissive | ishansd94/travisci-test | 1c83638f8576a931d7776394ee9926b8c7cf07bc | 2edd64e655f8536bddd9655992d1397a10835ed7 | refs/heads/master | 2020-03-30T09:09:52.755411 | 2018-10-01T11:36:52 | 2018-10-01T11:36:52 | 151,063,228 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 239 | py | import unittest
from title import convert
class TestConvert(unittest.TestCase):
def test_convert(self):
str = convert("hello world")
self.assertEqual(str, "Hello World")
if __name__ == '__main__':
unittest.main() | [
"ishan.dassanayake@pearson.com"
] | ishan.dassanayake@pearson.com |
f2a89ce80fd4ce27424ff0a7e83f15dcf071cf2b | 4994d9e1e3bbf4fc3d2ce7b5f3a41b3d0ff1e165 | /lib/handle/StaticHandler.py | 053961a57bf511f3f79a3ace3ee78ea3d3c07234 | [
"MIT"
] | permissive | francoricci/sapspid | 228ab362c123cf3d28c9eab215f5daafe4aa293b | db335f2335824ba4f7aa7a01cd15c235bc815a47 | refs/heads/master | 2021-07-13T02:19:12.923503 | 2021-03-03T15:59:43 | 2021-03-03T15:59:43 | 97,124,842 | 2 | 0 | null | null | null | null | UTF-8 | Python | false | false | 212 | py | import response
class StaticFileHandler(response.StaticFileHandler):
def write_error(self, status_code, **kwargs):
super(StaticFileHandler, self).write_error(status_code, errorcode = '3', **kwargs) | [
"pippo@pippo.com"
] | pippo@pippo.com |
457107eadcf898c5b73e8308dd0dda559a8e82c0 | 1eefc1fc19dd4b0ded6eaad75d232450d00e0eba | /bench/pya.py | e72d2be49042f8110990e00bb7955677fd30d4af | [
"MIT"
] | permissive | pskopnik/apq | 2dbfa3e56c6e5c836b9d38a9e4b6bdf2f83eb44a | 827e722ec604d2f7c050f43748136613c3cd3d70 | refs/heads/master | 2021-07-07T08:32:40.204240 | 2020-02-12T13:33:41 | 2020-02-12T13:33:41 | 237,626,912 | 4 | 1 | MIT | 2021-04-20T19:28:42 | 2020-02-01T14:31:05 | Python | UTF-8 | Python | false | false | 2,319 | py | from . import bench, BenchTimer, main_bench_registered
from .utils import StringSource
from .py.keyedpq_a import PyKeyedPQA
from random import random as random_01
@bench()
def bench_add(b: BenchTimer) -> None:
s = StringSource()
s_offset = StringSource()
pq: PyKeyedPQA[str, None] = PyKeyedPQA()
for _ in range(10000):
pq.add(next(s), random_01(), None)
next(s_offset)
with b.time() as t:
for _ in t:
pq.add(next(s), random_01(), None)
with b.offset() as t:
for _ in t:
next(s_offset)
random_01()
@bench()
def bench_pop(b: BenchTimer) -> None:
s = StringSource()
pq: PyKeyedPQA[str, None] = PyKeyedPQA()
for _ in range(b.n + 10000):
pq.add(next(s), random_01(), None)
with b.time() as t:
for _ in t:
pq.pop()
with b.offset() as t:
for _ in t:
pass
@bench()
def bench_pop_add(b: BenchTimer) -> None:
s = StringSource()
s_offset = StringSource()
pq: PyKeyedPQA[str, None] = PyKeyedPQA()
for _ in range(10000):
pq.add(next(s), random_01(), None)
next(s_offset)
with b.time() as t:
for _ in t:
pq.pop()
pq.add(next(s), random_01(), None)
with b.offset() as t:
for _ in t:
next(s_offset)
random_01()
@bench()
def bench_change_value(b: BenchTimer) -> None:
s = StringSource()
pq: PyKeyedPQA[str, None] = PyKeyedPQA()
for _ in range(10000):
pq.add(next(s), random_01(), None)
with b.time() as t:
for _ in t:
key = s.rand_existing()
pq.change_value(key, random_01())
with b.offset() as t:
for _ in t:
key = s.rand_existing()
random_01()
@bench()
def bench_remove(b: BenchTimer) -> None:
s = StringSource()
s_remove = StringSource()
s_offset = StringSource()
pq: PyKeyedPQA[str, None] = PyKeyedPQA()
for _ in range(b.n + 10000):
pq.add(next(s), random_01(), None)
next(s_offset)
with b.time() as t:
for _ in t:
key = next(s_remove)
del pq[key]
with b.offset() as t:
for _ in t:
key = next(s_offset)
if __name__ == '__main__':
main_bench_registered()
| [
"paul@skopnik.me"
] | paul@skopnik.me |
b548eedfdd00fe7c08f5ba00618fbe44e0cba7df | 2bdedcda705f6dcf45a1e9a090377f892bcb58bb | /src/main/output/pipeline/service_group/number_office/time/fact.py | e3cbfccb649de7dbf84162e340a4f0fe1510ddd6 | [] | no_license | matkosoric/GenericNameTesting | 860a22af1098dda9ea9e24a1fc681bb728aa2d69 | 03f4a38229c28bc6d83258e5a84fce4b189d5f00 | refs/heads/master | 2021-01-08T22:35:20.022350 | 2020-02-21T11:28:21 | 2020-02-21T11:28:21 | 242,123,053 | 1 | 0 | null | null | null | null | UTF-8 | Python | false | false | 3,575 | py | package textTranslator;
import java.io.*;
import java.net.*;
import java.util.*;
import com.google.gson.*;
import com.squareup.okhttp.*;
public class Translate {
String subscriptionKey = 'b58103fec253e2c21b0fdc1a24e16352';
String url = "https://api.cognitive.microsofttranslator.com/translate?api-version=3.0&to=";
public Translate(String subscriptionKey) {
this.subscriptionKey = subscriptionKey;
}
// Instantiates the OkHttpClient.
OkHttpClient client = new OkHttpClient();
// This function performs a POST request.
public String Post() throws IOException {
MediaType mediaType = MediaType.parse("application/json");
RequestBody body = RequestBody.create(mediaType,
"[{\n\t\"Text\": \"Welcome to Microsoft Translator. Guess how many languages I speak!\"\n}]");
Request request = new Request.Builder()
.url(url).post(body)
.addHeader("ec0c96a092ea0a3ba1041f4738a0b33a", subscriptionKey)
.addHeader("Content-type", "application/json").build();
Response response = client.newCall(request).execute();
return response.body().string();
}
public String Post(String bodyStr, String translateTo) throws IOException {
MediaType mediaType = MediaType.parse("application/json");
RequestBody body = RequestBody.create(mediaType,
"[{\n\t\"Text\": \"" + bodyStr + "\"\n}]");
Request request = new Request.Builder()
.url(url + translateTo).post(body)
.addHeader("f460aacf46d11f243d71d7221840dbe5", subscriptionKey)
.addHeader("Content-type", "application/json").build();
Response response = client.newCall(request).execute();
return response.body().string();
}
// This function prettifies the json response.
public static String prettify(String json_text) {
JsonParser parser = new JsonParser();
JsonElement json = parser.parse(json_text);
Gson gson = new GsonBuilder().setPrettyPrinting().create();
return gson.toJson(json);
}
public static String getTranslatedText(String jsonText) {
JsonParser parser = new JsonParser();
JsonArray json = parser.parse(jsonText).getAsJsonArray();
String translatedText = null;
for (int i = 0; i < json.size(); i++) {
if (translatedText != null)
break;
JsonObject jsonObj = json.get(i).getAsJsonObject();
JsonArray translations = jsonObj.getAsJsonArray("translations");
if (translations == null) return "";
for (int j = 0; j < translations.size(); j++) {
if (translatedText != null) break;
JsonObject translation = translations.get(j).getAsJsonObject();
JsonElement text = translation.get("text");
if (text == null) return "";
translatedText = text.getAsString();
}
}
return translatedText;
}
// public static void main(String[] args) {
// try {
// Translate translateRequest = new Translate(System.getenv("Translator"));
//// String response = translateRequest.Post();
//// System.out.println(prettify(response));
//
// String response = translateRequest.Post("Hello", "fr");
// System.out.println(Translate.prettify(response));
//
// System.out.println(getTranslatedText(response));
//
//
// } catch (Exception e) {
// System.out.println(e);
// }
// }
}
| [
"soric.matko@gmail.com"
] | soric.matko@gmail.com |
d17b534e12aa0ac404011b7b2baf87061f771646 | 893e09a68b636a214a75745453fb73ce6c618472 | /lab9_b.py | 40c65fe61ff69a6f3d9df4d95b6f349c17210272 | [] | no_license | pavangabani/DAA_Lab | 241b33d4a13156c2adaa79e82e36140a7f6c3a99 | c3cb265e412ce77106507a8491c24fd38612939f | refs/heads/main | 2023-08-15T01:22:38.135604 | 2021-09-14T16:36:06 | 2021-09-14T16:36:06 | 406,445,915 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 360 | py | def lcs(X , Y):
m = len(X)
n = len(Y)
L = [[None]*(n+1) for i in range(m+1)]
for i in range(m+1):
for j in range(n+1):
if i == 0 or j == 0 :
L[i][j] = 0
elif X[i-1] == Y[j-1]:
L[i][j] = L[i-1][j-1]+1
else:
L[i][j] = max(L[i-1][j] , L[i][j-1])
return L[m][n]
x="pavan"
y="gabani"
a=lcs(x,y)
print ("Length of LCS is ",a)
| [
"pavan.gabani@gmail.com"
] | pavan.gabani@gmail.com |
fd6b8c1a24799e9d64be437cb85b6a6c16c1e23c | 9fe79da67efcd12cae6c61ea360960a87e8fe805 | /web/urls.py | b7afe9636b634ea342a3349487fb98dd5bac0c64 | [] | no_license | ZuiYee/EducationSystem | 00abaa482c393c1b50dac598d4b3ddfdd6268cd7 | 07bf141ea4c213d966105273d3feb9ad997f556f | refs/heads/master | 2020-04-09T23:42:44.673994 | 2019-01-13T06:14:17 | 2019-01-13T06:14:17 | 160,663,085 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 413 | py | from django.conf.urls import url
from . import views
app_name = 'web'
urlpatterns = [
url(r'^studentProfile/', views.studentProfile, name='studentProfile'),
url(r'^teacherProfile/', views.teacherProfile, name='teacherProfile'),
url(r'^studentparseresult/', views.studentparseresult, name='studentparseresult'),
url(r'^teacherparseresult/', views.teacherparseresult, name='teacherparseresult'),
] | [
"39691460+ZuiYee@users.noreply.github.com"
] | 39691460+ZuiYee@users.noreply.github.com |
49bbf912ea16d6bf259a39904154cc010346a28a | 687306842e8082ed1c31441bbacf697352fe1d22 | /design.py | b7a6d0ecdb4066f6d35c6bb75f544e667966a54e | [] | no_license | Garyguo2011/Firewall | a77940d6fa0957fb2c2811cfcc5fa3c3b8982209 | 0906e947853c14b0a04fcccfd350202405b1c8f5 | refs/heads/master | 2020-05-21T11:37:02.562484 | 2014-12-03T15:45:44 | 2014-12-03T15:45:44 | 26,337,718 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 321 | py | class DNSArchive(Archive):
def __init__(self, ....):
self.app = "dns"
self.domainName
class TCPArchive (Archive):
def __init__(self, packet)
class Archive(object):
def __init__(self, ....):
self.direction
self.protocol
self.externalIP
self.countryCode
self.packet
self.matchRules
# Control Plane
| [
"xguo@berkeley.edu"
] | xguo@berkeley.edu |
ffe965efd83b48d88452e41df5c8274713eac169 | ca565548206583a58fe8d646bfd9a6f1ba51c673 | /problem2.py | fa5313404ef249962fe28fa2f3edd13684ba5711 | [] | no_license | GLAU-TND/python-programming-assignment2-kirtimansinghcs19 | fbd772f38fa3546e579ffc2bdf99cc2b34e9937b | 5dc16c8b24186a2e00c749e14eecaac426f51e90 | refs/heads/master | 2021-01-13T22:51:02.990390 | 2020-02-23T16:32:51 | 2020-02-23T16:32:51 | 242,519,926 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 282 | py | from itertools import permutations
def largest(l):
lst=[]
for i in permutations(l, len(l)):
lst.append(''.join(map(str,i)))
return max(lst)
ls=[]
n=int(input('Enter the no element'))
for i in range(0,n):
ls.append(int(input()))
print(largest(ls))
| [
"noreply@github.com"
] | noreply@github.com |
48763a0aeba89cb486860a35c4917e5e2660d135 | 7393b48ff1a403d163812ae77321981586707733 | /email_dl_sep.py | b2a8ef3c378ad7ee7dfe35898484f5b8c67a5ccc | [] | no_license | playerdefault/littlethings | 7121754cc5174c407adeb3d49e5e2f5d7796276c | 5b0fb17f51a7e9179f5523492c44a7eed2f0e47e | refs/heads/master | 2021-07-05T08:05:28.183727 | 2020-08-27T06:41:43 | 2020-08-27T06:41:43 | 168,153,514 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 398 | py | # This program separates a list of semi-colon separated emails from a text file
# and prints out the number of emails
path = input(str("Enter the relative path of the file with the DL List: "))
DLListInputFile = open(path, 'r')
DLInput = DLListInputFile.read()
numberOfEmails = 0
for char in DLInput:
if(char==";"):
numberOfEmails += 1
print("The number of emails is: " + str(numberOfEmails))
| [
"swaraj.mohapatra@outlook.in"
] | swaraj.mohapatra@outlook.in |
49e35d732d050a8de37689f7459907f5c429e2fa | 5a52ccea88f90dd4f1acc2819997fce0dd5ffb7d | /alipay/aop/api/request/AlipayCloudDevopsDictQueryRequest.py | 980977a75413bdc34d657ea1e8ece1c6b0ddb700 | [
"Apache-2.0"
] | permissive | alipay/alipay-sdk-python-all | 8bd20882852ffeb70a6e929038bf88ff1d1eff1c | 1fad300587c9e7e099747305ba9077d4cd7afde9 | refs/heads/master | 2023-08-27T21:35:01.778771 | 2023-08-23T07:12:26 | 2023-08-23T07:12:26 | 133,338,689 | 247 | 70 | Apache-2.0 | 2023-04-25T04:54:02 | 2018-05-14T09:40:54 | Python | UTF-8 | Python | false | false | 3,955 | py | #!/usr/bin/env python
# -*- coding: utf-8 -*-
import json
from alipay.aop.api.FileItem import FileItem
from alipay.aop.api.constant.ParamConstants import *
from alipay.aop.api.domain.AlipayCloudDevopsDictQueryModel import AlipayCloudDevopsDictQueryModel
class AlipayCloudDevopsDictQueryRequest(object):
def __init__(self, biz_model=None):
self._biz_model = biz_model
self._biz_content = None
self._version = "1.0"
self._terminal_type = None
self._terminal_info = None
self._prod_code = None
self._notify_url = None
self._return_url = None
self._udf_params = None
self._need_encrypt = False
@property
def biz_model(self):
return self._biz_model
@biz_model.setter
def biz_model(self, value):
self._biz_model = value
@property
def biz_content(self):
return self._biz_content
@biz_content.setter
def biz_content(self, value):
if isinstance(value, AlipayCloudDevopsDictQueryModel):
self._biz_content = value
else:
self._biz_content = AlipayCloudDevopsDictQueryModel.from_alipay_dict(value)
@property
def version(self):
return self._version
@version.setter
def version(self, value):
self._version = value
@property
def terminal_type(self):
return self._terminal_type
@terminal_type.setter
def terminal_type(self, value):
self._terminal_type = value
@property
def terminal_info(self):
return self._terminal_info
@terminal_info.setter
def terminal_info(self, value):
self._terminal_info = value
@property
def prod_code(self):
return self._prod_code
@prod_code.setter
def prod_code(self, value):
self._prod_code = value
@property
def notify_url(self):
return self._notify_url
@notify_url.setter
def notify_url(self, value):
self._notify_url = value
@property
def return_url(self):
return self._return_url
@return_url.setter
def return_url(self, value):
self._return_url = value
@property
def udf_params(self):
return self._udf_params
@udf_params.setter
def udf_params(self, value):
if not isinstance(value, dict):
return
self._udf_params = value
@property
def need_encrypt(self):
return self._need_encrypt
@need_encrypt.setter
def need_encrypt(self, value):
self._need_encrypt = value
def add_other_text_param(self, key, value):
if not self.udf_params:
self.udf_params = dict()
self.udf_params[key] = value
def get_params(self):
params = dict()
params[P_METHOD] = 'alipay.cloud.devops.dict.query'
params[P_VERSION] = self.version
if self.biz_model:
params[P_BIZ_CONTENT] = json.dumps(obj=self.biz_model.to_alipay_dict(), ensure_ascii=False, sort_keys=True, separators=(',', ':'))
if self.biz_content:
if hasattr(self.biz_content, 'to_alipay_dict'):
params['biz_content'] = json.dumps(obj=self.biz_content.to_alipay_dict(), ensure_ascii=False, sort_keys=True, separators=(',', ':'))
else:
params['biz_content'] = self.biz_content
if self.terminal_type:
params['terminal_type'] = self.terminal_type
if self.terminal_info:
params['terminal_info'] = self.terminal_info
if self.prod_code:
params['prod_code'] = self.prod_code
if self.notify_url:
params['notify_url'] = self.notify_url
if self.return_url:
params['return_url'] = self.return_url
if self.udf_params:
params.update(self.udf_params)
return params
def get_multipart_params(self):
multipart_params = dict()
return multipart_params
| [
"jishupei.jsp@alibaba-inc.com"
] | jishupei.jsp@alibaba-inc.com |
df6c20b6c5095c0d72d68b742ed9c6b48614b69e | 73de83162fd26ea60b0d07a3bb0a9ced63499d43 | /scripts/show_result.py | f1fd3005a1e8db8cd77c90185924775cb3cb8c28 | [
"GPL-3.0-or-later",
"MIT"
] | permissive | Geonhee-LEE/PythonLinearNonlinearControl | fa6c3dec2a7234ddb182388ac2e21074162e2155 | 2a2467098108641483778c09ceb7906cb49f6cee | refs/heads/master | 2023-07-10T03:48:45.566076 | 2021-08-21T12:55:30 | 2021-08-21T12:55:30 | 276,524,383 | 0 | 0 | MIT | 2020-07-02T02:00:18 | 2020-07-02T02:00:17 | null | UTF-8 | Python | false | false | 1,636 | py | import os
import argparse
import pickle
import numpy as np
import matplotlib.pyplot as plt
from PythonLinearNonlinearControl.plotters.plot_func import load_plot_data, \
plot_multi_result
def run(args):
controllers = ["iLQR", "DDP", "CEM", "MPPI"]
history_xs = None
history_us = None
history_gs = None
# load data
for controller in controllers:
history_x, history_u, history_g = \
load_plot_data(args.env, controller,
result_dir=args.result_dir)
if history_xs is None:
history_xs = history_x[np.newaxis, :]
history_us = history_u[np.newaxis, :]
history_gs = history_g[np.newaxis, :]
continue
history_xs = np.concatenate((history_xs,
history_x[np.newaxis, :]), axis=0)
history_us = np.concatenate((history_us,
history_u[np.newaxis, :]), axis=0)
history_gs = np.concatenate((history_gs,
history_g[np.newaxis, :]), axis=0)
plot_multi_result(history_xs, histories_g=history_gs, labels=controllers,
ylabel="x")
plot_multi_result(history_us, histories_g=np.zeros_like(history_us),
labels=controllers, ylabel="u", name="input_history")
def main():
parser = argparse.ArgumentParser()
parser.add_argument("--env", type=str, default="FirstOrderLag")
parser.add_argument("--result_dir", type=str, default="./result")
args = parser.parse_args()
run(args)
if __name__ == "__main__":
main()
| [
"quick1st97@gmail.com"
] | quick1st97@gmail.com |
27a16c27e996906294903e01303e65e7a1d5d0ff | b0769f847d8f2c945f2552891ea7b48e2fad3a0f | /tweetTrend/bin/wheel | f7b6559dcc982dc09a8bff038e4c0507de0ea03c | [] | no_license | jiayangli2/twittmap | b808090acd66a6533e7e17a818512d9a01912632 | 772b9a868c0a9a0fe52c87bae77d43bc71fd9003 | refs/heads/master | 2021-03-22T01:27:09.600457 | 2017-04-09T20:38:37 | 2017-04-09T20:38:37 | 84,241,520 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 252 | #!/home/emmittxu/Desktop/TweetTrend/tweetTrend/bin/python2
# -*- coding: utf-8 -*-
import re
import sys
from wheel.tool import main
if __name__ == '__main__':
sys.argv[0] = re.sub(r'(-script\.pyw|\.exe)?$', '', sys.argv[0])
sys.exit(main())
| [
"gx2127@columbia.edu"
] | gx2127@columbia.edu | |
142f68111255fe08b0cfa29a4378494361ef2c57 | 8ee5dcbdbd407eb5f294d430813b16eca22f571c | /data/HW5/hw5_253.py | 628a39851ed1f06194065eadcb2c20d9da276de9 | [] | no_license | MAPLE-Robot-Subgoaling/IPT | 5e60e2ee4d0a5688bc8711ceed953e76cd2ad5d9 | f512ea3324bfdceff8df63b4c7134b5fcbb0514e | refs/heads/master | 2021-01-11T12:31:00.939051 | 2018-08-13T23:24:19 | 2018-08-13T23:24:19 | 79,373,489 | 1 | 1 | null | null | null | null | UTF-8 | Python | false | false | 383 | py | def main():
width = int(input("please enter the width of the box "))
height = int(input("please enter the height of thebox "))
sym = input("please enter a symbol for the outline ")
fill = input("please enter a fill symbol ")
for h in range(height):
for w in range(width):
print(sym if h in(0,height-1) or w in(0,width-1) else fill, end = ' ')
print()
main()
| [
"mneary1@umbc.edu"
] | mneary1@umbc.edu |
201f42a6dc8b4593fc50814c1c71e25270c0c730 | 0318d24670acc083b67d27027961ba2e060857b4 | /naiveBayes_logisticRegression/utility.py | 455d08018463e44195b6d09999257d67ea39ffe2 | [] | no_license | HC15/Machine-Learning-Email-Classification | 7d4e4ef9c76d41884c29c72179f6df8f204529a7 | fe4945bc01ac9055aec143478d6bedaa8f71eda6 | refs/heads/master | 2020-03-08T04:00:52.183730 | 2018-05-07T15:30:57 | 2018-05-07T15:30:57 | 127,908,833 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 3,819 | py | from re import sub
from os import scandir
from nltk.stem import SnowballStemmer
from Classifier import Classifier
def normalize_text(text):
return sub("[^a-z]+", ' ', text.lower())
def read_data(directory_name):
data = {}
with scandir(directory_name) as data_directory:
for class_entry in data_directory:
if class_entry.is_dir():
classification = class_entry.name
if classification not in data:
data[classification] = []
with scandir(class_entry.path) as class_directory:
for file_entry in class_directory:
if file_entry.is_file() and file_entry.name.endswith(".txt"):
with open(file_entry.path, 'r', encoding="utf8", errors="ignore") as file:
data[classification].append(normalize_text(file.read()))
return data
def read_data_python35(directory_name):
data = {}
data_directory = scandir(directory_name)
for class_entry in data_directory:
if class_entry.is_dir():
classification = class_entry.name
if classification not in data:
data[classification] = []
class_directory = scandir(class_entry.path)
for file_entry in class_directory:
if file_entry.is_file() and file_entry.name.endswith(".txt"):
file = open(file_entry.path, 'r', encoding="utf8", errors="ignore")
data[classification].append(normalize_text(file.read()))
file.close()
return data
def get_stop_words(stop_words_on):
if stop_words_on:
return [" ", "a", "about", "above", "after", "again", "against", "all", "am", "an", "and", "any", "are",
"arent", "as", "at", "be", "because", "been", "before", "being", "below", "between", "both", "but",
"by", "cant", "cannot", "could", "couldnt", "did", "didnt", "do", "does", "doesnt", "doing", "dont",
"down", "during", "each", "few", "for", "from", "further", "had", "hadnt", "has", "hasnt", "have",
"havent", "having", "he", "hed", "hell", "hes", "her", "here", "heres", "hers", "herself", "him",
"himself", "his", "how", "hows", "i", "id", "ill", "im", "ive", "if", "in", "into", "is", "isnt",
"it", "its", "its", "itself", "lets", "me", "more", "most", "mustnt", "my", "myself", "no", "nor",
"not", "of", "off", "on", "once", "only", "or", "other", "ought", "our", "ours", "ourselves", "out",
"over", "own", "same", "shant", "she", "shed", "shell", "shes", "should", "shouldnt", "so", "some",
"such", "than", "that", "thats", "the", "their", "theirs", "them", "themselves", "then", "there",
"theres", "these", "they", "theyd", "theyll", "theyre", "theyve", "this", "those", "through", "to",
"too", "under", "until", "up", "very", "was", "wasnt", "we", "wed", "well", "were", "weve", "werent",
"what", "whats", "when", "whens", "where", "wheres", "which", "while", "who", "whos", "whom", "why",
"whys", "with", "wont", "would", "wouldnt", "you", "youd", "youll", "youre", "youve", "your", "yours",
"yourself", "yourselves"]
else:
return [" "]
def data_to_classifiers(data, filter_stop_words):
classifiers = []
stop_words = get_stop_words(filter_stop_words)
stemmer = SnowballStemmer("english")
for classification, text_files in data.items():
for text in text_files:
classifier_new = Classifier(classification)
classifier_new.count_words(stop_words, stemmer, text)
classifiers.append(classifier_new)
return classifiers
| [
"harvc015@gmail.com"
] | harvc015@gmail.com |
40ce5ec818cd4be194c39a8e93b7069f16945d43 | 1ec2e018d63d15486110aea7923ffbbf62ecbd5b | /SVM_classification__.py | a6893ea17659bb15ef9207d6898ec74c58171c78 | [] | no_license | DDeman/svm | 88f51cefed530821e5a13df259602bf3c67f0128 | 743c6edcbd12a34ffc45dd43ef97c36792d308ac | refs/heads/master | 2021-05-24T17:04:42.727047 | 2020-04-07T02:39:47 | 2020-04-07T02:39:47 | 253,668,816 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 6,489 | py | #!/usr/bin/env python
# -*- coding: utf-8 -*-
"""
__title__ = ''
__author__ = '任晓光'
__mtime__ = '2020/4/3'
# code is far away from bugs with the god animal protecting
I love animals. They taste delicious.
┏┓ ┏┓
┏┛┻━━━┛┻┓
┃ ☃ ┃
┃ ┳┛ ┗┳ ┃
┃ ┻ ┃
┗━┓ ┏━┛
┃ ┗━━━┓
┃ 神兽保佑 ┣┓
┃ 永无BUG! ┏┛
┗┓┓┏━┳┓┏┛
┃┫┫ ┃┫┫
┗┻┛ ┗┻┛
"""
import numpy as np
import pandas as pd
from sklearn.datasets import load_breast_cancer
# np.set_printoptions(suppress=True)
class SVM__smo_simple():
def __init__(self):
pass
def select_j(self, i, m):
while True:
j = int(np.random.uniform(0, m))
if j != i:
break
return j
def calcute_L_H(self, i, j, C, y_array, alpha):
if y_array[i] != y_array[j]:
L = max(0, alpha[j] - alpha[i])
H = min(C, C + alpha[j] - alpha[i])
else:
L = max(0, alpha[i] + alpha[j] - C)
H = min(C, alpha[i] + alpha[j])
return L, H
def cat_alpha(self, L, H, alpha):
if alpha > H:
return H
elif alpha < L:
return L
else:
return alpha
def calcute_Ei(self,alpha,x_array,y_array,b):
for idx in range(len(alpha)):
gx = sum(np.dot(x_array,x_array[idx]) * y_array * alpha) + b
self.E[idx] = gx - y_array[idx]
def fit(self, data, C, toler, max_iters):
x_pd = data.iloc[:, :-1]
y_pd = data.iloc[:, -1]
m, n = x_pd.shape
alpha = np.zeros((m,))
b = 0
x_array = np.array(x_pd)
y_array = np.array(y_pd)
iters = 0
max_iters = max_iters
self.E = np.zeros((m,1))
self.calcute_Ei(alpha,x_array,y_array,b)
kk = 0
while iters < max_iters:
i = self.select_i(alpha, y_array, x_array, C, b)
E_i = self.E[i]
# print((E_i))
# b = alpha[i]
if ((y_array[i] * E_i < - toler) and (alpha[i] < C)) or ((y_array[i] * E_i > toler) and (alpha[i] > 0)):
j = self.select_j(i, m)
E_j = sum(np.dot(x_array, x_array[j]) * y_array * alpha) + b - y_array[j]
L, H = self.calcute_L_H(i, j, C, y_array, alpha)
if L == H:
continue
eta = np.dot(x_array[i], x_array[i]) + np.dot(x_array[j], x_array[j]) - 2 * np.dot(x_array[i],
x_array[j])
if eta <= 0:
continue
alpha_j_new_unc = alpha[j] + y_array[j] * (E_i - E_j) / eta
alpha_j_new = self.cat_alpha(L, H, alpha_j_new_unc)
alpha_i_new = alpha[i] + y_array[i] * y_array[j] * (alpha[j] - alpha_j_new)
bi_new = b - E_i - y_array[i] * (alpha_i_new - alpha[i]) * np.dot(x_array[i], x_array[i]) - y_array[
j] * (
alpha_j_new -
alpha[
j]) * np.dot(
x_array[j], x_array[i])
bj_new = b - E_j - y_array[i] * (alpha_i_new - alpha[i]) * np.dot(x_array[i], x_array[j]) - y_array[
j] * (
alpha_j_new -
alpha[
j]) * np.dot(
x_array[j], x_array[j])
if 0 < alpha_i_new < C:
b = bi_new
elif 0 < alpha_j_new < C:
b = bj_new
else:
b = (bi_new + bj_new) / 2
alpha[i], alpha[j] = alpha_i_new, alpha_j_new
# print(alpha)
iters += 1
print('iters is :', iters)
# print(alpha)
self.w = np.dot(alpha * y_array, x_array)
j = None
for i in range(m):
if alpha[i] > 0:
j = i
continue
self.b = y_array[j] - alpha * y_array * np.dot(x_array, x_array[j])
return self.w, self.b
def predict(self, x_array):
pred = np.dot(x_array, self.w) + self.b
return pred
def select_i(self, alpha, y_array, x_array, C, b):
for idx in range(len(alpha)):
gx = sum(np.dot(x_array,x_array[idx]) * y_array * alpha) + b
if alpha[idx] == 0:
if y_array[idx] * gx < 1:
return idx
elif 0 < alpha[idx] < C:
if y_array[idx] * gx != 1:
return idx
elif alpha[idx] == C:
if y_array[idx] * gx > 1:
return idx
i = np.random.uniform(0, len(alpha))
return alpha[i]
if __name__ == '__main__':
data = load_breast_cancer()
x, y = data.data, data.target
x_pd = pd.DataFrame(x, columns=data.feature_names)
y_pd = pd.DataFrame(y, columns=['result']).replace([0, 1], [-1, 1])
data_pd = pd.concat([x_pd, y_pd], axis=1)
svm = SVM__smo_simple()
w, b = svm.fit(data_pd, 0.6, 0.001, 5000)
x_array = np.array(x_pd)
y_pred = svm.predict(x_array)
print(y_pred)
y_p = []
for i in y_pred:
if i > 0:
y_p.append(1)
else:
y_p.append(-1)
from sklearn.metrics import accuracy_score
print(y_p)
print(np.array(y_pd).tolist())
acc = accuracy_score(y_pd, y_p)
print(acc)
| [
"rxg15506009565"
] | rxg15506009565 |
e8d0823a474e0ac46065eca508bc73d8188113d9 | 29c4f16b2bd95203fc58f1b43ada634116aabb8d | /Customer.py | fa14e037618593e8a9bba2823b3fae772e9ee149 | [] | no_license | AlexandreGheraibia/banquePython | 2c715c7013c8cd89f87ce80fcadb63ce47830b76 | df2f1b79f8dd5219f65bd5cf0b9a249a12e11caa | refs/heads/master | 2020-03-22T01:06:23.046419 | 2018-06-30T22:27:29 | 2018-06-30T22:27:29 | 139,283,290 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 320 | py | class Customer:
def setId(this,id):
return id
def getId(this):
return this.id
def setName(this,name):
return name
def getName(this):
return this.name
def __init__(this):
return
def __init__(this,id,name):
this.id=id
this.name=name
| [
"gheraibia@hotmail.com"
] | gheraibia@hotmail.com |
f7ee63e6b92678782ec9da34b96b0addaf69997c | b9571590d8cc83a99293d777f57e5ebeea5bcc92 | /spiders/DoctorSpider.py | 1cc8539b8017fa62c7ea2ce5c7a731be27f7fec8 | [] | no_license | LiuQL2/Crawler_xywy_doctor_communication | 585a0a3230f397640e5fc54506cd6585bfd04f57 | 3374f08ea34ae8ea7e96501188a4fec247c72b5d | refs/heads/master | 2020-06-30T13:28:01.048195 | 2017-08-04T07:29:19 | 2017-08-04T07:29:19 | 74,369,626 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 1,508 | py | #! /usr/bin/env python
# -*- coding: utf-8 -*-
"""
用来获取病例和心得帖子内容的类,传入一个帖子的URL,调用不同的方法得到不同的数据。
"""
# Author: Liu Qianlong <LiuQL2@163.com>
# Date: 2016.12.08
import datetime
import json
import sys
import urllib2
from BaseSpider import BaseSpider
reload(sys)
sys.setdefaultencoding('utf-8')
class DoctorSpider(BaseSpider):
def __init__(self,url, crawl_number, try_number = 20):
self.target_url = url
request = urllib2.Request(url=self.target_url, headers=self.get_header())
self.status = True
self.try_number = try_number
self.crawl_number = crawl_number
self.selector = None
self.number_url = 'http://club.xywy.com/doctorShare/index.php?type=share_operation&uid=' + self.target_url.split('/')[4] + '&stat=14'
def get_number(self):
doc = self.process_url_request(self.number_url,xpath_type=False)
if doc != None:
doc = json.loads(doc)
crawl_time = datetime.datetime.now().strftime('%Y-%m-%d')
return {'attention_number':str(doc['attenNum']), 'fans_number':str(doc['fansNum']),'web_number':str(doc['wbNum']),'doctor_url':self.target_url, 'crawl_time':crawl_time, 'crawl_number':self.crawl_number}
else:
return None
if __name__ == '__main__':
doctor = DoctorSpider(url='http://club.xywy.com/doc_card/55316663/blog')
print doctor.get_number() | [
"LiuQL2@sina.com"
] | LiuQL2@sina.com |
0fe08899b3a8f27f944baf7bfb39b3fcdf8ebdff | f576f0ea3725d54bd2551883901b25b863fe6688 | /sdk/synapse/azure-synapse-accesscontrol/azure/synapse/accesscontrol/aio/__init__.py | 8eafa989fcbc836fcc407acd2ea0859726442db7 | [
"LicenseRef-scancode-generic-cla",
"MIT",
"LGPL-2.1-or-later"
] | permissive | Azure/azure-sdk-for-python | 02e3838e53a33d8ba27e9bcc22bd84e790e4ca7c | c2ca191e736bb06bfbbbc9493e8325763ba990bb | refs/heads/main | 2023-09-06T09:30:13.135012 | 2023-09-06T01:08:06 | 2023-09-06T01:08:06 | 4,127,088 | 4,046 | 2,755 | MIT | 2023-09-14T21:48:49 | 2012-04-24T16:46:12 | Python | UTF-8 | Python | false | false | 558 | py | # coding=utf-8
# --------------------------------------------------------------------------
# Copyright (c) Microsoft Corporation. All rights reserved.
# Licensed under the MIT License. See License.txt in the project root for license information.
# Code generated by Microsoft (R) AutoRest Code Generator.
# Changes may cause incorrect behavior and will be lost if the code is regenerated.
# --------------------------------------------------------------------------
from ._access_control_client import AccessControlClient
__all__ = ['AccessControlClient']
| [
"noreply@github.com"
] | noreply@github.com |
9691a7aad57450dba8acad44cb0105ff32ac9d5c | c171d1b819d599c294afe67d17f09f7ef5e358af | /Blind_Search/8_Puzzle_DFS.py | 088de9e8ec7e148eb1df2ea660d39d7c67ccbbe2 | [] | no_license | bawejagb/Artificial_Intelligence | 0e18e9f9cb0ee2663ae5d3ffd4ad9b2560295947 | 9caafe026e93cd50a0d01d3d83e079ce221c5a9d | refs/heads/main | 2023-04-21T19:20:49.713415 | 2021-05-04T15:54:48 | 2021-05-04T15:54:48 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 2,840 | py | '''
Q1- 8 puzzle problem
Made By: Gaurav Baweja, 102097005, CSE4
'''
import copy as cp
def Show(arr):
print("------")
for lis in arr:
for elm in lis:
print(elm,end="|")
print()
print("------")
def Position(val,arr):
for i in range(len(arr)):
for j in range(len(arr[i])):
if(arr[i][j] == val):
return (i,j)
def Swap(i1,j1,i2,j2,arr):
temp = arr[i1][j1]
arr[i1][j1] = arr[i2][j2]
arr[i2][j2] = temp
def MoveUp(i,j,arr):
if(i < 1):
return arr
temp = cp.deepcopy(arr)
Swap(i,j,i-1,j,temp)
return temp
def MoveLeft(i,j,arr):
if(j < 1):
return arr
temp = cp.deepcopy(arr)
Swap(i,j,i,j-1,temp)
return temp
def MoveRight(i,j,arr):
if(j > 1):
return arr
temp = cp.deepcopy(arr)
Swap(i,j,i,j+1,temp)
return temp
def MoveDown(i,j,arr):
if(i > 1):
return arr
temp = cp.deepcopy(arr)
Swap(i,j,i+1,j,temp)
return temp
def Compare(arr1, arr2):
if(arr1 == arr2):
return True
return False
def enqueue(que, arr):
que.append(arr)
def dequeue(que):
if(len(que) != 0):
del que[-1]
def front(que):
if(len(que) > 0):
return que[0]
def end(que):
if(len(que) > 0):
return que[-1]
def DFS(start, goal):
itr_count = 0
queue = []
visited = []
enqueue(queue, start)
while(len(queue) != 0):
itr_count += 1
temp = end(queue)
#Show(temp)
dequeue(queue)
visited.append((temp))
row,col = Position(0,temp) # Check Position of Empty Slide(0)
for state in range(1,5):
if(state == 1): #MoveUp
nextState = MoveUp(row,col,temp)
if(state == 2): #MoveDown
nextState = MoveDown(row,col,temp)
if(state == 3): #MoveLeft
nextState = MoveLeft(row,col,temp)
if(state == 4): #MoveRight
nextState = MoveRight(row,col,temp)
if(nextState == goal): # Check Goal State
print("Achieved Goal State:")
print("Total Iteration: ", itr_count)
Show(nextState)
return True
if(nextState not in queue and
nextState not in visited): #Enqueue
enqueue(queue, nextState)
return False
if (__name__ == "__main__"):
#Start
""" StartState = [[2,0,3],
[1,8,4],
[7,6,5]]
GoalState = [[1,2,3],[8,0,4],[7,6,5]] """
StartState = [[2,0,3],
[1,8,4],
[7,6,5]]
GoalState = [[1,2,3],
[8,0,4],
[7,6,5]]
status = DFS(StartState, GoalState)
print("State Possible: ",end = "")
if(status):
print("Yes")
else:
print("No")
| [
"gaurav.baweja2508@gmail.com"
] | gaurav.baweja2508@gmail.com |
3f80f32810fc412f915873a0c635f787b0603cd6 | 7f07515310c075c95033354a91f9f82557b98092 | /heatmap.py | cb8fd55587cb71d1707428b4f0ae92759ce9cb53 | [] | no_license | HegemanLab/VanKrevelen | 805d33e9e0515a1250fbb87f27b7d56af1de759f | ea82f284f3ade1b43adb6bc5041b5bb14c166c2f | refs/heads/master | 2020-04-16T23:39:22.636160 | 2016-08-17T17:54:16 | 2016-08-17T17:54:16 | 47,356,866 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 7,028 | py | '''
Thanks to jjguy @ http://jjguy.com/heatmap/
Minimal edits made to the original code, mostly just comments
'''
import ctypes
import os
import platform
import sys
from PIL import Image
import colorschemes
class Heatmap:
"""
Create heatmaps from a list of 2D coordinates.
Heatmap requires the Python Imaging Library and Python 2.5+ for ctypes.
Coordinates autoscale to fit within the image dimensions, so if there are
anomalies or outliers in your dataset, results won't be what you expect. You
can override the autoscaling by using the area parameter to specify the data bounds.
The output is a PNG with transparent background, suitable alone or to overlay another
image or such. You can also save a KML file to use in Google Maps if x/y coordinates
are lat/long coordinates. Make your own wardriving maps or visualize the footprint of
your wireless network.
Most of the magic starts in heatmap(), see below for description of that function.
"""
KML = """<?xml version="1.0" encoding="UTF-8"?>
<kml xmlns="http://www.opengis.net/kml/2.2">
<Folder>
<GroundOverlay>
<Icon>
<href>%s</href>
</Icon>
<LatLonBox>
<north>%2.16f</north>
<south>%2.16f</south>
<east>%2.16f</east>
<west>%2.16f</west>
<rotation>0</rotation>
</LatLonBox>
</GroundOverlay>
</Folder>
</kml>"""
def __init__(self, libpath=None):
self.minXY = ()
self.maxXY = ()
self.img = None
if libpath:
self._heatmap = ctypes.cdll.LoadLibrary(libpath)
else:
# establish the right library name, based on platform and arch. Windows
# are pre-compiled binaries; linux machines are compiled during setup.
self._heatmap = None
libname = "cHeatmap.so"
if "cygwin" in platform.system().lower():
libname = "cHeatmap.dll"
if "windows" in platform.system().lower():
libname = "cHeatmap-x86.dll"
if "64" in platform.architecture()[0]:
libname = "cHeatmap-x64.dll"
# now rip through everything in sys.path to find them. Should be in site-packages
# or local dir
for d in sys.path:
if os.path.isfile(os.path.join(d, libname)):
self._heatmap = ctypes.cdll.LoadLibrary(
os.path.join(d, libname))
if not self._heatmap:
raise Exception("Heatmap shared library not found in PYTHONPATH.")
def heatmap(self, points, dotsize=150, opacity=128, size=(1024, 1024), scheme="classic", area=None):
"""
points -> an iterable list of tuples, where the contents are the
x,y coordinates to plot. e.g., [(1, 1), (2, 2), (3, 3)]
dotsize -> the size of a single coordinate in the output image in
pixels, default is 150px. Tweak this parameter to adjust
the resulting heatmap.
opacity -> the strength of a single coordiniate in the output image.
Tweak this parameter to adjust the resulting heatmap.
size -> tuple with the width, height in pixels of the output PNG
scheme -> Name of color scheme to use to color the output image.
Use schemes() to get list. (images are in source distro)
area -> Specify bounding coordinates of the output image. Tuple of
tuples: ((minX, minY), (maxX, maxY)). If None or unspecified,
these values are calculated based on the input data.
"""
self.dotsize = dotsize
self.opacity = opacity
self.size = size
self.points = points
if area is not None:
self.area = area
self.override = 1
else:
self.area = ((0, 0), (0, 0))
self.override = 0
if scheme not in self.schemes():
tmp = "Unknown color scheme: %s. Available schemes: %s" % (
scheme, self.schemes())
raise Exception(tmp)
arrPoints = self._convertPoints(points)
arrScheme = self._convertScheme(scheme)
arrFinalImage = self._allocOutputBuffer()
ret = self._heatmap.tx(
arrPoints, len(points) * 2, size[0], size[1], dotsize,
arrScheme, arrFinalImage, opacity, self.override,
ctypes.c_float(self.area[0][0]), ctypes.c_float(
self.area[0][1]),
ctypes.c_float(self.area[1][0]), ctypes.c_float(self.area[1][1]))
if not ret:
raise Exception("Unexpected error during processing.")
self.img = Image.frombuffer('RGBA', (self.size[0], self.size[1]),
arrFinalImage, 'raw', 'RGBA', 0, 1)
return self.img
def _allocOutputBuffer(self):
return (ctypes.c_ubyte * (self.size[0] * self.size[1] * 4))()
def _convertPoints(self, pts):
""" flatten the list of tuples, convert into ctypes array """
flat = []
for i, j in pts:
flat.append(i)
flat.append(j)
# Build array of input points
arr_pts = (ctypes.c_float * (len(pts) * 2))(*flat)
return arr_pts
def _convertScheme(self, scheme):
""" flatten the list of RGB tuples, convert into ctypes array """
flat = []
for r, g, b in colorschemes.schemes[scheme]:
flat.append(r)
flat.append(g)
flat.append(b)
arr_cs = (
ctypes.c_int * (len(colorschemes.schemes[scheme]) * 3))(*flat)
return arr_cs
def _ranges(self, points):
""" walks the list of points and finds the
max/min x & y values in the set """
minX = points[0][0]
minY = points[0][1]
maxX = minX
maxY = minY
for x, y in points:
minX = min(x, minX)
minY = min(y, minY)
maxX = max(x, maxX)
maxY = max(y, maxY)
return ((minX, minY), (maxX, maxY))
def saveKML(self, kmlFile):
"""
Saves a KML template to use with google earth. Assumes x/y coordinates
are lat/long, and creates an overlay to display the heatmap within Google
Earth.
kmlFile -> output filename for the KML.
"""
if self.img is None:
raise Exception("Must first run heatmap() to generate image file.")
tilePath = os.path.splitext(kmlFile)[0] + ".png"
self.img.save(tilePath)
if self.override:
((east, south), (west, north)) = self.area
else:
((east, south), (west, north)) = self._ranges(self.points)
bytes = self.KML % (tilePath, north, south, east, west)
file(kmlFile, "w").write(bytes)
def schemes(self):
"""
Return a list of available color scheme names.
"""
return colorschemes.valid_schemes()
| [
"roden026@umn.edu"
] | roden026@umn.edu |
cf3970f95cd0df134c66e7e2c608fab1e79e582a | 0031bd210e25f9602a8ee3cf581c44e8e8f3a00f | /Junior/COSC0023-Py/Exercise/数据画图.py | a0a7647e524821d9fb8ef041ddf829e2fe99deef | [
"MIT"
] | permissive | TiffanyChou21/University | d991d30cad3b28bb5abc929faa6d530219a1d844 | 9584fa6b052a59ce01a256efc77add5bbec68d98 | refs/heads/master | 2020-09-29T10:54:00.297491 | 2020-08-16T03:47:57 | 2020-08-16T03:47:57 | 227,021,880 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 2,043 | py | #!/usr/bin/env python
# coding: utf-8
# In[52]:
import numpy as np
import pandas as pd
import matplotlib.pyplot as plt
import warnings
# plt.rcParams['font.sans-serif'] = [u'SimHei'] #显示不了中文放弃了不显示了
# plt.rcParams['axes.unicode_minus'] = False
warnings.filterwarnings('ignore')
pd.set_option('display.max_rows',None)
iris=pd.read_csv('iris.csv',names=['花萼长度','花萼宽度','花瓣长度','花瓣宽度','类别'])
# ## 作业一
# 画出花萼长度和花萼宽度的散点图
# In[19]:
plt.xlabel('length')
plt.ylabel('width')
plt.scatter(iris['花萼长度'],iris['花萼宽度'],c='r',marker='.')
# ## 作业二
# 按照花萼长度排序作为X轴,画出花萼宽度随着花萼长度变化的折线图,包括图表标题、轴标签、刻度等
# In[20]:
plt.xlabel('length')
plt.ylabel('width')
plt.title("width-length")
plt.plot(iris.sort_index(by='花萼长度')['花萼长度'],iris['花萼宽度'],'r')
# ## 作业三
# 画出花瓣长度和花瓣宽度的散点图,要求不同类别花样本点的颜色不同。
# In[42]:
plt.xlabel('length')
plt.ylabel('width')
plt.scatter(iris['花瓣长度'], iris['花瓣宽度'], c=iris['类别'])
# ## 作业四
# 计算每个特征的平均值,画出直方图
# In[30]:
iris1=iris.drop(['类别'],axis=1)
m=np.array(iris1.mean())
# In[45]:
# labels = ['花萼长度', '花萼宽度', '花瓣长度', '花瓣宽度']#中文显示有问题花萼对应s 花瓣对应f
labels = ['slength', 'swidth', 'flength', 'fwidth']
plt.bar(np.arange(4)+1,m,color='c',tick_label=labels)
for x, y in zip(np.arange(4)+1, m):
plt.text(x , y, '%.2f' % y, ha='center', va='bottom')
# ## 作业五
# 计算每种花的样本数量百分比,画出饼状图。
# In[67]:
count_df = iris.groupby('类别').count()
test_df = pd.DataFrame(count_df)
perc=test_df/test_df.sum()
perc=perc.drop(['花萼宽度','花瓣长度','花瓣宽度'],axis=1)
perc=np.array(perc)
# In[70]:
plt.pie(perc,labels=['0','1','2'],autopct='%1.1f')
| [
"TiffanyChou21@163.com"
] | TiffanyChou21@163.com |
ee391734bbe1d920f7349971047cc74c0c565f36 | e9ef3cd143478660d098668a10e67544a42b5878 | /Lib/corpuscrawler/crawl_mpx.py | 71bb3a7ee49333cc9c4fc1cee863a89f398c5aa2 | [
"Apache-2.0"
] | permissive | google/corpuscrawler | a5c790c19b26e6397b768ce26cf12bbcb641eb90 | 10adaecf4ed5a7d0557c8e692c186023746eb001 | refs/heads/master | 2023-08-26T04:15:59.036883 | 2022-04-20T08:18:11 | 2022-04-20T08:18:11 | 102,909,145 | 119 | 40 | NOASSERTION | 2022-04-20T08:18:12 | 2017-09-08T22:21:03 | Python | UTF-8 | Python | false | false | 799 | py | # coding: utf-8
# Copyright 2017 Google LLC
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
from __future__ import absolute_import, print_function, unicode_literals
import re
def crawl(crawler):
out = crawler.get_output(language='mpx')
crawler.crawl_pngscriptures_org(out, language='mpx')
| [
"sascha@brawer.ch"
] | sascha@brawer.ch |
87eea3930704d7e2f8216d0c4e219c57beb148a0 | f6973cc740981bf4ce80b5d1dc8e12f90d38ef42 | /XDG_CACHE_HOME/Microsoft/Python Language Server/stubs.v1/sOkxR-HicgX79sgQZAyBR1W09XbOY-yF9DMdXi8nVww=/rwobject.cpython-36m-x86_64-linux-gnu.pyi | fdf731a3699e1443db437aea28576483fb0d807e | [] | no_license | fdr896/Trumper-Jumper | 88783cf97979a0e9f8d2a3f64c606a67d4dd1719 | 77f813b7267451f3156cf6bddad76081c29a25f0 | refs/heads/master | 2020-06-02T08:41:50.978639 | 2019-06-10T05:36:24 | 2019-06-10T05:36:24 | 191,102,422 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 707 | pyi | pygame 1.9.4
Hello from the pygame community. https://www.pygame.org/contribute.html
import builtins as _mod_builtins
_PYGAME_C_API = _mod_builtins.PyCapsule()
__doc__ = 'SDL_RWops support'
__file__ = '/home/fed/.local/lib/python3.6/site-packages/pygame/rwobject.cpython-36m-x86_64-linux-gnu.so'
__name__ = 'pygame.rwobject'
__package__ = 'pygame'
def encode_file_path(obj=None, etype=None):
'encode_file_path([obj [, etype]]) -> bytes or None\nEncode a Unicode or bytes object as a file system path'
pass
def encode_string(obj=None, encoding=None, errors=None, etype=None):
'encode_string([obj [, encoding [, errors [, etype]]]]) -> bytes or None\nEncode a Unicode or bytes object'
pass
| [
"fiosetrova@gmail.com"
] | fiosetrova@gmail.com |
f19728bf8e40186ce8d1a8b60514928cac527607 | 51a48a4144ef3aabb354e75033293da45a813c8f | /sourceCodes/python/pfile/wrapper.py | b7c9f99060862916eef3125a38ab52e2f6a0a2ff | [] | no_license | melodi-lab/SGM | a53f79c8a0e4cbc7dd64f269002b6a0d49cd5d92 | 36d9b7b6f864ee93220c960065f8c7e216c5009c | refs/heads/master | 2020-04-05T23:11:23.626090 | 2017-06-19T23:30:40 | 2017-06-19T23:30:40 | 64,819,641 | 1 | 2 | null | null | null | null | UTF-8 | Python | false | false | 4,343 | py | #!/usr/bin/env python
#
# Copyright 2011 <fill in later>
__authors__ = [ 'Ajit Singh <ajit@ee.washington.edu>' ]
import os
import sys
import types
try:
import libpfile as lib
except ImportError:
dirname = os.path.dirname(__file__)
print >> sys.stderr, "Run make in %s to compile libpfile" % dirname
exit(-1)
class PFile(object):
def __init__(self, nf, ni, f, doswap = None):
"""Constructor
Creates a PFile with the name specified in fn. Each segment contains
nf floats, followed by ni integers.
Arguments:
nf: Number of floats in a segment.
ni: Number of integers (labels) in a segment.
fn: Name of the pfile to create, no extension is forced.
doswap: If set, it will force a particular byte order in the
generated PFile. Useful if you're writing a PFile on one
platform for use on another platform. If None, use whatever
sys.byteorder returns.
"""
if not doswap:
if sys.byteorder == 'little':
doswap = 1
elif sys.byteorder == 'big':
doswap = 0
else:
raise Exception("Could not infer byteorder.")
index = 1
if type(f) == types.FileType:
self.f = f
elif type(f) == types.StringType:
self.f = open(f, 'w')
else:
raise Exception("Bad filename argument: %s" % str(f))
self.nf = nf
self.ni = ni
self.doswap = doswap
self.pf = lib.OutFtrLabStream_PFile(0, '', self.f, nf, ni, index,
doswap)
# Create buffers for translating Python lists of floats or ints to
# float* and unsigneed int*
self.buf_floats = lib.new_doubleArray(self.nf)
self.buf_ints = lib.new_uintArray(self.ni)
def __del__(self):
"""Destructor.
TODO(ajit): Calling pfile.fclose of self.pf causes a segmentation fault.
Determine where the file is really being deleted (it may only be on
exit, or deletion of the class).
"""
del self.pf
lib.delete_doubleArray(self.buf_floats)
lib.delete_uintArray(self.buf_ints)
@property
def name(self):
return self.f.name
def check_frame(self, *args):
if len(args) != self.nf + self.ni:
raise Exception("Wrong length %d vs. %d" % (len(args),
self.nf + self.ni))
for i in xrange(0, self.nf, 1):
if not type(args[i]) == types.FloatType:
raise Exception("Wrong type arg[%d]: wanted float, got %s" % (
i, str(type(args[i]))))
for i in xrange(self.nf, self.nf+self.ni, 1):
if not type(args[i]) == types.IntType:
raise Exception("Wrong type arg[%d]: wanted int, got %s" % (
i, str(type(args[i]))))
def add_frame(self, *args):
for i in xrange(0, self.nf, 1):
lib.doubleArray_setitem(self.buf_floats, i, args[i])
for i in xrange(self.nf, self.nf + self.ni, 1):
lib.uintArray_setitem(self.buf_ints, i-self.nf, args[i])
self.pf.write_ftrslabs(1, self.buf_floats, self.buf_ints)
def add_segment(self, nframes, floats, ints):
"""Copy a whole sentence in one shot.
Can be useful in reducing the Python -> C++ overhead required to
generate one sentence: e.g., creating one list for all the floats
in a sentence, instead of one list per-frame. You do not need to
call end_segment after using this function.
TODO(ajit): It's not clear whether the segment ID is actually used
anywhere. The code in pfile.cc:doneseg does not appear to use the
segment ID, and ignoring it doesn't seem to cause any problems.
Arguments:
nframes: Number of frames in the sentence.
floats: Iterable with all of the floats in the sentence. First,
all the floats in frame 0, then frame 1, etc.
ints: Iterable with all the integers in the sentence.
"""
pass
def end_segment(self, i = None):
if not i:
i = lib.SEGID_UNKNOWN
self.pf.doneseg(i)
| [
"baiwenruo@gmail.com"
] | baiwenruo@gmail.com |
f8597c8ce3dfbc755d8bf76575047963a0ec8beb | 6c74c8babd2f94cbed185af75940774a2750f3e5 | /src/georinex/base.py | ccfff852795a64c572afd92589a410550c92cf2e | [
"MIT"
] | permissive | geospace-code/georinex | c28c8a17196bb1fa8093c818ce43bcb74ec52171 | c689a5a6bc2ffb68bc055f150f1da1b6bab12812 | refs/heads/main | 2023-04-13T15:01:50.903458 | 2022-12-27T19:25:58 | 2022-12-27T19:26:15 | 34,296,204 | 106 | 40 | MIT | 2023-04-10T02:54:45 | 2015-04-21T01:19:29 | Python | UTF-8 | Python | false | false | 7,148 | py | from __future__ import annotations
import typing as T
from pathlib import Path
import xarray
from datetime import datetime, timedelta
import logging
from .rio import rinexinfo
from .obs2 import rinexobs2
from .obs3 import rinexobs3
from .nav2 import rinexnav2
from .nav3 import rinexnav3
from .sp3 import load_sp3
from .utils import _tlim
# for NetCDF compression. too high slows down with little space savings.
ENC = {"zlib": True, "complevel": 1, "fletcher32": True}
def load(
rinexfn: T.TextIO | str | Path,
out: Path = None,
use: set[str] = None,
tlim: tuple[datetime, datetime] = None,
useindicators: bool = False,
meas: list[str] = None,
verbose: bool = False,
*,
overwrite: bool = False,
fast: bool = True,
interval: float | int | timedelta = None,
):
"""
Reads OBS, NAV in RINEX 2.x and 3.x
Files / StringIO input may be plain ASCII text or compressed (including Hatanaka)
"""
if verbose:
logging.basicConfig(level=logging.INFO)
if isinstance(rinexfn, (str, Path)):
rinexfn = Path(rinexfn).expanduser()
# %% determine if/where to write NetCDF4/HDF5 output
outfn = None
if out:
out = Path(out).expanduser()
if out.is_dir():
outfn = out / (
rinexfn.name + ".nc"
) # not with_suffix to keep unique RINEX 2 filenames
elif out.suffix == ".nc":
outfn = out
else:
raise ValueError(f"not sure what output is wanted: {out}")
# %% main program
if tlim is not None:
if len(tlim) != 2:
raise ValueError("time bounds are specified as start stop")
if tlim[1] < tlim[0]:
raise ValueError("stop time must be after start time")
info = rinexinfo(rinexfn)
if info["rinextype"] == "nav":
return rinexnav(rinexfn, outfn, use=use, tlim=tlim, overwrite=overwrite)
elif info["rinextype"] == "obs":
return rinexobs(
rinexfn,
outfn,
use=use,
tlim=tlim,
useindicators=useindicators,
meas=meas,
verbose=verbose,
overwrite=overwrite,
fast=fast,
interval=interval,
)
assert isinstance(rinexfn, Path)
if info["rinextype"] == "sp3":
return load_sp3(rinexfn, outfn)
elif rinexfn.suffix == ".nc":
# outfn not used here, because we already have the converted file!
try:
nav = rinexnav(rinexfn)
except LookupError:
nav = None
try:
obs = rinexobs(rinexfn)
except LookupError:
obs = None
if nav is not None and obs is not None:
return {"nav": nav, "obs": rinexobs(rinexfn)}
elif nav is not None:
return nav
elif obs is not None:
return obs
else:
raise ValueError(f"No data of known format found in {rinexfn}")
else:
raise ValueError(f"What kind of RINEX file is: {rinexfn}")
def batch_convert(
path: Path,
glob: str,
out: Path,
use: set[str] = None,
tlim: tuple[datetime, datetime] = None,
useindicators: bool = False,
meas: list[str] = None,
verbose: bool = False,
*,
fast: bool = True,
):
path = Path(path).expanduser()
flist = (f for f in path.glob(glob) if f.is_file())
for fn in flist:
try:
load(
fn,
out,
use=use,
tlim=tlim,
useindicators=useindicators,
meas=meas,
verbose=verbose,
fast=fast,
)
except ValueError as e:
logging.error(f"{fn.name}: {e}")
def rinexnav(
fn: T.TextIO | str | Path,
outfn: Path = None,
use: set[str] = None,
group: str = "NAV",
tlim: tuple[datetime, datetime] = None,
*,
overwrite: bool = False,
) -> xarray.Dataset:
"""Read RINEX 2 or 3 NAV files"""
if isinstance(fn, (str, Path)):
fn = Path(fn).expanduser()
if fn.suffix == ".nc":
try:
return xarray.open_dataset(fn, group=group)
except OSError as e:
raise LookupError(f"Group {group} not found in {fn} {e}")
tlim = _tlim(tlim)
info = rinexinfo(fn)
if int(info["version"]) == 2:
nav = rinexnav2(fn, tlim=tlim)
elif int(info["version"]) == 3:
nav = rinexnav3(fn, use=use, tlim=tlim)
else:
raise LookupError(f"unknown RINEX {info} {fn}")
# %% optional output write
if outfn:
outfn = Path(outfn).expanduser()
wmode = _groupexists(outfn, group, overwrite)
enc = {k: ENC for k in nav.data_vars}
nav.to_netcdf(outfn, group=group, mode=wmode, encoding=enc)
return nav
# %% Observation File
def rinexobs(
fn: T.TextIO | Path,
outfn: Path = None,
use: set[str] = None,
group: str = "OBS",
tlim: tuple[datetime, datetime] = None,
useindicators: bool = False,
meas: list[str] = None,
verbose: bool = False,
*,
overwrite: bool = False,
fast: bool = True,
interval: float | int | timedelta = None,
):
"""
Read RINEX 2.x and 3.x OBS files in ASCII or GZIP (or Hatanaka)
"""
if isinstance(fn, (str, Path)):
fn = Path(fn).expanduser()
# %% NetCDF4
if fn.suffix == ".nc":
try:
return xarray.open_dataset(fn, group=group)
except OSError as e:
raise LookupError(f"Group {group} not found in {fn} {e}")
tlim = _tlim(tlim)
# %% version selection
info = rinexinfo(fn)
if int(info["version"]) in (1, 2):
obs = rinexobs2(
fn,
use,
tlim=tlim,
useindicators=useindicators,
meas=meas,
verbose=verbose,
fast=fast,
interval=interval,
)
elif int(info["version"]) == 3:
obs = rinexobs3(
fn,
use,
tlim=tlim,
useindicators=useindicators,
meas=meas,
verbose=verbose,
fast=fast,
interval=interval,
)
else:
raise ValueError(f"unknown RINEX {info} {fn}")
# %% optional output write
if outfn:
outfn = Path(outfn).expanduser()
wmode = _groupexists(outfn, group, overwrite)
enc = {k: ENC for k in obs.data_vars}
# Pandas >= 0.25.0 requires this, regardless of xarray version
if obs.time.dtype != "datetime64[ns]":
obs["time"] = obs.time.astype("datetime64[ns]")
obs.to_netcdf(outfn, group=group, mode=wmode, encoding=enc)
return obs
def _groupexists(fn: Path, group: str, overwrite: bool) -> str:
print(f"saving {group}:", fn)
if overwrite or not fn.is_file():
return "w"
# be sure there isn't already NAV in it
try:
xarray.open_dataset(fn, group=group)
raise ValueError(f"{group} already in {fn}")
except OSError:
pass
return "a"
| [
"scivision@users.noreply.github.com"
] | scivision@users.noreply.github.com |
e0c09849f0aec5951bf94adaa9bc3656ac75f05f | abc72a2f2072ab7a5a338e41d81c354324943b09 | /MC 102 (Exemplos de aula)/eliminar_repeticao.py | 55c15d25c81d25f12a60900b67da3c9af6354681 | [] | no_license | gigennari/mc102 | a3d39fd9a942c97ef477a9b59d7955f4269b202a | fce680d5188a8dfb0bc1832d6f430cbcaf68ef55 | refs/heads/master | 2023-04-05T01:40:58.839889 | 2020-07-27T20:33:56 | 2020-07-27T20:33:56 | 354,130,720 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 453 | py |
def eliminar_repeticao(lista1, lista2):
lista_sem_rep = []
freq_sem_rep = []
for i in range(len(lista1)):
if lista1[i] not in lista_sem_rep:
lista_sem_rep.append(lista1[i])
freq_sem_rep.append(lista2[i])
return lista_sem_rep, freq_sem_rep
def main():
lista1 = [3, 3, 6, 5, 8, 8, 10]
lista2 = [2, 2, 1, 1, 2, 2, 1]
lista3, lista4 = eliminar_repeticao(lista1, lista2)
print(lista3)
main() | [
"g198010@dac.unicamp.br"
] | g198010@dac.unicamp.br |
b48d0fdf80d00e5128a64d51260df0314579ca35 | 4160b450b052830e17457a0412e29414f67caea5 | /order/migrations/0010_auto_20210822_0755.py | 295b200f20b97da9b7772cdca4c5ea7279ce3aaa | [] | no_license | mnogoruk/fastcustoms | 6ad7b058607ddf4d2b56a09e23e66fcfb43be1a7 | 4c3bf7f9f1d4af2851f957a084b6adc2b7b7f681 | refs/heads/master | 2023-08-23T15:54:08.415613 | 2021-10-31T12:21:29 | 2021-10-31T12:21:29 | 372,066,847 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 962 | py | # Generated by Django 3.2.3 on 2021-08-22 07:55
from django.db import migrations, models
class Migration(migrations.Migration):
dependencies = [
('order', '0009_auto_20210821_2220'),
]
operations = [
migrations.AlterField(
model_name='orderagent',
name='comment',
field=models.TextField(blank=True, default='', max_length=1000),
),
migrations.AlterField(
model_name='orderagent',
name='company_name',
field=models.CharField(blank=True, max_length=250, null=True),
),
migrations.AlterField(
model_name='orderagent',
name='email',
field=models.EmailField(blank=True, max_length=120, null=True),
),
migrations.AlterField(
model_name='orderagent',
name='phone',
field=models.CharField(blank=True, max_length=50, null=True),
),
]
| [
"danii.litvinenko@x5.ru"
] | danii.litvinenko@x5.ru |
379b155116a31b53d67b638b0922f8cf82ef99a4 | 6aef2fdd5b98038fc6ecc7551dd76dccf370c4ae | /without_variance/GPOMDP_SVRG_WV_ada_bv_wbas.py | 25116e70f8b7cc295ad1394c40f7e36f3d85bfe8 | [
"LicenseRef-scancode-generic-cla",
"MIT"
] | permissive | Bobeye/rllab | 29b1cf3f29b748f93af4ac103d1a0eaa40290e7f | 53c0afb73f93c4a78ff21507914d7f7735c21ea9 | refs/heads/master | 2020-05-02T07:18:17.323566 | 2019-03-26T02:34:02 | 2019-03-26T02:34:02 | 177,814,299 | 0 | 0 | NOASSERTION | 2019-03-26T15:14:46 | 2019-03-26T15:14:45 | null | UTF-8 | Python | false | false | 18,229 | py | from rllab.envs.box2d.cartpole_env import CartpoleEnv
from rllab.policies.gaussian_mlp_policy import GaussianMLPPolicy
from rllab.envs.normalized_env import normalize
import numpy as np
import theano
import theano.tensor as TT
from rllab.sampler import parallel_sampler
from lasagne.updates import sgd
import matplotlib.pyplot as plt
from rllab.envs.gym_env import GymEnv
def unpack(i_g):
i_g_arr = [np.array(x) for x in i_g]
res = i_g_arr[0].reshape(i_g_arr[0].shape[0]*i_g_arr[0].shape[1])
res = np.concatenate((res,i_g_arr[1]))
res = np.concatenate((res,i_g_arr[2][0]))
res = np.concatenate((res,i_g_arr[3]))
return res
def compute_snap_batch(observations,actions,d_rewards,n_traj,n_part):
n=n_traj
i=0
svrg_snap=list()
while(n-np.int(n_traj/n_part)>=0):
n=n-np.int(n_traj/n_part)
s_g = compute_grad_snap(observations[i:i+np.int(n_traj/n_part)],
actions[i:i+np.int(n_traj/n_part)],d_rewards[i:i+np.int(n_traj/n_part)],np.int(n_traj/n_part),T)
i += np.int(n_traj/n_part)
svrg_snap.append(unpack(s_g))
return svrg_snap
def estimate_variance(observations,actions,d_rewards,snap_grads,n_traj,n_traj_s,n_part,M,N):
n=n_traj
i=0
svrg=list()
j=0
while(n-np.int(n_traj/n_part)>=0):
n=n-np.int(n_traj/n_part)
x = unpack(compute_grad_svrg(observations[i:i+np.int(n_traj/n_part)],actions[i:i+np.int(n_traj/n_part)],
d_rewards[i:i+np.int(n_traj/n_part)],np.int(n_traj/n_part),T,False,None))*np.sqrt(np.int(n_traj/n_part)/M)
g = snap_grads[j]*np.sqrt(np.int(n_traj_s/n_part)/N)+x
g=g/n_traj*n_part
i+=np.int(n_traj/n_part)
j+=1
svrg.append(g)
return (np.diag(np.cov(np.matrix(svrg),rowvar=False)).sum())
def compute_grad_snap(observations,actions,d_rewards,N,T):
minT=T
cum_num = []
cum_den = []
for ob,ac,rw in zip(observations,actions,d_rewards):
if minT>len(ob):
minT=len(ob)
x=f_baseline_g(ob, ac)
z = [y**2 for y in x]
index2 = np.arange(len(rw))
prov_der_num = [y[i]*rw[i] for i in index2 for y in z ]
prov_der_den = [y[i] for i in index2 for y in z]
cum_num.append(prov_der_num)
cum_den.append(prov_der_den)
mean_num = []
mean_den = []
baseline = []
for i in range(minT):
mean_num.append(cum_num[0][len(x)*i:len(x)*(i+1)])
mean_den.append(cum_den[0][len(x)*i:len(x)*(i+1)])
index = np.arange(len(mean_num[0]))
for i in range(minT):
for j in range(1,len(cum_den)):
mean_num[i] = [mean_num[i][pos] + cum_num[j][len(x)*i:len(x)*(i+1)][pos] for pos in index]
mean_den[i] = [mean_den[i][pos] + cum_den[j][len(x)*i:len(x)*(i+1)][pos] for pos in index]
for i in range(minT):
mean_num[i] = [mean_num[i][pos]/N for pos in index]
mean_den[i] = [mean_den[i][pos]/N for pos in index]
baseline.append([mean_num[i][pos]/mean_den[i][pos] for pos in index])
zero_grad = [mean_den[0][pos]*0 for pos in index]
for i in range(minT,T):
baseline.append(zero_grad)
cum = zero_grad
s_g = f_train(observations[0], actions[0])
index2 = np.arange(len(d_rewards[0]))
s_g = [y[i] for i in index2 for y in s_g]
for i in range(len(observations[0])):
R = [(d_rewards[0][i]-baseline[i][pos])*s_g[len(zero_grad)*i:len(zero_grad)*(i+1)][pos] for pos in index]
cum = [R[pos]+cum[pos] for pos in index]
for ob,ac,rw in zip(observations[1:],actions[1:],d_rewards[1:]):
s_g = f_train(ob, ac)
index2 = np.arange(len(rw))
s_g = [y[i] for i in index2 for y in s_g]
for i in range(len(ob)):
R = [(rw[i]-baseline[i][pos])*s_g[len(zero_grad)*i:len(zero_grad)*(i+1)][pos] for pos in index]
cum = [R[pos]+cum[pos] for pos in index]
cum = [cum[pos]/N for pos in index]
return cum
def compute_grad_svrg(observations,actions,d_rewards,M,T,add_full,fg):
minT=T
cum_num = []
cum_den = []
for ob,ac,rw in zip(observations,actions,d_rewards):
if minT>len(ob):
minT=len(ob)
x=f_baseline_g(ob, ac)
index = np.arange(len(x))
x_snap=f_baseline_g_snap(ob, ac)
iw = f_importance_weights(ob,ac)
index2 = np.arange(len(rw))
x_iw = [y[i]*iw[i] for i in index2 for y in x ]
x_snap_bv = [y[i] for i in index2 for y in x_snap]
index3 = np.arange(len(x_snap_bv))
x_dif = [x_iw[i]-x_snap_bv[i] for i in index3]
z = [y**2 for y in x_dif]
prov_der_num = [z[len(x)*i:len(x)*(i+1)][pos]*rw[i] for i in index2 for pos in index ]
prov_der_den = z
cum_num.append(prov_der_num)
cum_den.append(prov_der_den)
mean_num = []
mean_den = []
baseline = []
for i in range(minT):
mean_num.append(cum_num[0][len(x)*i:len(x)*(i+1)])
mean_den.append(cum_den[0][len(x)*i:len(x)*(i+1)])
for i in range(minT):
for j in range(1,len(cum_den)):
mean_num[i] = [mean_num[i][pos] + cum_num[j][len(x)*i:len(x)*(i+1)][pos] for pos in index]
mean_den[i] = [mean_den[i][pos] + cum_den[j][len(x)*i:len(x)*(i+1)][pos] for pos in index]
for i in range(minT):
mean_num[i] = [mean_num[i][pos]/M for pos in index]
mean_den[i] = [mean_den[i][pos]/M+1e-16 for pos in index]
baseline.append([mean_num[i][pos]/mean_den[i][pos] for pos in index])
zero_grad = [mean_den[0][pos]*0 for pos in index]
for i in range(minT,T):
baseline.append(zero_grad)
cum = zero_grad
s_g = f_baseline_g(observations[0], actions[0])
s_g_snap_p=f_baseline_g_snap(observations[0], actions[0])
iw = f_importance_weights(observations[0], actions[0])
index2 = np.arange(len(d_rewards[0]))
s_g_iw = [y[i]*iw[i] for i in index2 for y in s_g ]
s_g_snap = [y[i] for i in index2 for y in s_g_snap_p]
for i in range(len(observations[0])):
R = [(d_rewards[0][i]-baseline[i][pos])*(s_g_iw[len(zero_grad)*i:len(zero_grad)*(i+1)][pos]-s_g_snap[len(zero_grad)*i:len(zero_grad)*(i+1)][pos]) for pos in index]
cum = [R[pos]+cum[pos] for pos in index]
for ob,ac,rw in zip(observations[1:],actions[1:],d_rewards[1:]):
s_g = f_baseline_g(ob, ac)
s_g_snap=f_baseline_g_snap(ob, ac)
iw = f_importance_weights(ob, ac)
index2 = np.arange(len(rw))
s_g_iw = [y[i]*iw[i] for i in index2 for y in s_g ]
s_g_snap = [y[i] for i in index2 for y in s_g_snap]
for i in range(len(ob)):
R = [(rw[i]-baseline[i][pos])*(-s_g_iw[len(zero_grad)*i:len(zero_grad)*(i+1)][pos]+s_g_snap[len(zero_grad)*i:len(zero_grad)*(i+1)][pos]) for pos in index]
cum = [R[pos]+cum[pos] for pos in index]
cum = [cum[pos]/M for pos in index]
if (add_full):
cum = [cum[pos] + fg[pos] for pos in index]
return cum
load_policy=True
# normalize() makes sure that the actions for the environment lies
# within the range [-1, 1] (only works for environments with continuous actions)
env = normalize(CartpoleEnv())
#env = GymEnv("InvertedPendulum-v1")
# Initialize a neural network policy with a single hidden layer of 8 hidden units
policy = GaussianMLPPolicy(env.spec, hidden_sizes=(8,),learn_std=False)
snap_policy = GaussianMLPPolicy(env.spec, hidden_sizes=(8,),learn_std=False)
back_up_policy = GaussianMLPPolicy(env.spec, hidden_sizes=(8,),learn_std=False)
parallel_sampler.populate_task(env, snap_policy)
# policy.distribution returns a distribution object under rllab.distributions. It contains many utilities for computing
# distribution-related quantities, given the computed dist_info_vars. Below we use dist.log_likelihood_sym to compute
# the symbolic log-likelihood. For this example, the corresponding distribution is an instance of the class
# rllab.distributions.DiagonalGaussian
dist = policy.distribution
snap_dist = snap_policy.distribution
# We will collect 100 trajectories per iteration
N = 100
# Each trajectory will have at most 100 time steps
T = 100
#We will collect M secondary trajectories
M = 10
#Number of sub-iterations
#m_itr = 100
# Number of iterations
#n_itr = np.int(10000/(m_itr*M+N))
# Set the discount factor for the problem
discount = 0.99
# Learning rate for the gradient update
learning_rate = 0.00005
#perc estimate
perc_est = 0.6
#tot trajectories
s_tot = 10000
partition = 3
porz = np.int(perc_est*N)
observations_var = env.observation_space.new_tensor_variable(
'observations',
# It should have 1 extra dimension since we want to represent a list of observations
extra_dims=1
)
actions_var = env.action_space.new_tensor_variable(
'actions',
extra_dims=1
)
d_rewards_var = TT.vector('d_rewards')
importance_weights_var = TT.vector('importance_weight')
bl = TT.vector()
# policy.dist_info_sym returns a dictionary, whose values are symbolic expressions for quantities related to the
# distribution of the actions. For a Gaussian policy, it contains the mean and (log) standard deviation.
dist_info_vars = policy.dist_info_sym(observations_var)
snap_dist_info_vars = snap_policy.dist_info_sym(observations_var)
surr = - dist.log_likelihood_sym_1traj_GPOMDP(actions_var, dist_info_vars)
params = policy.get_params(trainable=True)
snap_params = snap_policy.get_params(trainable=True)
importance_weights = dist.likelihood_ratio_sym_1traj_GPOMDP(actions_var,snap_dist_info_vars,dist_info_vars)
grad = TT.jacobian(surr, params)
eval_grad1 = TT.matrix('eval_grad0',dtype=grad[0].dtype)
eval_grad2 = TT.vector('eval_grad1',dtype=grad[1].dtype)
eval_grad3 = TT.col('eval_grad3',dtype=grad[2].dtype)
eval_grad4 = TT.vector('eval_grad4',dtype=grad[3].dtype)
surr_on1 = TT.sum(- dist.log_likelihood_sym_1traj_GPOMDP(actions_var,dist_info_vars)*d_rewards_var*importance_weights_var)
surr_on2 = TT.sum(snap_dist.log_likelihood_sym_1traj_GPOMDP(actions_var,snap_dist_info_vars)*d_rewards_var)
grad_SVRG =[sum(x) for x in zip([eval_grad1, eval_grad2, eval_grad3, eval_grad4], theano.grad(surr_on1,params),theano.grad(surr_on2,snap_params))]
grad_SVRG_4v = [sum(x) for x in zip(theano.grad(surr_on1,params),theano.grad(surr_on2,snap_params))]
grad_var = theano.grad(surr_on1,params)
cum_likelihood = dist.log_likelihood_sym_1traj_GPOMDP(actions_var, dist_info_vars)
cum_likelihood_snap = dist.log_likelihood_sym_1traj_GPOMDP(actions_var, snap_dist_info_vars)
all_der, update_scan = theano.scan(lambda i, cum_likelihood: theano.grad(cum_likelihood[i], params),
sequences=TT.arange(cum_likelihood.shape[0]),
non_sequences=cum_likelihood)
all_der_snap, update_scan = theano.scan(lambda i, cum_likelihood_snap: theano.grad(cum_likelihood_snap[i], snap_params),
sequences=TT.arange(cum_likelihood_snap.shape[0]),
non_sequences=cum_likelihood_snap)
f_train = theano.function(
inputs = [observations_var, actions_var],
outputs = grad
)
f_update = theano.function(
inputs = [eval_grad1, eval_grad2, eval_grad3, eval_grad4],
outputs = None,
updates = sgd([eval_grad1, eval_grad2, eval_grad3, eval_grad4], params, learning_rate=learning_rate)
)
f_importance_weights = theano.function(
inputs = [observations_var, actions_var],
outputs = importance_weights
)
f_update_SVRG = theano.function(
inputs = [eval_grad1, eval_grad2, eval_grad3, eval_grad4],
outputs = None,
updates = sgd([eval_grad1, eval_grad2, eval_grad3, eval_grad4], params, learning_rate=learning_rate)
)
f_train_SVRG = theano.function(
inputs=[observations_var, actions_var, d_rewards_var, eval_grad1, eval_grad2, eval_grad3, eval_grad4,importance_weights_var],
outputs=grad_SVRG,
)
f_train_SVRG_4v = theano.function(
inputs=[observations_var, actions_var, d_rewards_var,importance_weights_var],
outputs=grad_SVRG_4v,
)
var_SVRG = theano.function(
inputs=[observations_var, actions_var, d_rewards_var, importance_weights_var],
outputs=grad_var,
)
f_baseline_g = theano.function(
inputs = [observations_var, actions_var],
outputs = all_der
)
f_baseline_g_snap = theano.function(
inputs = [observations_var, actions_var],
outputs = all_der_snap
)
alla = []
alla2 = []
alla3 = []
for k in range(10):
alla4=[]
if (load_policy):
snap_policy.set_param_values(np.loadtxt('policy_novar.txt'), trainable=True)
policy.set_param_values(np.loadtxt('policy_novar.txt'), trainable=True)
avg_return = np.zeros(s_tot)
#np.savetxt("policy_novar.txt",snap_policy.get_param_values(trainable=True))
j=0
while j<s_tot-N:
paths = parallel_sampler.sample_paths_on_trajectories(snap_policy.get_param_values(),N,T,show_bar=False)
#baseline.fit(paths)
j+=N
observations = [p["observations"] for p in paths]
actions = [p["actions"] for p in paths]
d_rewards = [p["rewards"] for p in paths]
temp = list()
for x in d_rewards:
z=list()
t=1
for y in x:
z.append(y*t)
t*=discount
temp.append(np.array(z))
d_rewards=temp
s_g = compute_grad_snap(observations,actions,d_rewards,N,T)
b=compute_snap_batch(observations[0:porz],actions[0:porz],d_rewards[0:porz],porz,partition)
f_update(s_g[0],s_g[1],s_g[2],s_g[3])
avg_return[j-N:j] = np.repeat(np.mean([sum(p["rewards"]) for p in paths]),N)
var_sgd = np.cov(np.matrix(b),rowvar=False)
var_batch = (var_sgd)*(porz/partition)/M
print(str(j-1)+' Average Return:', avg_return[j-1])
back_up_policy.set_param_values(policy.get_param_values(trainable=True), trainable=True)
n_sub = 0
while j<s_tot-M:
iw_var = f_importance_weights(observations[0],actions[0])
var_svrg = (estimate_variance(observations[porz:],actions[porz:],d_rewards[porz:],b,N-porz,porz,partition,M,N))
var_dif = var_svrg-(np.diag(var_batch).sum())
alla2.append(var_svrg)
alla3.append((np.diag(var_batch).sum()))
alla4.append(np.mean(iw_var))
if (var_dif>0):
policy.set_param_values(back_up_policy.get_param_values(trainable=True), trainable=True)
break
j += M
n_sub+=1
sub_paths = parallel_sampler.sample_paths_on_trajectories(snap_policy.get_param_values(),M,T,show_bar=False)
#baseline.fit(paths)
sub_observations=[p["observations"] for p in sub_paths]
sub_actions = [p["actions"] for p in sub_paths]
sub_d_rewards = [p["rewards"] for p in sub_paths]
temp = list()
for x in sub_d_rewards:
z=list()
t=1
for y in x:
z.append(y*t)
t*=discount
temp.append(np.array(z))
sub_d_rewards=temp
iw = f_importance_weights(sub_observations[0],sub_actions[0])
back_up_policy.set_param_values(policy.get_param_values(trainable=True), trainable=True)
g = compute_grad_svrg(observations,actions,d_rewards,M,T,True,s_g)
f_update(g[0],g[1],g[2],g[3])
avg_return[j-M:j] = np.repeat(np.mean([sum(p["rewards"]) for p in sub_paths]),M)
#print(str(j)+' Average Return:', avg_return[j])
snap_policy.set_param_values(policy.get_param_values(trainable=True), trainable=True)
plt.plot(avg_return[::10])
plt.show()
plt.plot(alla2)
plt.plot(alla3)
plt.show()
alla.append(avg_return)
alla_mean = [np.mean(x) for x in zip(*alla)]
plt.plot(alla_mean)
plt.plot()
np.savetxt("GPOMDP_SVRG_wbas",alla_mean)
gpomdp = np.loadtxt("GPOMDP_l5e-05")
gpomdpbas = np.loadtxt("GPOMDP_with_base")
#gpomdp_svrg=np.loadtxt("GPOMDP_SVRG_5e-5")
#gpomdp_svrg_ada_wb = np.loadtxt("GPOMDP_SVRG_5e-5_ada_wb")
plt.plot(gpomdp )
plt.plot(gpomdpbas)
plt.plot(gpomdp_svrg_ada_wb[::10])
plt.plot(alla_mean[::10])
plt.legend(['gpomdp','gpomdp baseline','gpomdp_svrg','gpomdp_svrg baseline'], loc='lower right')
plt.savefig("baseline_verbaseline.jpg", figsize=(32, 24), dpi=160)
#gpomdp_svrg_ada_wb_bv_m7 = np.loadtxt("GPOMDP_SVRG_5e-5_ada_b2")
#gpomdp_svrg_ada_wb_bv_m5 = np.loadtxt("GPOMDP_SVRG_5e-5_ada_b2_m5")
#gpomdp_svrg_ada_wb_bv_m3 = np.loadtxt("GPOMDP_SVRG_5e-5_ada_b2_m3")
#gpomdp_svrg_ada_wb_bv_s15 = np.loadtxt("GPOMDP_SVRG_5e-5_ada_b2_s15")
#
plt.plot(gpomdp)
#plt.plot(gpomdp_svrg)
#plt.plot(gpomdp_svrg_ada_wb[::10])
plt.plot(gpomdp_svrg_ada_wb_bv_m7[::10])
#plt.plot(gpomdp_svrg_ada_wb_bv_m3[::10])
#plt.plot(gpomdp_svrg_ada_wb_bv_m5[::10])
#plt.plot(gpomdp_svrg_ada_wb_bv_s15[::10])
#plt.legend(['gpomdp','gpomdp_svrg','gpomdp_svrg_ada_wb','gpomdp_svrg_m7','gpomdp_svrg_s15'], loc='lower right')
#plt.savefig("adapt_nnv.jpg", figsize=(32, 24), dpi=160)
plt.show()
#uni = np.ones(640,dtype=np.int)
#for i in range(40):
# uni[i*16]=10
#scal_svrg = np.repeat(gpondp_svrg,uni)
#plt.plot(gpondp)
#plt.plot(scal_svrg )
#plt.legend(['gpondp','gpondp_svrg'], loc='lower right')
#plt.savefig("gpondp_5e-6.jpg", figsize=(32, 24), dpi=160) | [
"damiano.binaghi@gmail.com"
] | damiano.binaghi@gmail.com |
e7e95d0bc9841ba1cce84f802176bbe2ef6d5e38 | 6c71226e2080c79a993fb086445c5af52b42bb95 | /randomForestRegression/random_forest_regression.py | 1df75f2bfa11cfc2b80c74e43cd52df35f2a8c6d | [] | no_license | ArakelyanEdgar/MachineLearningAlgorithms | 32f4046931d39cfacc58bd6c753031fd3af3d3aa | 07b64e83dafc9259f9a09aefcce07ca4b261b76f | refs/heads/master | 2020-03-08T13:27:31.750449 | 2018-04-05T04:27:44 | 2018-04-05T04:27:44 | 128,158,464 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 1,304 | py | # Random Forest Regression
# Importing the libraries
import numpy as np
import matplotlib.pyplot as plt
import pandas as pd
from sklearn.ensemble import RandomForestRegressor
# Importing the dataset
dataset = pd.read_csv('Position_Salaries.csv')
X = dataset.iloc[:, 1:2].values
y = dataset.iloc[:, 2].values
# Splitting the dataset into the Training set and Test set
"""from sklearn.cross_validation import train_test_split
X_train, X_test, y_train, y_test = train_test_split(X, y, test_size = 0.2, random_state = 0)"""
# Feature Scaling
"""from sklearn.preprocessing import StandardScaler
sc_X = StandardScaler()
X_train = sc_X.fit_transform(X_train)
X_test = sc_X.transform(X_test)
sc_y = StandardScaler()
y_train = sc_y.fit_transform(y_train)"""
#random forest regression
regressor = RandomForestRegressor(n_estimators=1000, random_state=0)
regressor.fit(X, y)
#Predicting a new result
y_pred = regressor.predict(6.5)
print(y_pred)
# Visualising the Random Forest Regression results (higher resolution)
X_grid = np.arange(min(X), max(X), 0.01)
X_grid = X_grid.reshape((len(X_grid), 1))
plt.scatter(X, y, color = 'red')
plt.plot(X_grid, regressor.predict(X_grid), color = 'blue')
plt.title('Truth or Bluff (Random Forest Regression)')
plt.xlabel('Position level')
plt.ylabel('Salary')
plt.show() | [
"edgararakelyan123@gmail.com"
] | edgararakelyan123@gmail.com |
4a80747e2268ca965179dddbee104338c66849bc | af65714ea99ea2a1edd6b372609f682399a7d64d | /your_app_name/manage.py | 5578a3e697c59017498982b4457f470dcf7b70f3 | [
"MIT"
] | permissive | gibeongideon/django-github-action-runner-CICD | fbfb81b94bbb4ccc93fc90cbc452695a4949a502 | ddf02176dc83e3f7ed4944f8f48207c944e33f18 | refs/heads/master | 2023-06-05T21:39:30.002833 | 2021-06-23T20:21:46 | 2021-06-23T20:21:46 | 379,716,569 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 669 | py | #!/usr/bin/env python
"""Django's command-line utility for administrative tasks."""
import os
import sys
def main():
"""Run administrative tasks."""
os.environ.setdefault('DJANGO_SETTINGS_MODULE', 'your_app_name.settings')
try:
from django.core.management import execute_from_command_line
except ImportError as exc:
raise ImportError(
"Couldn't import Django. Are you sure it's installed and "
"available on your PYTHONPATH environment variable? Did you "
"forget to activate a virtual environment?"
) from exc
execute_from_command_line(sys.argv)
if __name__ == '__main__':
main()
| [
"kipngeno.gibeon@gmail.com"
] | kipngeno.gibeon@gmail.com |
f369d5667a7f0255f82296fbbee935075af34b7e | 7b5ec17918cb2328d53bf2edd876c153af26b38d | /scripts/ingestors/rwis/process_idot_awos.py | c29e696ecbcafd40fb720a5612021a2b033ca115 | [
"MIT"
] | permissive | Xawwell/iem | 78e62f749661f3ba292327f82acf4ef0f0c8d55b | 88177cc096b9a66d1bd51633fea448585b5e6573 | refs/heads/master | 2020-09-06T09:03:54.174221 | 2019-11-08T03:23:44 | 2019-11-08T03:23:44 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 2,063 | py | """Process AWOS METAR file"""
from __future__ import print_function
import re
import sys
import os
import datetime
import ftplib
import subprocess
import tempfile
from io import StringIO
from pyiem import util
INCOMING = "/mesonet/data/incoming"
def fetch_files():
"""Fetch files """
props = util.get_properties()
fn = "%s/iaawos_metar.txt" % (INCOMING,)
try:
ftp = ftplib.FTP("165.206.203.34")
except TimeoutError:
print("process_idot_awos FTP server timeout error")
sys.exit()
ftp.login("rwis", props["rwis_ftp_password"])
ftp.retrbinary("RETR METAR.txt", open(fn, "wb").write)
ftp.close()
return fn
def main():
"""Go Main"""
fn = fetch_files()
utc = datetime.datetime.utcnow().strftime("%Y%m%d%H%M")
data = {}
# Sometimes, the file gets gobbled it seems
for line in open(fn, "rb"):
line = line.decode("utf-8", "ignore")
match = re.match("METAR K(?P<id>[A-Z1-9]{3})", line)
if not match:
continue
gd = match.groupdict()
data[gd["id"]] = line
sio = StringIO()
sio.write("\001\r\r\n")
sio.write(
("SAUS00 KISU %s\r\r\n")
% (datetime.datetime.utcnow().strftime("%d%H%M"),)
)
sio.write("METAR\r\r\n")
for sid in data:
sio.write("%s=\r\r\n" % (data[sid].strip().replace("METAR ", ""),))
sio.write("\003")
sio.seek(0)
(tmpfd, tmpname) = tempfile.mkstemp()
os.write(tmpfd, sio.getvalue().encode("utf-8"))
os.close(tmpfd)
proc = subprocess.Popen(
(
"/home/ldm/bin/pqinsert -i -p 'data c %s "
"LOCDSMMETAR.dat LOCDSMMETAR.dat txt' %s"
)
% (utc, tmpname),
shell=True,
stdin=subprocess.PIPE,
stdout=subprocess.PIPE,
stderr=subprocess.STDOUT,
)
(stdout, stderr) = proc.communicate()
os.remove(tmpname)
if stdout != b"" or stderr is not None:
print("process_idot_awos\nstdout: %s\nstderr: %s" % (stdout, stderr))
if __name__ == "__main__":
main()
| [
"akrherz@iastate.edu"
] | akrherz@iastate.edu |
b132f2782265b56e98b88f6df9aaaa8c85d5bddc | 0db8fad3d630899a1e8389349b047c2cabdb6a27 | /meiduo_mall/meiduo_mall/apps/orders/migrations/0002_auto_20190531_0919.py | 6168072e180c8d36de7f6d08fa9748ee6b9c8253 | [] | no_license | zhujian2019/Django_Frontend | 30123b8e20a9e0fc6b48b49433c22744cd1aa9d1 | ee8fa8385487ee96b309f8230928e5945ddc5c86 | refs/heads/master | 2020-06-04T19:46:17.733452 | 2019-06-16T08:38:14 | 2019-06-16T08:38:14 | 192,167,921 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 420 | py | # -*- coding: utf-8 -*-
# Generated by Django 1.11.11 on 2019-05-31 09:19
from __future__ import unicode_literals
from django.db import migrations
class Migration(migrations.Migration):
dependencies = [
('orders', '0001_initial'),
]
operations = [
migrations.RenameField(
model_name='ordergoods',
old_name='oreder',
new_name='order',
),
]
| [
"zhujian_work@163.com"
] | zhujian_work@163.com |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.