blob_id
stringlengths 40
40
| directory_id
stringlengths 40
40
| path
stringlengths 2
616
| content_id
stringlengths 40
40
| detected_licenses
listlengths 0
69
| license_type
stringclasses 2
values | repo_name
stringlengths 5
118
| snapshot_id
stringlengths 40
40
| revision_id
stringlengths 40
40
| branch_name
stringlengths 4
63
| visit_date
timestamp[us] | revision_date
timestamp[us] | committer_date
timestamp[us] | github_id
int64 2.91k
686M
⌀ | star_events_count
int64 0
209k
| fork_events_count
int64 0
110k
| gha_license_id
stringclasses 23
values | gha_event_created_at
timestamp[us] | gha_created_at
timestamp[us] | gha_language
stringclasses 213
values | src_encoding
stringclasses 30
values | language
stringclasses 1
value | is_vendor
bool 2
classes | is_generated
bool 2
classes | length_bytes
int64 2
10.3M
| extension
stringclasses 246
values | content
stringlengths 2
10.3M
| authors
listlengths 1
1
| author_id
stringlengths 0
212
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
c5fa5fda6af41850d5525c2fbd4e5caab4f6ccee
|
73000d852404cc1fd3accfbfc8c83a0b88f6b4cb
|
/tests/test_xtbparser.py
|
17d521260fae33c10b5b68e03bef388a5d5191f7
|
[
"Apache-2.0"
] |
permissive
|
nomad-coe/nomad-parser-xtb
|
1580f90a2d391c63c3d291bc6a698779d599be10
|
09894dc15511fe6a9e9938027a0c48786dcdc084
|
refs/heads/master
| 2023-06-20T16:41:18.147131
| 2021-07-14T21:55:22
| 2021-07-14T21:55:22
| 386,077,099
| 1
| 0
| null | null | null | null |
UTF-8
|
Python
| false
| false
| 1,111
|
py
|
#
# Copyright The NOMAD Authors.
#
# This file is part of NOMAD. See https://nomad-lab.eu for further info.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
#
import pytest
import numpy as np
from nomad.datamodel import EntryArchive
from xtbparser import XTBParser
def approx(value, abs=0, rel=1e-6):
return pytest.approx(value, abs=abs, rel=rel)
@pytest.fixture(scope='module')
def parser():
return XTBParser()
def test_scf(parser):
archive = EntryArchive()
parser.parse('tests/data/gfn1/out', archive, None)
sec_run = archive.section_run[0]
assert sec_run.program_name == '6.4.1'
|
[
"ladinesalvinnoe@gmail.com"
] |
ladinesalvinnoe@gmail.com
|
fb28d6bd690ff888dfd3ea29b317ae4cf3d2be7a
|
8eca0a7a9ae207113f9f9ed98d093cbe21ffcd8a
|
/Maths_based_problems/grid_unique_paths/solution2_dp.py
|
eed4c34ad0482d36605fb37189b7a03c658eb218
|
[] |
no_license
|
Ranjit007ai/-Interviewbit-Maths
|
3238c720bb5e0765eef0e0e1a39549eff1ba788d
|
044627422fc38ee3e5aaa9cbfc8f00398d1f9bb5
|
refs/heads/main
| 2023-03-25T06:06:14.206384
| 2021-03-27T14:38:16
| 2021-03-27T14:38:16
| 352,065,531
| 0
| 0
| null | null | null | null |
UTF-8
|
Python
| false
| false
| 568
|
py
|
def unique_paths(m,n):
dp = [[0]*n for _ in range(0,m)]
# each position in dp show the no of way to reach their
# now the first row and first col will be 1 ,since their is 1 way to traverse from one position to another in single row or column vector
for row in range(0,m):
dp[row][0] = 1
for col in range(0,n):
dp[0][col] = 1
for row in range(1,m):
for col in range(1,n):
dp[row][col] = dp[row-1][col] + dp[row][col-1]
return dp[m-1][n-1]
m =3
n =
ans = unique_paths(m,n)
print(ans)
|
[
"noreply@github.com"
] |
Ranjit007ai.noreply@github.com
|
7ba0a06d9808d2f9a578e5b737327d2a711235eb
|
04a6f0914b128251bbd032bf13fbbfddb3c4eaa7
|
/pythonbind/src/head.py
|
8f67231978aa717ed41f4eb826b0187e5779b55a
|
[] |
no_license
|
YulinWu/qos_public
|
a8d6eb4cba65a17e292f7d072bfb6728e95c7815
|
58cb67f62c285d70f86cb8a79d147610abb7ff3f
|
refs/heads/master
| 2022-12-30T22:17:09.310242
| 2020-10-16T05:38:02
| 2020-10-16T05:38:02
| 304,203,187
| 0
| 1
| null | 2020-10-16T05:38:04
| 2020-10-15T03:48:59
|
Java
|
UTF-8
|
Python
| false
| false
| 3,284
|
py
|
import jpype
import jpype.imports
from jpype.types import *
import os
def matchArgument(patterns, args):
if len(patterns) != len(args):
return False
for p, v in zip(patterns, args):
if p.startswith('@NonNull ') and v == None:
return False
if p.startswith('@NonNull '):
p = p[9:]
vtype = p.split(' ')[-2]
if vtype[-2:] == '[]' and not isinstance(v, list):
return False
if vtype[-4:] == '[][]' and not isinstance(v[0], list):
return False
return True
def convertArgument(patterns, args):
typemap = {
'boolean': JBoolean,
'byte': JByte,
'char': JChar,
'short': JShort,
'int': JInt,
'long': JLong,
'float': JFloat,
'double': JDouble,
'String': JString,
'boolean[]': JArray(JBoolean),
'byte[]': JArray(JByte),
'char[]': JArray(JChar),
'short[]': JArray(JShort),
'int[]': JArray(JInt),
'long[]': JArray(JLong),
'float[]': JArray(JFloat),
'double[]': JArray(JDouble),
'String[]': JArray(JString),
'boolean[][]': JArray(JArray(JBoolean)),
'byte[][]': JArray(JArray(JByte)),
'char[][]': JArray(JArray(JChar)),
'short[][]': JArray(JArray(JShort)),
'int[][]': JArray(JArray(JInt)),
'long[][]': JArray(JArray(JLong)),
'float[][]': JArray(JArray(JFloat)),
'double[][]': JArray(JArray(JDouble)),
'String[][]': JArray(JArray(JString)),
'boolean[][][]': JArray(JArray(JArray(JBoolean))),
'byte[][][]': JArray(JArray(JArray(JByte))),
'char[][][]': JArray(JArray(JArray(JChar))),
'short[][][]': JArray(JArray(JArray(JShort))),
'int[][][]': JArray(JArray(JArray(JInt))),
'long[][][]': JArray(JArray(JArray(JLong))),
'float[][][]': JArray(JArray(JArray(JFloat))),
'double[][][]': JArray(JArray(JArray(JDouble))),
'String[][][]': JArray(JArray(JArray(JString))),
}
args2 = []
for p, v in zip(patterns, args):
vtype = p.split(' ')[-2]
args2.append(typemap[vtype](v))
return args2
jpype.startJVM(classpath=[os.path.abspath(os.path.dirname(
__file__))+'/../../rpcapi/target/qos-rpcapi-1.0.jar'])
# jpype.startJVM(classpath=[os.path.abspath(
# os.path.dirname(__file__))+'/../../rpcapi/target/classes'])
QOSServerApi = jpype.JClass('qc.qos.rpcapi.qpx1701.QOSServerApi')
class g:
Instance = None
def getInstance(*args):
patterns = ['@NonNull String userName',
'String password',
'@NonNull String host',
'int port']
if matchArgument(patterns, args):
args = convertArgument(patterns, args)
g.Instance = QOSServerApi(*args)
return
patterns = ['@NonNull String userName',
'String password',
'@NonNull String host',
'int port',
'boolean useSSL',
'String keyFile',
'char[] keyStorePassword']
if matchArgument(patterns, args):
args = convertArgument(patterns, args)
g.Instance = QOSServerApi(*args)
return
raise TypeError('no input pattern match')
|
[
"zhao.uv@gmail.com"
] |
zhao.uv@gmail.com
|
c0355743b4fe4bb2b1c3a1515697ae7ad09aba3e
|
10c9699a6f2f2af4986048b53e44bb063312c3cd
|
/StudentAnalysis/accounts/models.py
|
b8f83e1cb3a24f9aaa72325b44ba079324ed877f
|
[] |
no_license
|
vagdevik/GDG-VIT-Hackathon
|
60dea9e9a187825dde7854c8c3cff6e8f24cf690
|
0eb00ae8bf3b2122cbf4e3789a7ce544ff405799
|
refs/heads/master
| 2021-05-07T18:21:27.411760
| 2017-10-30T04:19:43
| 2017-10-30T04:19:43
| null | 0
| 0
| null | null | null | null |
UTF-8
|
Python
| false
| false
| 261
|
py
|
# -*- coding: utf-8 -*-
from __future__ import unicode_literals
from django.db import models
from django.contrib import auth
# class User(auth.models.User, auth.models.PermissionsMixin):
#
# def __str__(self):
# return "@{}".format(self.username)
|
[
"abhishek.m15@iiits.in"
] |
abhishek.m15@iiits.in
|
b3096c51d5f0148b23157700f003b048c28d4cb6
|
efd6c1d24b0a392a177679429d53dd2f515d0d95
|
/bi_auth/migrations/0001_initial.py
|
fd577d05d960ad8b413cd4b9a52c1be60fe0f81b
|
[] |
no_license
|
mitshel/TakedaAnalitic
|
5ccfb4aa83a056cbeaebce03df41819c7ece7985
|
b04b08fb053bff238a1ce68df423f99314827b48
|
refs/heads/master
| 2021-07-02T21:27:10.023816
| 2019-02-14T14:32:18
| 2019-02-14T14:32:42
| 153,908,777
| 0
| 0
| null | null | null | null |
UTF-8
|
Python
| false
| false
| 979
|
py
|
# Generated by Django 2.1.2 on 2018-12-15 11:16
from django.conf import settings
from django.db import migrations, models
import django.db.models.deletion
class Migration(migrations.Migration):
initial = True
dependencies = [
migrations.swappable_dependency(settings.AUTH_USER_MODEL),
]
operations = [
migrations.CreateModel(
name='UserProfile',
fields=[
('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')),
('is_orgadmin', models.BooleanField(verbose_name='Администратор организации')),
('user', models.OneToOneField(on_delete=django.db.models.deletion.CASCADE, related_name='user_profile', to=settings.AUTH_USER_MODEL)),
],
options={
'verbose_name': 'Профиль',
'verbose_name_plural': 'Профили',
},
),
]
|
[
"mitshel@mail.ru"
] |
mitshel@mail.ru
|
9004a5e73ad5c6897f0e9b223cb0cbfb05dd568f
|
5b301fb18d62953cbf48e284594abe9e14ae90d9
|
/hyperbolic/euclid/origin_line.py
|
7b6431e6d2b143972ab8adfd1ab5dd4fd364c572
|
[
"MIT"
] |
permissive
|
cduck/hyperbolic
|
70ba6bec20aae2e7d63382ffb25ed5b491c61496
|
bebe9b546e12f28d76ed07d0a2d9954d9d143571
|
refs/heads/master
| 2023-08-21T22:36:13.504014
| 2023-02-27T02:42:04
| 2023-02-27T05:01:31
| 115,620,902
| 119
| 30
|
MIT
| 2023-09-01T13:58:35
| 2017-12-28T12:24:28
|
Python
|
UTF-8
|
Python
| false
| false
| 893
|
py
|
from . import line
class OriginLine(line.Line):
def __init__(self, px, py):
self.px = px
self.py = py
def to_line(self):
return line.Line(0, 0, self.px, self.py)
@property
def x1(self): return 0
@property
def y1(self): return 0
@property
def x2(self): return self.px
@property
def y2(self): return self.py
def __repr__(self):
return '{}({}, {})'.format(type(self).__name__,
round(self.px, 3), round(self.py, 3))
def reverse(self):
self.px, self.py = -self.px, -self.py
def reversed(self):
return OriginLine(-self.px, -self.py)
@staticmethod
def from_points(x1, y1, **kwargs):
return OriginLine(x1, y1, **kwargs)
def to_drawables(self, **kwargs):
import drawsvg as draw
return (draw.Line(self.x1, self.y1, self.x2, self.y2, **kwargs),)
|
[
"cduckering@gmail.com"
] |
cduckering@gmail.com
|
1a5dd463ef4a3efe08755888c769206aba42c6a0
|
70f58b1dba311ea632257c7262c143f410378051
|
/corona/index.py
|
cc46ba839a4966150b4805cb64995c8d68071581
|
[] |
no_license
|
azizyilmaz/data-analysis-with-python
|
e295488ae7f0d938a4d0422d06801c1241794763
|
2ee39d3bf3d72e5b43a785ba4690cb21e1666966
|
refs/heads/master
| 2022-11-18T09:21:14.390509
| 2020-07-12T21:45:14
| 2020-07-12T21:45:14
| 277,184,867
| 0
| 0
| null | null | null | null |
UTF-8
|
Python
| false
| false
| 734
|
py
|
import pandas
dataset = pandas.read_csv("corona/coronavirusdata/covid_19_data.csv")
print(f'Covid19 veri listesi:\n{dataset}')
print(f'Satır ve sütun sayıları:\n{dataset.shape}')
print(f'Sütunları:\n{dataset.columns}')
print(f'Veri tipleri:\n{dataset.dtypes}')
print(f'İlk 10 veri listesi:\n{dataset.head(10)}')
print(f'Son 10 veri listesi:\n{dataset.tail(10)}')
print(f'Veri listesi bilgileri:\n{dataset.info()}')
print(f'Veri listesi özeti:\n{dataset.describe()}')
# En fazla ölümün olduğu 20 veriyi getirir
print(dataset.sort_values(by="Deaths", ascending=False).head(20))
# Türkiye sonuçları
print(dataset[dataset['Country/Region'] == 'Turkey']
.sort_values(by="Deaths", ascending=False).head(20))
|
[
"azizxyilmaz@outlook.com"
] |
azizxyilmaz@outlook.com
|
1bab277d30cc4f1a009426a5f5951a2f5b8d1ff1
|
4453c7b7051fd47a8b2fe7d24f6be338dee830aa
|
/lostfound/mainapp/views.py
|
3a5e39dc5f1bd9cfc8c97bd93a0718a0070ea398
|
[] |
no_license
|
travo-star/lost-and-found
|
c39f6a548698eab8ed6a2cdd75285d7509d36c18
|
d1fc17a9bb66d2b44f758f0cb9d5af90f59daa3b
|
refs/heads/master
| 2020-08-29T12:01:03.444263
| 2019-10-28T09:58:06
| 2019-10-28T09:58:06
| null | 0
| 0
| null | null | null | null |
UTF-8
|
Python
| false
| false
| 2,540
|
py
|
from django.shortcuts import render,redirect
from mainapp.models import Post
from django.http import HttpResponseRedirect
from .forms import Posts, ClaimForm
from django.contrib import messages
from django.contrib.auth.decorators import login_required
from django.views.generic import CreateView
from .models import Claim,Post
from django.views.generic import ListView
from users.models import Stratizen
from .filters import PostFilter
# Create your views here.
def homepage(request):
return render(request, 'mainapp/home.html', {'title':'Home'})
@login_required(login_url='/login/')
def claimitem(request):
return render(request, 'mainapp/claimitem.html', {'title':'Claim Item'})
@login_required(login_url='/login/')
def postitem(request):
if request.method == 'POST':
form = Posts(request.POST, request.FILES)
if form.is_valid():
form.save()
messages.success(request, f'The item has been posted successfully')
return redirect('posts_search')
else:
form = Posts()
return render(request,'mainapp/postitem.html',{'form':form},{'title':'Post Item'})
@login_required(login_url='/login/')
def viewitems(request):
return render(request, 'mainapp/viewitems.html', {'title':'View Item'})
@login_required(login_url='/login/')
def claimitem(request, post_id):
post = Post.objects.get(id=post_id)
if request.method == 'POST':
claim_form = ClaimForm(request.POST, request.FILES)
claim_form.instance.claimed_by = request.user
claim_form.instance.item_claimed = post
if claim_form.is_valid():
claim_form.save()
messages.success(request, f'Claim Submitted Successfully!')
return redirect('posts_search')
else:
claim_form = ClaimForm()
context = {
'claim_form': claim_form,
'post': Post.objects.get(id=post_id),
'title':'Claim Item'
}
return render(request, 'mainapp/claimitem.html', context)
class ClaimCreateView(CreateView):
model = Claim
fields = ['description','location','time_from','time_to','image']
class PostListView(ListView):
context_object_name = 'posts'
def get_queryset(self):
return Post.objects.filter(claimed=False).order_by('-When')
def posts_search(request):
posts = Post.objects.filter(claimed=False).order_by('-When')
post_filter = PostFilter(request.GET, queryset=posts)
return render(request, 'mainapp/posts_filter.html', {'filter':post_filter})
|
[
"notelgamwe@gmail.com"
] |
notelgamwe@gmail.com
|
20222593e4e4c86708abb70c46aac55bb19e07aa
|
61361e75dda44c69b5c152ea932380fdd70c1397
|
/Lecture code/MongoDB 1-3.py
|
bc337c619de0522010d6bdb009168503f3fcecde
|
[] |
no_license
|
dsj529/DataWrangling
|
34b5d5d544e26491dbbe27aa264d1b11075052dd
|
446ec3a41f7b1eb8fa0ca9536d255b7c17dcdaee
|
refs/heads/master
| 2020-04-02T06:54:22.430939
| 2016-07-13T04:32:54
| 2016-07-13T04:32:54
| 63,209,058
| 0
| 0
| null | null | null | null |
UTF-8
|
Python
| false
| false
| 1,650
|
py
|
# To experiment with this code freely you will have to run this code locally.
# We have provided an example json output here for you to look at,
# but you will not be able to run any queries through our UI.
import json
import requests
BASE_URL = "http://musicbrainz.org/ws/2/"
ARTIST_URL = BASE_URL + "artist/"
query_type = { "simple": {},
"atr": {"inc": "aliases+tags+ratings"},
"aliases": {"inc": "aliases"},
"releases": {"inc": "releases"}}
def query_site(url, params, uid="", fmt="json"):
params["fmt"] = fmt
r = requests.get(url + uid, params=params)
print "requesting", r.url
if r.status_code == requests.codes.ok:
return r.json()
else:
r.raise_for_status()
def query_by_name(url, params, name):
params["query"] = "artist:" + name
return query_site(url, params)
def pretty_print(data, indent=4):
if type(data) == dict:
print json.dumps(data, indent=indent, sort_keys=True)
else:
print data
def main():
results = query_by_name(ARTIST_URL, query_type["simple"], "Nirvana")
pretty_print(results)
artist_id = results["artists"][1]["id"]
print "\nARTIST:"
pretty_print(results["artists"][1])
artist_data = query_site(ARTIST_URL, query_type["releases"], artist_id)
releases = artist_data["releases"]
print "\nONE RELEASE:"
pretty_print(releases[0], indent=2)
release_titles = [r["title"] for r in releases]
print "\nALL TITLES:"
for t in release_titles:
print t
if __name__ == '__main__':
main()
|
[
"noreply@github.com"
] |
dsj529.noreply@github.com
|
cade017587a1fddf10ba53e82b78b74261e5d548
|
c5bf6fd200e76dc388770c86b365ba3a0721416a
|
/bcassessment/bcassessment/items.py
|
33fc306e236f3bb27d15c9d521ef9650033b0228
|
[] |
no_license
|
b15h4/ScrapingProjects
|
48a98ee8c23794cbd3ca0e62aeabf2866fdd5851
|
3906091811409ccb9ce12a5099a511e3e6771d50
|
refs/heads/master
| 2020-05-09T11:41:47.420389
| 2020-01-28T22:30:48
| 2020-01-28T22:30:48
| 181,089,587
| 0
| 0
| null | null | null | null |
UTF-8
|
Python
| false
| false
| 1,236
|
py
|
# -*- coding: utf-8 -*-
# Define here the models for your scraped items
#
# See documentation in:
# https://docs.scrapy.org/en/latest/topics/items.html
import scrapy
class BcassessmentItem(scrapy.Item):
# define the fields for your item here like:
# name = scrapy.Field()
PID = scrapy.Field()
Address = scrapy.Field()
Description = scrapy.Field()
Bedrooms = scrapy.Field()
Bathrooms = scrapy.Field()
Carports = scrapy.Field()
Garages = scrapy.Field()
Land_Size = scrapy.Field()
First_Flor_Area = scrapy.Field()
Second_Flor_Area = scrapy.Field()
Basement_Finish_Area = scrapy.Field()
Strata_Area = scrapy.Field()
Building_Storeys = scrapy.Field()
Gross_Leasable_Area = scrapy.Field()
Net_Leasable_Area = scrapy.Field()
No_Of_Apartament_Units = scrapy.Field()
Sold_Date = scrapy.Field()
Sold_Price = scrapy.Field()
Total_Value = scrapy.Field()
Land = scrapy.Field()
Buildings = scrapy.Field()
Previous_Year_Value = scrapy.Field()
Previous_Land = scrapy.Field()
Previous_Buildings = scrapy.Field()
Year_Built = scrapy.Field()
Area_Jurisdiction_Roll = scrapy.Field()
Legal_Description_and_Parcel_ID = scrapy.Field()
|
[
"noreply@github.com"
] |
b15h4.noreply@github.com
|
d55b8c361a5e9a259b03d03acde5dbaabcbf9062
|
f9f698e577ac561e34f6e7851dea807e3429bc57
|
/DevTree/models/admin.py
|
a605ba4f05ca31238bcdd933e0a4ebbbe4c495f3
|
[] |
no_license
|
zjd2035/DevTree
|
22d5a628ad7f2fd03f4065243e5f1f21b6a3e249
|
99a7142aa5ef1f6d4991a2de2516419f83a13dbe
|
refs/heads/master
| 2021-08-26T08:16:05.837792
| 2017-11-22T13:39:09
| 2017-11-22T13:39:09
| 111,687,133
| 0
| 0
| null | null | null | null |
UTF-8
|
Python
| false
| false
| 836
|
py
|
from django.contrib import admin
from DevSky.models.models import *
class DeveloperAdmin(admin.ModelAdmin):
#Fields for creating Developer object
fields = ['user', 'about', 'rating']
#Fields to display for developer
#list_display = ('username',)
class ProjectAdmin(admin.ModelAdmin):
#Fields for creating a Project
fields = ['title', 'languages', 'quick_description', 'full_description', 'owner', 'developers', 'rating' ]
#Fields to display for project
list_display = ('title', 'owner', 'quick_description')
#Search fields
search_fields = ['title', 'owner']
admin.site.register(Developer, DeveloperAdmin)
admin.site.register(Project, ProjectAdmin)
admin.site.register(DeveloperRating)
admin.site.register(ProjectRating)
admin.site.register(DeveloperComment)
admin.site.register(ProjectComment)
|
[
"zjd2035@gmail.com"
] |
zjd2035@gmail.com
|
5fe19acc4de946e7408e43a378612cded89edc88
|
aa91f6e8d59286f65e7f6ed065823c80b7694439
|
/scripts/analysis/baostock/yangxian/yangxian.py
|
adf483a1883a24304215af71aef322817a97af98
|
[
"MIT"
] |
permissive
|
davidyuqiwei/davidyu_stock
|
7f93bcc2c50a0e2c356e3b517dbf7e2e7467093f
|
c234911f49d5980f2dff651333f8ca957333e094
|
refs/heads/master
| 2023-01-07T02:08:04.330559
| 2023-01-02T16:31:37
| 2023-01-02T16:31:37
| 163,694,812
| 13
| 2
| null | 2022-12-18T03:55:10
| 2018-12-31T20:07:30
|
Jupyter Notebook
|
UTF-8
|
Python
| false
| false
| 1,725
|
py
|
from davidyu_cfg import *
from functions.data_dir import *
from functions.get_datetime import *
from functions.run_combine_all_csv import *
from functions.colNames import *
from functions.day_history.kLines import klineDate
from functions.LinearReg import *
from functions.common.dfProcess import *
from functions.common.loadModule.load_module_kdj import *
from scipy.stats import linregress
def stock_data(stock_index,start_date,end_date):
df_dir = os.path.join(data_path,"history_data","baostock","2020-12-17")
df1 = pd.read_csv(os.path.join(df_dir,stock_index+".csv"))
df1 = df1[(df1["dt"]>=start_date)&(df1["dt"]<=end_date)]
df1 = df1.drop_duplicates()
df1 = df1.sort_values("date")
df1["stock_index"] = [ x[3:9] for x in df1["code"]]
return df1
def get_3_pos_line(df1):
df1["line"] = df1["close"] - df1["open"]
df1["line"][df1["line"]>0]=1
df1["line"][df1["line"]<=0]=0
df1['mv_close'] = df1.close.rolling(window=3).mean()
df1['mv_close_120'] = df1.close.rolling(window=120).mean()
df1['mv_close_250'] = df1.close.rolling(window=250).mean()
df1['line_check_5'] = df1.line.rolling(window=5).sum()
df1['line_check_3'] = df1.line.rolling(window=3).sum()
df2 = df1[(df1["line_check_3"]==3)&(df1["close"]<df1['mv_close_250'])]
return df2
if __name__ =='__main__':
stock_index = sys.argv[1]
start_date = '2017-01-01'
end_date = '2020-12-17'
try:
df1 = stock_data(stock_index,start_date,end_date)
df2 = get_3_pos_line(df1)
#df3 = df2.tail(1)
#print("{},{}".format(df2['date'].values,df2['code'].values))
print(df2[["date","code"]].to_string(index=False,header=None))
except:
pass
|
[
"davidyuqiwei@outlook.com"
] |
davidyuqiwei@outlook.com
|
1710dc2802c6eab6b4e818684ead157e2a79fa5c
|
9661db98a656462bd1b666b153f4727d54b436de
|
/sql_queries.py
|
f1b959bd3bccd723530a076000588aab98516936
|
[] |
no_license
|
gingersmog/Udacity-Data-Engineer
|
0f26703e6ba2bfa65b704414a22b6be9971fd940
|
a091acf2bf30d87e3e76ecb40975969d9acc5da5
|
refs/heads/master
| 2020-11-30T07:24:46.081833
| 2019-12-27T00:43:48
| 2019-12-27T00:43:48
| 230,346,961
| 0
| 0
| null | null | null | null |
UTF-8
|
Python
| false
| false
| 1,681
|
py
|
# DROP TABLES
songplay_table_drop = "DROP TABLE IF EXISTS songplays"
user_table_drop = "DROP TABLE IF EXISTS users"
song_table_drop = "DROP TABLE IF EXISTS songs"
artist_table_drop = "DROP TABLE IF EXISTS artists"
time_table_drop = "DROP TABLE IF EXISTS time"
# CREATE TABLES
songplay_table_create = ("CREATE TABLE IF NOT EXISTS songplays (songplay_id varchar, start_time int, user_id int, level varchar, song_id varchar, artist_id varchar, session_id int, location varchar, user_agent varchar)")
user_table_create = ("CREATE TABLE IF NOT EXISTS users (user_id int, first_name varchar, last_name varchar, gender varchar, level varchar)")
song_table_create = ("CREATE TABLE IF NOT EXISTS songs (song_id varchar, title varchar, artist_id varchar, year int, duration numeric)")
artist_table_create = ("CREATE TABLE IF NOT EXISTS artists (artist_id varchar, name varchar, location varchar, latitude numeric, longitude numeric)")
time_table_create = ("CREATE TABLE IF NOT EXISTS time (start_time time, hour int, day int, week int, month int, year int, weekday int)")
# INSERT RECORDS
songplay_table_insert = ("""
""")
user_table_insert = ("""
""")
song_table_insert = ("INSERT INTO songs (song_id, title, artist_id, year, duration) VALUES (%s, %s, %s, %s, %s)", song_data)
artist_table_insert = ("""
""")
time_table_insert = ("""
""")
# FIND SONGS
song_select = ("""
""")
# QUERY LISTS
create_table_queries = [songplay_table_create, user_table_create, song_table_create, artist_table_create, time_table_create]
drop_table_queries = [songplay_table_drop, user_table_drop, song_table_drop, artist_table_drop, time_table_drop]
|
[
"noreply@github.com"
] |
gingersmog.noreply@github.com
|
b1c19b8460916ca3e33a02b99ead0145444c29b0
|
e04c445698d94b7d98dfae2c92b481ce42647e9f
|
/main/views.py
|
e3aa66c2105bd9f85b7b155086bbfa7aa9931ba7
|
[] |
no_license
|
sersergious/taskmanager
|
48475bcbc94d737b5d80c7e7d58d3ab19a009348
|
7005712dfc7e755b89dd176df92ccefa3334b7f1
|
refs/heads/master
| 2023-03-11T07:22:46.016276
| 2021-02-24T22:03:05
| 2021-02-24T22:03:05
| null | 0
| 0
| null | null | null | null |
UTF-8
|
Python
| false
| false
| 704
|
py
|
from django.shortcuts import render, redirect
from .models import Task
from .forms import TaskForm
def index(request):
tasks = Task.objects.all()
return render(request, 'main/index.html', {'title': 'Main page', 'tasks': tasks})
def about(request):
return render(request, 'main/about.html')
def create(request):
error = ''
if request.method == 'POST':
form = TaskForm(request.POST)
if form.is_valid():
form.save()
return redirect('home')
else:
error = 'The form was not correct'
form = TaskForm
context = {
'form': form,
'error': error
}
return render(request, 'main/create.html', context)
|
[
"sergeykuzmin495@gmail.com"
] |
sergeykuzmin495@gmail.com
|
651296c899c3fc4e8987c32de2d56bef52ab2d79
|
57d51b3ec0db9a91054b81c5e33f6926da3cdb09
|
/cesiumpy/util/html.py
|
151be315b75ccce9ba288ba917f54c426e5511de
|
[
"Apache-2.0"
] |
permissive
|
cksammons7/cesiumpy
|
f75aa73f2b836bc16e148d9c66d4056942fa6501
|
0ffa7509fdac03644f0e2fb91385106c40284aa1
|
refs/heads/master
| 2022-05-28T09:45:56.385989
| 2020-04-20T17:20:55
| 2020-04-20T17:20:55
| 257,345,562
| 1
| 0
|
Apache-2.0
| 2020-04-20T16:47:33
| 2020-04-20T16:47:31
| null |
UTF-8
|
Python
| false
| false
| 1,477
|
py
|
#!/usr/bin/env python
# coding: utf-8
from __future__ import unicode_literals
import os
import six
import warnings
def _check_uri(sourceUri):
if not os.path.exists(sourceUri):
msg = "Unable to read specified path, be sure to the output HTML can read the path: {0}"
warnings.warn(msg.format(sourceUri))
return True
def _wrap_uri(sourceUri):
if sourceUri.endswith('.js'):
return '<script src="{0}"></script>'.format(sourceUri)
elif sourceUri.endswith('.css'):
return '<link rel="stylesheet" href="{0}" type="text/css">'.format(sourceUri)
else:
raise ValueError(sourceUri)
def _wrap_script(script):
if not isinstance(script, list):
script = [script]
# filter None and empty str
script = [s for s in script if s is not None and len(s) > 0]
script = _add_indent(script)
return ["""<script type="text/javascript">"""] + script + ["""</script>"""]
def _add_indent(script, indent=2):
""" Indent list of script with specfied number of spaces """
if not isinstance(script, list):
script = [script]
indent = ' ' * indent
return [indent + s for s in script]
def _build_html(*args):
results = []
for a in args:
if isinstance(a, list):
results.extend(a)
elif isinstance(a, six.string_types):
results.append(a)
else:
raise ValueError(type(a))
return os.linesep.join(results)
|
[
"noreply@github.com"
] |
cksammons7.noreply@github.com
|
ef57aa3b2a686f3f147faa1556323040b18f9c63
|
662aef24b143f21b44b74b1db3b7c0a5e6c4e202
|
/python/python标准库urllib.py
|
accaec446456e4a5a2031d688ff317af71f84573
|
[] |
no_license
|
Chan0619/FIS03
|
aa8724bbde0af19389288ecd605c4276a8658195
|
969823101daf1e761c63a959afaaebf56c28e79c
|
refs/heads/master
| 2023-06-23T06:00:22.896841
| 2021-07-20T04:56:31
| 2021-07-20T04:56:31
| 316,534,191
| 0
| 0
| null | null | null | null |
UTF-8
|
Python
| false
| false
| 263
|
py
|
import urllib.request
response = urllib.request.urlopen('http://www.baidu.com')
print(response)
print(response.status)
print(response.read())
print(response.headers) # 头部信息
import math
print(math.ceil(1.5))
print(math.floor(1.5))
print(math.sqrt(25))
|
[
"phoebe.yanxi@gmail.com"
] |
phoebe.yanxi@gmail.com
|
4ffd5dae6f79f999c134250acdaedc6bc61e49ee
|
2e6a926805852fb3d4db9d8d80602aa5cb9d86d3
|
/7.tests/stress_test/locustfile.py
|
7f7db16a514f10de1e0fade7f1f574c0abe82a91
|
[
"Apache-2.0"
] |
permissive
|
Sunao-Yoshii/WelcomePythonExamples
|
4db495ed70ba802e5837344c302a9179b8e6b96b
|
0f929aff7d72075e659e025bca3ce272b952e0a1
|
refs/heads/master
| 2020-08-02T16:10:43.762925
| 2019-11-24T02:03:32
| 2019-11-24T02:03:32
| 211,423,930
| 2
| 0
| null | null | null | null |
UTF-8
|
Python
| false
| false
| 1,315
|
py
|
from locust import HttpLocust, TaskSet, task
import re
class UserBehavior(TaskSet):
def on_start(self):
self.login()
def on_stop(self):
self.logout()
def login(self):
# Cookie などを受け取る目的で
self.client.get("/wp-login.php")
parameters = {
'log': 'TestAccount',
'pwd': 'Fire!!1192',
'redirect_to': 'http://localhost:9000/wp-admin/',
'testcookie': '1'
}
self.client.post("/wp-login.php", parameters)
def logout(self):
response = self.client.get("/wp-admin/")
# ログアウト用 URL 取得
logout_url = re.search(
r'<a class="screen-reader-shortcut" href="http:\/\/localhost\:9000(.+)">ログアウト</a>',
response.text).group(1)
self.client.get(logout_url)
@task
def top(self):
self.client.get("/")
@task(2)
def mypage(self):
with self.client.get("/wp-admin/customize.php", catch_response=True) as response:
if response.status_code != 200:
response.failure("not authenticated???")
@task
def projects(self):
self.client.get("/wp-admin/")
class Wordpress(HttpLocust):
task_set = UserBehavior
min_wait = 500
max_wait = 1000
|
[
"2810066+Sunao-Yoshii@users.noreply.github.com"
] |
2810066+Sunao-Yoshii@users.noreply.github.com
|
7696736a1e831bf9068ea45731dc549e00c9e8c7
|
aa25165d891ae591decc0c214761411bd62681a9
|
/utils/__init__.py
|
1575d6c975074a1397ca2ff39b320a6b8131795f
|
[] |
no_license
|
RaduSoro/MetinBot
|
8a94fa4fac0be2b30eb98049d795859cf1694dd5
|
8936eba0f53cc2e620a80ffabb6b93ad31ea296e
|
refs/heads/master
| 2023-08-26T22:37:07.233612
| 2021-11-02T08:18:55
| 2021-11-02T08:18:55
| 398,293,772
| 0
| 0
| null | null | null | null |
UTF-8
|
Python
| false
| false
| 87
|
py
|
from .utils import *
from .vision import Vision, SnowManFilter, SnowManFilterRedForest
|
[
"radu.sorostinean@ausy.com"
] |
radu.sorostinean@ausy.com
|
1698c854294c33913a617563ebf95c58f4f0da0e
|
30082a1920ed9082864c24f1d3dbaff9bd2cb9a7
|
/nexusdatabroker/TLSNXAPITool1.0/TLSNXAPITool.py
|
a60df0a92657345aa569590e55ee3e85d5b153e5
|
[] |
no_license
|
NARESH9286/nexus9000
|
7885ade015f83d86b96e10a7cc3fbc262d157686
|
b6536915d1d22b087b0c93f775641d4be0bb2307
|
refs/heads/master
| 2021-06-24T22:05:47.759161
| 2017-09-12T06:52:38
| 2017-09-12T06:52:38
| null | 0
| 0
| null | null | null | null |
UTF-8
|
Python
| false
| false
| 5,547
|
py
|
import os
import sys
import yaml
import requests
import subprocess
import logging
import paramiko
# pylint: disable-msg=E0611
from requests.packages.urllib3.exceptions import InsecureRequestWarning
from requests.packages.urllib3.exceptions import SNIMissingWarning
requests.packages.urllib3.disable_warnings(InsecureRequestWarning)
requests.packages.urllib3.disable_warnings(SNIMissingWarning)
class Server:
def __init__(self):
self.conn_type = "https"
with open(INPUTFILE) as file_ptr:
confi = yaml.safe_load(file_ptr)
self.server_ip = confi['ServerIP']['ServerIP1']['ip']
self.username = confi['ServerIP']['ServerIP1']['user']
self.password = confi['ServerIP']['ServerIP1']['password']
self.port = '8443'
self.web_url = ""
self.login_url = ""
self.add_device_url = ""
self.device_response = 0
self.xnc_pwd = str(confi['xnc_password'])
self.xnc_usr = str(confi['xnc_username'])
def ndb_servrer_login(self):
try:
self.web_url = self.conn_type+"://"+self.server_ip+":"\
+self.port+"/monitor/"
self.login_url = self.conn_type+"://"+self.server_ip+":"\
+self.port+"/monitor/j_security_check"
login_payload = {"j_username" : self.xnc_usr, "j_password" : self.xnc_pwd}
with open(INPUTFILE) as file_ptr:
dev_info = yaml.safe_load(file_ptr)
for dic in sorted(dev_info['IP'].keys()):
add_device_payload = dev_info['IP'][dic]
add_device_payload['connectiontype'] = 'NXAPI'
add_device_payload['auxnode'] = 'false'
for key in add_device_payload:
add_device_payload[key] = str(add_device_payload[key])
self.add_device_url = str(self.conn_type+"://"+\
str(self.server_ip)+":"+str(self.port)+\
"/controller/web/devices/extended//element/add")
#pylint: disable=maybe-no-member
with requests.session() as ses:
ses.get(self.web_url, verify=False)
ses.post(self.login_url, data=login_payload, verify=False)
ses.post(self.add_device_url, data=add_device_payload, verify=False)
LOGGER.info("Device - "+add_device_payload['address']+\
" Device added successfully")
except paramiko.SSHException:
LOGGER.error("Device - "+add_device_payload['address']+\
" Failed to add device in NDB")
if __name__ == "__main__":
FILE1 = '/etc/ssh/ssh_config'
DIR = os.path.dirname(__file__)
#sys.stdout = os.devnull
if not os.path.isdir('./Utilities/Log'):
os.mkdir("./Utilities/Log")
#sys.stdout = open(os.devnull, "w")
if len(sys.argv) == 1:
FILENAME = os.path.join(DIR, './Utilities/Log/Logfile.log')
LOGGER = logging.getLogger(__name__)
LOGGER.setLevel(logging.DEBUG)
CON_LOG_HANDLER = logging.StreamHandler()
FILE_LOG_HANDLER = logging.FileHandler(FILENAME)
FILE_LOG_HANDLER.setLevel(logging.DEBUG)
CON_LOG_HANDLER.setLevel(logging.DEBUG)
FORMATTER = logging.Formatter(
'%(asctime)s - %(name)s - %(levelname)s - %(message)s')
FILE_LOG_HANDLER.setFormatter(FORMATTER)
CON_LOG_HANDLER.setFormatter(FORMATTER)
LOGGER.addHandler(FILE_LOG_HANDLER)
LOGGER.addHandler(CON_LOG_HANDLER)
elif len(sys.argv) == 2:
if '--quiet' in sys.argv:
FILENAME = os.path.join(DIR, './Utilities/Log/Logfile.log')
LOGGER = logging.getLogger(__name__)
LOGGER.setLevel(logging.DEBUG)
FILE_LOG_HANDLER = logging.FileHandler(FILENAME)
FILE_LOG_HANDLER.setLevel(logging.DEBUG)
FORMATTER = logging.Formatter(
'%(asctime)s - %(name)s - %(levelname)s - %(message)s')
FILE_LOG_HANDLER.setFormatter(FORMATTER)
LOGGER.addHandler(FILE_LOG_HANDLER)
else:
FILENAME = os.path.join(DIR, './Utilities/Log/Logfile.log')
LOGGER = logging.getLogger(__name__)
LOGGER.setLevel(logging.DEBUG)
CON_LOG_HANDLER = logging.StreamHandler()
FILE_LOG_HANDLER = logging.FileHandler(FILENAME)
FILE_LOG_HANDLER.setLevel(logging.DEBUG)
CON_LOG_HANDLER.setLevel(logging.DEBUG)
FORMATTER = logging.Formatter(
'%(asctime)s - %(name)s - %(levelname)s - %(message)s')
FILE_LOG_HANDLER.setFormatter(FORMATTER)
CON_LOG_HANDLER.setFormatter(FORMATTER)
LOGGER.addHandler(FILE_LOG_HANDLER)
LOGGER.addHandler(CON_LOG_HANDLER)
LOGGER.error(" Run python script without arguments or along "+\
"with --quiet argument")
sys.exit(0)
else:
LOGGER.error(" Run python script without arguments or along "+\
"with --quiet argument")
sys.exit(0)
if '--quiet' in sys.argv:
subprocess.call(" python TLSScript.py --quiet", shell=True)
subprocess.call(" python OpenSSL.py --quiet", shell=True)
else:
subprocess.call(" python TLSScript.py 1", shell=True)
subprocess.call(" python OpenSSL.py 1", shell=True)
INPUTFILE = os.path.join(DIR, './Utilities/Input/inputfile.yaml')
DEV = Server()
DEV.ndb_servrer_login()
os.system("rm -rf ./Utilities/TlsCerts/temp")
os.system("rm -rf ./Utilities/TlsCerts/xnc.log")
|
[
"stumbala@cisco.com"
] |
stumbala@cisco.com
|
82d4bb5adde44d0c3f5d9dcf5fd38838388ed335
|
d1c4189ebec01e2e603618418537039fe88b6572
|
/RNA_1/bls_main.py
|
cee12d48be639efa530ce3a2155dcdaed3b691ae
|
[] |
no_license
|
arashk7/BLS_RNA_Classifier
|
0ee32a19dd8a12f72564a0289df0359aa6b222bf
|
6b320b2e538f54cc4960af338e368468be63ff18
|
refs/heads/master
| 2022-12-24T20:41:49.983691
| 2020-09-25T03:30:10
| 2020-09-25T03:30:10
| 298,180,974
| 0
| 0
| null | null | null | null |
UTF-8
|
Python
| false
| false
| 1,741
|
py
|
import pandas as pd
from sklearn.model_selection import train_test_split
from sklearn.ensemble import RandomForestClassifier
from sklearn import metrics
import seaborn as sn
import matplotlib.pyplot as plt
import numpy as np
from RNA_1.BroadLearningSystem import BLS,BLS_AddEnhanceNodes
import pandas as pd
features1 = pd.read_csv('../Dataset/Sample1.csv')
features1.head()
features2 = pd.read_csv('../Dataset/Sample2.csv')
features2.head()
features = pd.concat([features1, features2])
features.head()
# print(features)
features = features.replace('mod', 0)
features = features.replace('unm', 1)
features = features.replace(np.nan, 0, regex=True)
# print(features)
X = features[['q1', 'q2', 'q3', 'q4', 'q5', 'mis1', 'mis2', 'mis3', 'mis4', 'mis5']].astype(float)
Y = features['sample'].astype(int)
from sklearn.preprocessing import MinMaxScaler
sc = MinMaxScaler(feature_range=(0, 1))
X = sc.fit_transform(X)
X_train, X_test, y_train, y_test = train_test_split(X, Y, test_size=0.25, shuffle=True)
N1 = 10 # # of nodes belong to each window 10
N2 = 10 # # of windows -------Feature mapping layer 10
N3 = 500 # # of enhancement nodes -----Enhance layer 500
L = 5 # # of incremental steps
M1 = 50 # # of adding enhance nodes 50
s = 0.8 # shrink coefficient
C = 2 ** -30 # Regularization coefficient
y_train=y_train.values
y_test = y_test.values
y_train=np.array([y_train]).transpose()
y_test =np.array([y_test]).transpose()
# X_train=X_train[:2]
# y_train=y_train[:2]
print('-------------------BLS_BASE---------------------------')
BLS(X_train, y_train, X_test, y_test, s, C, N1, N2, N3)
print('-------------------BLS_ENHANCE------------------------')
BLS_AddEnhanceNodes(X_train, y_train, X_test, y_test,s,C,N1,N2,N3,L,M1)
|
[
"ahmad.karambakhsh@gmail.com"
] |
ahmad.karambakhsh@gmail.com
|
f09be31c2f91323210f68ba6571fd449ab30ea4b
|
33360ff5f52c1cfbe01bcc001f0533900cee38a0
|
/bme280.py
|
1097289a031a069a2dc516eed3fae10908a2ca0a
|
[] |
no_license
|
diyron/uP_WifiNode
|
49d4090e402618a8b2373c95ea1db834f9ccac45
|
724e82663efab7c67a972b73cc2a8e6789b1020a
|
refs/heads/master
| 2021-01-05T03:14:07.349548
| 2020-02-23T11:32:54
| 2020-02-23T11:32:54
| 240,858,457
| 0
| 0
| null | null | null | null |
UTF-8
|
Python
| false
| false
| 6,575
|
py
|
import time
from ustruct import unpack, unpack_from
from array import array
# BME280 default address.
BME280_I2CADDR = 0x76
# Operating Modes
BME280_OSAMPLE_1 = 1
BME280_OSAMPLE_2 = 2
BME280_OSAMPLE_4 = 3
BME280_OSAMPLE_8 = 4
BME280_OSAMPLE_16 = 5
BME280_REGISTER_CONTROL_HUM = 0xF2
BME280_REGISTER_CONTROL = 0xF4
class BME280:
def __init__(self,
mode=BME280_OSAMPLE_1,
address=BME280_I2CADDR,
i2c=None,
**kwargs):
# Check that mode is valid.
if mode not in [BME280_OSAMPLE_1, BME280_OSAMPLE_2, BME280_OSAMPLE_4,
BME280_OSAMPLE_8, BME280_OSAMPLE_16]:
raise ValueError(
'Unexpected mode value {0}. Set mode to one of '
'BME280_ULTRALOWPOWER, BME280_STANDARD, BME280_HIGHRES, or '
'BME280_ULTRAHIGHRES'.format(mode))
self._mode = mode
self.address = address
if i2c is None:
raise ValueError('An I2C object is required.')
self.i2c = i2c
# load calibration data
dig_88_a1 = self.i2c.readfrom_mem(self.address, 0x88, 26)
dig_e1_e7 = self.i2c.readfrom_mem(self.address, 0xE1, 7)
self.dig_T1, self.dig_T2, self.dig_T3, self.dig_P1, \
self.dig_P2, self.dig_P3, self.dig_P4, self.dig_P5, \
self.dig_P6, self.dig_P7, self.dig_P8, self.dig_P9, \
_, self.dig_H1 = unpack("<HhhHhhhhhhhhBB", dig_88_a1)
self.dig_H2, self.dig_H3 = unpack("<hB", dig_e1_e7)
e4_sign = unpack_from("<b", dig_e1_e7, 3)[0]
self.dig_H4 = (e4_sign << 4) | (dig_e1_e7[4] & 0xF)
e6_sign = unpack_from("<b", dig_e1_e7, 5)[0]
self.dig_H5 = (e6_sign << 4) | (dig_e1_e7[4] >> 4)
self.dig_H6 = unpack_from("<b", dig_e1_e7, 6)[0]
self.i2c.writeto_mem(self.address, BME280_REGISTER_CONTROL,
bytearray([0x3F]))
self.t_fine = 0
# temporary data holders which stay allocated
self._l1_barray = bytearray(1)
self._l8_barray = bytearray(8)
self._l3_resultarray = array("i", [0, 0, 0])
def read_raw_data(self, result):
""" Reads the raw (uncompensated) data from the sensor.
Args:
result: array of length 3 or alike where the result will be
stored, in temperature, pressure, humidity order
Returns:
None
"""
self._l1_barray[0] = self._mode
self.i2c.writeto_mem(self.address, BME280_REGISTER_CONTROL_HUM,
self._l1_barray)
self._l1_barray[0] = self._mode << 5 | self._mode << 2 | 1
self.i2c.writeto_mem(self.address, BME280_REGISTER_CONTROL,
self._l1_barray)
sleep_time = 1250 + 2300 * (1 << self._mode)
sleep_time = sleep_time + 2300 * (1 << self._mode) + 575
sleep_time = sleep_time + 2300 * (1 << self._mode) + 575
time.sleep_us(sleep_time) # Wait the required time
# burst readout from 0xF7 to 0xFE, recommended by datasheet
self.i2c.readfrom_mem_into(self.address, 0xF7, self._l8_barray)
readout = self._l8_barray
# pressure(0xF7): ((msb << 16) | (lsb << 8) | xlsb) >> 4
raw_press = ((readout[0] << 16) | (readout[1] << 8) | readout[2]) >> 4
# temperature(0xFA): ((msb << 16) | (lsb << 8) | xlsb) >> 4
raw_temp = ((readout[3] << 16) | (readout[4] << 8) | readout[5]) >> 4
# humidity(0xFD): (msb << 8) | lsb
raw_hum = (readout[6] << 8) | readout[7]
result[0] = raw_temp
result[1] = raw_press
result[2] = raw_hum
def read_compensated_data(self, result=None):
""" Reads the data from the sensor and returns the compensated data.
Args:
result: array of length 3 or alike where the result will be
stored, in temperature, pressure, humidity order. You may use
this to read out the sensor without allocating heap memory
Returns:
array with temperature, pressure, humidity. Will be the one from
the result parameter if not None
"""
self.read_raw_data(self._l3_resultarray)
raw_temp, raw_press, raw_hum = self._l3_resultarray
# temperature
var1 = ((raw_temp >> 3) - (self.dig_T1 << 1)) * (self.dig_T2 >> 11)
var2 = (((((raw_temp >> 4) - self.dig_T1) *
((raw_temp >> 4) - self.dig_T1)) >> 12) * self.dig_T3) >> 14
self.t_fine = var1 + var2
temp = (self.t_fine * 5 + 128) >> 8
# pressure
var1 = self.t_fine - 128000
var2 = var1 * var1 * self.dig_P6
var2 = var2 + ((var1 * self.dig_P5) << 17)
var2 = var2 + (self.dig_P4 << 35)
var1 = (((var1 * var1 * self.dig_P3) >> 8) +
((var1 * self.dig_P2) << 12))
var1 = (((1 << 47) + var1) * self.dig_P1) >> 33
if var1 == 0:
pressure = 0
else:
p = 1048576 - raw_press
p = (((p << 31) - var2) * 3125) // var1
var1 = (self.dig_P9 * (p >> 13) * (p >> 13)) >> 25
var2 = (self.dig_P8 * p) >> 19
pressure = ((p + var1 + var2) >> 8) + (self.dig_P7 << 4)
# humidity
h = self.t_fine - 76800
h = (((((raw_hum << 14) - (self.dig_H4 << 20) -
(self.dig_H5 * h)) + 16384)
>> 15) * (((((((h * self.dig_H6) >> 10) *
(((h * self.dig_H3) >> 11) + 32768)) >> 10) +
2097152) * self.dig_H2 + 8192) >> 14))
h = h - (((((h >> 15) * (h >> 15)) >> 7) * self.dig_H1) >> 4)
h = 0 if h < 0 else h
h = 419430400 if h > 419430400 else h
humidity = h >> 12
if result:
result[0] = temp
result[1] = pressure
result[2] = humidity
return result
return array("i", (temp, pressure, humidity))
@property
def values(self):
""" human readable values """
t, p, h = self.read_compensated_data()
p = p // 256
pi = p // 100
#pd = p - pi * 100
hi = h // 1024
#hd = h * 100 // 1024 - hi * 100
t = ((t - (t % 10)) / 10) / 10
p = pi
h = hi
return {"Temperatur": t, "Luftfeuchte": h, "Luftdruck": p}
# original:
# return ("{ '{}".format(t / 100), "{}.{:02d} hPa".format(pi, pd),
# "{}.{:02d} %".format(hi, hd))
|
[
"andre.lange.84@googlemail.com"
] |
andre.lange.84@googlemail.com
|
96df10c1e180c5db30031a1555f1186f51078ddd
|
7de48c4ffee8f6ea59c3398a66fc90a27cf563ef
|
/Algorithms/leetcode/[820]单词的压缩编码.py
|
07f6dfe733617bb2c26efaefa608da3f3d3cc325
|
[] |
no_license
|
XiaoqiMa/StudyNotes
|
9a936e119387b65257b08c2b534fb76a95687a34
|
b4f0d27e647f7a25c93b18d148caf143b91337ce
|
refs/heads/master
| 2022-12-03T16:39:19.206183
| 2020-08-25T14:22:40
| 2020-08-25T14:22:40
| 258,927,625
| 0
| 0
| null | null | null | null |
UTF-8
|
Python
| false
| false
| 1,934
|
py
|
# 给定一个单词列表,我们将这个列表编码成一个索引字符串 S 与一个索引列表 A。
#
# 例如,如果这个列表是 ["time", "me", "bell"],我们就可以将其表示为 S = "time#bell#" 和 indexes = [0,
# 2, 5]。
#
# 对于每一个索引,我们可以通过从字符串 S 中索引的位置开始读取字符串,直到 "#" 结束,来恢复我们之前的单词列表。
#
# 那么成功对给定单词列表进行编码的最小字符串长度是多少呢?
#
#
#
# 示例:
#
# 输入: words = ["time", "me", "bell"]
# 输出: 10
# 说明: S = "time#bell#" , indexes = [0, 2, 5] 。
#
#
#
#
# 提示:
#
#
# 1 <= words.length <= 2000
# 1 <= words[i].length <= 7
# 每个单词都是小写字母 。
#
#
# leetcode submit region begin(Prohibit modification and deletion)
from collections import defaultdict
class TrieNode(object):
def __init__(self):
self.children = defaultdict(TrieNode)
self.word_end = False
class Trie(object):
def __init__(self):
self.root = TrieNode()
def insert(self, word):
node = self.root
for w in word:
node = node.children[w]
node.word_end = True
def start_with(self, prefix):
node = self.root
for p in prefix:
if p not in node.children:
return False
node = node.children[p]
return True
class Solution(object):
def minimumLengthEncoding(self, words):
"""
:type words: List[str]
:rtype: int
"""
words = sorted(words, reverse=True, key=len)
count = 0
trie = Trie()
for word in words:
if not trie.start_with(word[::-1]):
count += len(word) + 1
trie.insert(word[::-1])
return count
# leetcode submit region end(Prohibit modification and deletion)
|
[
"xiaoqima2013@gmail.com"
] |
xiaoqima2013@gmail.com
|
5a36493f79cd6e0eef01ad3fd5e7a886f6864e24
|
642a6fe8873e651c64a1f62ce1411dbb12cebd98
|
/Python/Python Basic/w3resources/basic_practice_part1/Q26.py
|
157cfd670068ff4c0d280ab3de36bb7def77b268
|
[] |
no_license
|
lydia0423/Machine_Learning
|
4dc9cdb8049c0e26e527ebd35e100eab60186d7a
|
f4b241efe9522dbcb9dfa923f80d028d64a6054d
|
refs/heads/main
| 2023-03-26T18:17:37.029260
| 2021-03-29T04:17:01
| 2021-03-29T04:17:01
| 336,692,282
| 0
| 0
| null | null | null | null |
UTF-8
|
Python
| false
| false
| 296
|
py
|
#?26. Write a Python program to create a histogram from a given list of integers.
def histogram(items):
for i in items:
output = ""
times = i
while(times > 0):
output += "*"
times -= 1
print(output)
print(histogram([1, 2, 3, 4, 5]))
|
[
"lydia08248@yahoo.com"
] |
lydia08248@yahoo.com
|
c49070446f9c6e5c95466fbae9f419d4d29905da
|
8d87bf5b2c7125fdaf470b065f90f39a9fadc15c
|
/other/subarray.py
|
c70f60225023c6212fbc9f4eedfec13d8bbd360b
|
[] |
no_license
|
mujavidb/algorithm_solutions
|
50ddb65fa89803fbd1d216f0bf09f18552853b33
|
8c5352d9c2381dd0aeae82c77ae61fbbb3583262
|
refs/heads/master
| 2021-01-10T06:50:07.906460
| 2017-05-13T13:48:21
| 2017-05-13T13:48:21
| 46,529,143
| 1
| 0
| null | null | null | null |
UTF-8
|
Python
| false
| false
| 492
|
py
|
# Length of longest subarray of sum less than or equal to k
def max_length(s, k): # These two mark the start and end of the subarray that current used to be.
start = end = total = 0
maxLength = -1 # returns -1 if there is no subsequence that adds up to k.
for i in s:
total += i
end += 1
while total > k: # Shrink the array from the left, until the sum is <= k.
total -= s[start]
start += 1
maxLength = max(maxLength, end - start)
return maxLength
|
[
"sirmuj@gmail.com"
] |
sirmuj@gmail.com
|
93cfdd6f6d45a191d083e9d52f5f9dbc9fb63fdc
|
bb459bb1496ba48b2e4a640c97de28541735a337
|
/Pfeature_scripts/bin_at_wp.py
|
051afaab0ad35c6467e41877c19eb861bc8a55f9
|
[] |
no_license
|
Raman1121/AlgPred
|
43f5f38bdc61b887221c23ef7872f2b8f6942c73
|
c437cc930d76e3a123eafac82cae7e656c91eb03
|
refs/heads/master
| 2020-05-30T15:54:57.319206
| 2019-06-23T06:03:52
| 2019-06-23T06:03:52
| 189,832,818
| 0
| 0
| null | null | null | null |
UTF-8
|
Python
| false
| false
| 3,532
|
py
|
import sys
import os
import pandas as pd
import getopt
def bin_at_wp(file,outt) :
filename, file_extension = os.path.splitext(file)
df=pd.read_csv(file,header=None)
############binary matrix for atoms
f = open('matrix_atom.out', 'w')
sys.stdout = f
print("C,H,N,O,S,")
x = []
for i in range(0,5) :
x.append([])
for j in range(0,5) :
if i == j :
x[i].append(1)
else :
x[i].append(0)
print(x[i][j], end=",")
print("")
f.close()
##############associate binary values to atoms
mat = pd.read_csv("matrix_atom.out")
mat1 = mat.iloc[:,:-1]
mat2 = mat1.transpose()
df1 = pd.read_csv("Data/atom.csv",header=None)
zz = []
kk = pd.DataFrame()
df1 = pd.read_csv("Data/atom.csv",header=None)
for i in range(0,len(df1)) :
zz.append([])
for j in range(0,len(df1[1][i])) :
temp = df1[1][i][j]
zz[i].append(mat2.loc[temp])
f1 = open('bin_atom', 'w')
sys.stdout = f1
for i in range(0,len(zz)) :
for row in zz[i]:
print(",".join(map(str,row)), end=",")
print("")
f1.close()
with open('bin_atom', 'r') as f:
g = list(f)
for i in range(0,len(g)) :
g[i] = g[i].replace(",\n","")
df1["bin"]=g
#########binary atom values for given file
xx=[]
jj = 0
for i in range(0,len(df)) :
xx.append([])
while jj < len(df[0][i]) :
temp=df[0][i][jj]
for k in range(0,len(df1)) :
if temp == df1[0][k][0] :
xx[i].append(df1.iloc[k,2])
jj += 1
jj = 0
f2 = open(outt, 'w')
sys.stdout = f2
for i in range(0,len(xx)) :
for row in xx[i]:
print("".join(map(str,row)), end=",")
print("")
f2.truncate()
os.remove("matrix_atom.out")
os.remove("bin_atom")
def main(argv):
global inputfile
global outputfile
inputfile = ''
outputfile = ''
#option = 1
if len(argv[1:]) == 0:
print ("\nUsage: bin_at_wp.py -i inputfile -o outputfile\n")
print('inputfile : file of peptide/protein sequences for which descriptors need to be generated\n')
print('outputfile : is the file of feature vectors\n')
sys.exit()
try:
opts, args = getopt.getopt(argv,"i:o:",["ifile=","ofile="])
except getopt.GetoptError:
print ("\nUsage: bin_at_wp.py -i inputfile -o outputfile\n")
print('inputfile : file of peptide/protein sequences for which descriptors need to be generated\n')
print('outputfile : is the file of feature vectors\n')
sys.exit(2)
for opt, arg in opts:
if opt == '--help' or opt == '--h':
print ('\nbin_at_wp.py -i inputfile -o outputfile\n')
print('inputfile : file of peptide/protein sequences for which descriptors need to be generated\n')
print('outputfile : is the file of feature vectors\n')
sys.exit()
elif opt in ("-i", "--ifile"):
inputfile = arg
elif opt in ("-o", "--ofile"):
outputfile = arg
bin_at_wp(sys.argv[2],sys.argv[4])
if __name__ == '__main__':
#print(sys.argv)
main(sys.argv[1:])
|
[
"happy07.dutt@gmail.com"
] |
happy07.dutt@gmail.com
|
cb57c6ba8f92e88518798d39afee69873a5fab85
|
77bd1f8e9f9a8e790007db9ddbd346ed851b6880
|
/client/verta/verta/_internal_utils/_artifact_utils.py
|
d2d328ebbbb4d39f3fffa902c1d7d37ead4aa958
|
[
"Apache-2.0"
] |
permissive
|
eyadsibai/modeldb
|
3f5d3fa54c67c3e53e8fa971fc9e659e310d4665
|
b172bd7324d26edb3245f6671ec8351ea03b5832
|
refs/heads/master
| 2023-02-18T08:34:13.269578
| 2023-01-24T21:02:53
| 2023-01-24T21:02:53
| 89,958,297
| 1
| 0
| null | 2017-05-01T20:03:21
| 2017-05-01T20:03:21
| null |
UTF-8
|
Python
| false
| false
| 17,776
|
py
|
# -*- coding: utf-8 -*-
import hashlib
import os
import pickle
import shutil
import tempfile
import zipfile
import cloudpickle
from ..external import six
from .. import __about__
from .importer import maybe_dependency, get_tensorflow_major_version
# default chunk sizes
# these values were all chosen arbitrarily at different times
_64MB = 64 * (10**6) # used for artifact uploads
_32MB = 32 * (10**6) # used in _request_utils
_5MB = 5 * (10**6) # used in this module
# for zip_dir()
# dirs zipped by client need an identifiable extension to unzip durig d/l
ZIP_EXTENSION = "dir.zip"
# NOTE: keep up-to-date with Deployment API
CUSTOM_MODULES_KEY = "custom_modules"
REGISTRY_MODEL_KEY = "model"
MODEL_KEY = "model.pkl" # currently used by experiment run
MODEL_API_KEY = "model_api.json"
# TODO: maybe bind constants for other keys used throughout client
# NOTE: if blocklisting more keys, update the docstrings of
# - RegisteredModel.create_standard_model()
# - RegisteredModelVersion.log_artifact()
# - ExperimentRun.log_artifact()
BLOCKLISTED_KEYS = {
CUSTOM_MODULES_KEY,
MODEL_KEY,
MODEL_API_KEY,
"requirements.txt",
"train_data",
"tf_saved_model",
"setup_script",
}
KERAS_H5PY_ERROR = RuntimeError( # https://github.com/h5py/h5py/issues/1732
"Keras encountered an error saving/loading the model due to a bug in h5py v3.0.0;"
' consider downgrading with `pip install "h5py<3.0.0"`'
)
def validate_key(key):
"""
Validates user-specified artifact key.
Parameters
----------
key : str
Name of artifact.
Raises
------
ValueError
If `key` is blocklisted.
"""
if key in BLOCKLISTED_KEYS:
msg = '"{}" is reserved for internal use; please use a different key'.format(
key
)
raise ValueError(msg)
def get_file_ext(file):
"""
Obtain the filename extension of `file`.
This method assumes `file` is accessible on the user's filesystem.
Parameters
----------
file : str or file handle
Filepath or on-disk file stream.
Returns
-------
str
Filename extension without the leading period.
Raises
------
TypeError
If a filepath cannot be obtained from the argument.
ValueError
If the filepath lacks an extension.
"""
if isinstance(file, six.string_types) and not os.path.isdir(file):
filepath = file
elif hasattr(file, "read") and hasattr(file, "name"): # `open()` object
filepath = file.name
else:
raise TypeError(
"unable to obtain filepath from object of type {}".format(type(file))
)
filename = os.path.basename(filepath).lstrip(".")
try:
_, extension = filename.split(os.extsep, 1)
except ValueError:
six.raise_from(ValueError('no extension found in "{}"'.format(filepath)), None)
else:
return extension
def ext_from_method(method):
"""
Returns an appropriate file extension for a given model serialization method.
Parameters
----------
method : str
The return value of `method` from ``serialize_model()``.
Returns
-------
str or None
Filename extension without the leading period.
"""
if method == "keras":
return "hdf5"
elif method in ("joblib", "cloudpickle", "pickle"):
return "pkl"
elif method == "zip":
return "zip"
elif method == ZIP_EXTENSION: # zipped by client
return ZIP_EXTENSION
elif method is None:
return None
else:
raise ValueError("unrecognized method value: {}".format(method))
def reset_stream(stream):
"""
Resets the cursor of a stream to the beginning.
This is implemented with a try-except because not all file-like objects are guaranteed to have
a ``seek()`` method, so we carry on if we cannot reset the pointer.
Parameters
----------
stream : file-like
A stream that may or may not implement ``seek()``.
"""
try:
stream.seek(0)
except AttributeError:
pass
def ensure_bytestream(obj):
"""
Converts an object into a bytestream.
If `obj` is file-like, its contents will be read into memory and then wrapped in a bytestream.
This has a performance cost, but checking beforehand whether an arbitrary file-like object
returns bytes rather than encoded characters is an implementation nightmare.
If `obj` is not file-like, it will be serialized and then wrapped in a bytestream.
Parameters
----------
obj : file-like or object
Object to convert into a bytestream.
Returns
-------
bytestream : file-like
Buffered bytestream of the serialized artifacts.
method : {"joblib", "cloudpickle", "pickle", None}
Serialization method used to produce the bytestream.
Raises
------
pickle.PicklingError
If `obj` cannot be serialized.
ValueError
If `obj` contains no data.
"""
if hasattr(obj, "read"): # if `obj` is file-like
reset_stream(obj) # reset cursor to beginning in case user forgot
# read first element to check if bytes
try:
chunk = obj.read(1)
except TypeError: # read() doesn't take an argument
pass # fall through to read & cast full stream
else:
if chunk and isinstance(chunk, bytes): # contents are indeed bytes
reset_stream(obj)
return obj, None
else:
pass # fall through to read & cast full stream
# read full stream and cast to bytes
reset_stream(obj)
contents = obj.read() # read to cast into binary
reset_stream(obj) # reset cursor to beginning as a courtesy
if not len(contents):
# S3 raises unhelpful error on empty upload, so catch here
raise ValueError("object contains no data")
bytestring = six.ensure_binary(contents)
bytestream = six.BytesIO(bytestring)
bytestream.seek(0)
return bytestream, None
else: # `obj` is not file-like
bytestream = six.BytesIO()
try:
cloudpickle.dump(obj, bytestream)
except pickle.PicklingError: # can't be handled by cloudpickle
pass
else:
bytestream.seek(0)
return bytestream, "cloudpickle"
if maybe_dependency("joblib"):
try:
maybe_dependency("joblib").dump(obj, bytestream)
except (
NameError, # joblib not installed
pickle.PicklingError,
): # can't be handled by joblib
pass
else:
bytestream.seek(0)
return bytestream, "joblib"
try:
pickle.dump(obj, bytestream)
except pickle.PicklingError: # can't be handled by pickle
six.raise_from(pickle.PicklingError("unable to serialize artifact"), None)
else:
bytestream.seek(0)
return bytestream, "pickle"
def serialize_model(model):
"""
Serializes a model into a bytestream, attempting various methods.
Parameters
----------
model : object or file-like
Model to convert into a bytestream.
Returns
-------
bytestream : file-like
Buffered bytestream of the serialized model.
method : {"joblib", "cloudpickle", "pickle", "keras", None}
Serialization method used to produce the bytestream.
model_type : {"torch", "sklearn", "xgboost", "tensorflow", "custom", "callable"}
Framework with which the model was built.
"""
# if `model` is filesystem path
if isinstance(model, six.string_types):
if os.path.isdir(model):
return zip_dir(model), ZIP_EXTENSION, None
else: # filepath
# open and continue
model = open(model, "rb")
# if `model` is file-like
if hasattr(model, "read"):
try: # attempt to deserialize
reset_stream(model) # reset cursor to beginning in case user forgot
model = deserialize_model(model.read())
except (TypeError, pickle.UnpicklingError):
# unrecognized serialization method and model type
return model, None, None # return bytestream
finally:
reset_stream(model) # reset cursor to beginning as a courtesy
# if `model` is a class
if isinstance(model, six.class_types):
model_type = "class"
bytestream, method = ensure_bytestream(model)
return bytestream, method, model_type
# if`model` is an instance
pyspark_ml_base = maybe_dependency("pyspark.ml.base")
if pyspark_ml_base:
# https://spark.apache.org/docs/latest/api/python/_modules/pyspark/ml/base.html
pyspark_base_classes = (
pyspark_ml_base.Estimator,
pyspark_ml_base.Model,
pyspark_ml_base.Transformer,
)
if isinstance(model, pyspark_base_classes):
temp_dir = tempfile.mkdtemp()
try:
spark_model_dir = os.path.join(temp_dir, "spark-model")
model.save(spark_model_dir)
bytestream = zip_dir(spark_model_dir)
finally:
shutil.rmtree(temp_dir)
# TODO: see if more info would be needed to deserialize in model service
return bytestream, "zip", "pyspark"
for class_obj in model.__class__.__mro__:
module_name = class_obj.__module__
if not module_name:
continue
elif module_name.startswith("torch"):
model_type = "torch"
bytestream, method = ensure_bytestream(model)
break
elif module_name.startswith("sklearn"):
model_type = "sklearn"
bytestream, method = ensure_bytestream(model)
break
elif module_name.startswith("xgboost"):
model_type = "xgboost"
bytestream, method = ensure_bytestream(model)
break
elif module_name.startswith(
"tensorflow.python.keras"
) or module_name.startswith("keras"):
model_type = "tensorflow"
tempf = tempfile.NamedTemporaryFile()
try:
if (
get_tensorflow_major_version() == 2
): # save_format param may not exist in TF 1.X
model.save(
tempf.name, save_format="h5"
) # TF 2.X uses SavedModel by default
else:
model.save(tempf.name)
except TypeError as e:
h5py = maybe_dependency("h5py")
if (
str(e) == "a bytes-like object is required, not 'str'"
and h5py is not None
and h5py.__version__ == "3.0.0"
):
# h5py v3.0.0 improperly checks if a `bytes` contains a `str`.
# Encountering this generic error message here plus the fact
# that h5py==3.0.0 suggests that this is the problem.
six.raise_from(KERAS_H5PY_ERROR, e)
else:
six.raise_from(e, None)
tempf.seek(0)
bytestream = tempf
method = "keras"
break
else:
if hasattr(model, "predict"):
model_type = "custom"
elif callable(model):
model_type = "callable"
else:
model_type = None
bytestream, method = ensure_bytestream(model)
return bytestream, method, model_type
def deserialize_model(bytestring, error_ok=False):
"""
Deserializes a model from a bytestring, attempting various methods.
If the model is unable to be deserialized, the bytes will be returned as a buffered bytestream.
Parameters
----------
bytestring : bytes
Bytes representing the model.
error_ok : bool, default False
Whether to return the serialized bytes if the model cannot be
deserialized. If False, an ``UnpicklingError`` is raised instead.
Returns
-------
model : obj or file-like
Model or buffered bytestream representing the model.
Raises
------
pickle.UnpicklingError
If `bytestring` cannot be deserialized into an object, and `error_ok`
is False.
"""
keras = maybe_dependency("tensorflow.keras")
if keras is not None:
# try deserializing with Keras (HDF5)
with tempfile.NamedTemporaryFile() as tempf:
tempf.write(bytestring)
tempf.seek(0)
try:
return keras.models.load_model(tempf.name)
except AttributeError as e:
h5py = maybe_dependency("h5py")
if (
str(e) == "'str' object has no attribute 'decode'"
and h5py is not None
and h5py.__version__ == "3.0.0"
):
# h5py v3.0.0 returns a `str` instead of a `bytes` to Keras.
# Encountering this generic error message here plus the fact
# that h5py==3.0.0 suggests that this is the problem.
six.raise_from(KERAS_H5PY_ERROR, e)
else:
six.raise_from(e, None)
except (
NameError, # Tensorflow not installed
IOError,
OSError,
): # not a Keras model
pass
bytestream = six.BytesIO(bytestring)
torch = maybe_dependency("torch")
if torch is not None:
try:
return torch.load(bytestream)
except: # not something torch can deserialize
bytestream.seek(0)
# try deserializing with cloudpickle
try:
return cloudpickle.load(bytestream)
except: # not a pickled object
bytestream.seek(0)
if error_ok:
return bytestream
else:
raise pickle.UnpicklingError("unable to deserialize model")
def get_stream_length(stream, chunk_size=_5MB):
"""
Get the length of the contents of a stream.
Parameters
----------
stream : file-like
Stream.
chunk_size : int, default 5 MB
Number of bytes (or whatever `stream` contains) to read into memory at a time.
Returns
-------
length : int
Length of `stream`.
"""
# if it's file handle, get file size without reading stream
filename = getattr(stream, "name", None)
if filename is not None:
try:
return os.path.getsize(filename)
except OSError: # can't access file
pass
# read stream in chunks to get length
length = 0
try:
part_lengths = iter(lambda: len(stream.read(chunk_size)), 0)
for (
part_length
) in (
part_lengths
): # could be sum() but not sure GC runs during builtin one-liner
length += part_length
finally:
reset_stream(stream) # reset cursor to beginning as a courtesy
return length
def calc_sha256(bytestream, chunk_size=_5MB):
"""
Calculates the SHA-256 checksum of a bytestream.
Parameters
----------
bytestream : file-like opened in binary mode
Bytestream.
chunk_size : int, default 5 MB
Number of bytes to read into memory at a time.
Returns
-------
checksum : str
SHA-256 hash of `bytestream`'s contents.
Raises
------
TypeError
If `bytestream` is opened in text mode instead of binary mode.
"""
checksum = hashlib.sha256()
try:
parts = iter(lambda: bytestream.read(chunk_size), b"")
for part in parts:
checksum.update(part)
finally:
reset_stream(bytestream) # reset cursor to beginning as a courtesy
return checksum.hexdigest()
def zip_dir(dirpath, followlinks=True):
"""
ZIPs a directory.
Parameters
----------
dirpath : str
Directory path.
Returns
-------
tempf : :class:`tempfile.NamedTemporaryFile`
ZIP file handle.
"""
e_msg = "{} is not a directory".format(str(dirpath))
if not isinstance(dirpath, six.string_types):
raise TypeError(e_msg)
if not os.path.isdir(dirpath):
raise ValueError(e_msg)
os.path.expanduser(dirpath)
tempf = tempfile.NamedTemporaryFile(suffix="." + ZIP_EXTENSION)
with zipfile.ZipFile(tempf, "w") as zipf:
for root, _, files in os.walk(dirpath, followlinks=followlinks):
for filename in files:
filepath = os.path.join(root, filename)
zipf.write(filepath, os.path.relpath(filepath, dirpath))
tempf.seek(0)
return tempf
def global_read_zipinfo(filename):
"""
Returns a :class:`zipfile.ZipInfo` with ``644`` permissions.
:meth:`zipfile.ZipFile.writestr` creates files with ``600`` [1]_ [2]_,
which means non-owners are unable to read the file, which can be
problematic for custom modules in deployment.
Parameters
----------
filename : str
Name to assign to the file in the ZIP archive.
Returns
-------
zip_info : :class:`zipfile.ZipInfo`
File metadata; the first arg to :meth:`zipfile.ZipFile.writestr`.
References
----------
.. [1] https://github.com/python/cpython/blob/3.9/Lib/zipfile.py#L1791
.. [2] https://bugs.python.org/msg69937
"""
zip_info = zipfile.ZipInfo(filename)
zip_info.external_attr = 0o644 << 16 # ?rw-r--r--
return zip_info
|
[
"noreply@github.com"
] |
eyadsibai.noreply@github.com
|
11e6f191c6447181e1e8bbb7d69eb8cc0edac896
|
65b37165306dff35942a3ce5ba7bb586b85357d9
|
/manage.py
|
890442822014016635d68790f0dddd489a4c63aa
|
[] |
no_license
|
iankigen/drf-pet-store
|
81b803f671c58038da939324a4c28bc21c0b3d25
|
2c87fb5ecde30deac69bfb910d68d495ad932293
|
refs/heads/master
| 2022-05-01T15:01:26.667589
| 2021-03-01T08:40:59
| 2021-03-01T08:40:59
| 197,615,754
| 0
| 0
| null | 2022-04-22T21:50:09
| 2019-07-18T15:45:35
|
Python
|
UTF-8
|
Python
| false
| false
| 811
|
py
|
#!/usr/bin/env python
import os
import sys
if __name__ == "__main__":
os.environ.setdefault("DJANGO_SETTINGS_MODULE", "pet_store_app.settings")
try:
from django.core.management import execute_from_command_line
except ImportError:
# The above import may fail for some other reason. Ensure that the
# issue is really that Django is missing to avoid masking other
# exceptions on Python 2.
try:
import django
except ImportError:
raise ImportError(
"Couldn't import Django. Are you sure it's installed and "
"available on your PYTHONPATH environment variable? Did you "
"forget to activate a virtual environment?"
)
raise
execute_from_command_line(sys.argv)
|
[
"ian.Gabe.ian@gmail.com"
] |
ian.Gabe.ian@gmail.com
|
6f46f4705ff8b0afcec45c239afd53eecb646c90
|
304b0dfb4ae0261d7769963260e99e3b37f588b9
|
/crm1/urls.py
|
f324ade9fceedf306cd1c15467abfdd72c11b5cc
|
[] |
no_license
|
mohdfazlie/crm1
|
bc634af4ae2307e87f3956ec5cd39b9f1c72dc05
|
4c29fa466b6e32f2390bed401464d8621aa497d5
|
refs/heads/main
| 2023-07-13T05:07:21.542412
| 2021-08-26T14:53:21
| 2021-08-26T14:53:21
| null | 0
| 0
| null | null | null | null |
UTF-8
|
Python
| false
| false
| 795
|
py
|
"""crm1 URL Configuration
The `urlpatterns` list routes URLs to views. For more information please see:
https://docs.djangoproject.com/en/3.2/topics/http/urls/
Examples:
Function views
1. Add an import: from my_app import views
2. Add a URL to urlpatterns: path('', views.home, name='home')
Class-based views
1. Add an import: from other_app.views import Home
2. Add a URL to urlpatterns: path('', Home.as_view(), name='home')
Including another URLconf
1. Import the include() function: from django.urls import include, path
2. Add a URL to urlpatterns: path('blog/', include('blog.urls'))
"""
from django.contrib import admin
from django.urls import path, include
urlpatterns = [
path('', include('accounts.urls')),
path('admin/', admin.site.urls),
]
|
[
"59987681+mohdfazlie@users.noreply.github.com"
] |
59987681+mohdfazlie@users.noreply.github.com
|
b468d83e6f86299cc5a6da5cc3813594132a55dc
|
30b232051b10753e9103a70d88a387dfa1aca63f
|
/164.py
|
2f3c183e1682f60665db64f6c436ac296f2bf23b
|
[] |
no_license
|
samrithasudhagar/guvi2
|
fe6d7af8a73cef515991524d7abad754c3700dc5
|
f7eb8a8b2cd701c2708c414939cc139414d3310d
|
refs/heads/master
| 2020-04-20T12:27:47.748859
| 2019-05-26T09:45:42
| 2019-05-26T09:45:42
| 168,843,977
| 0
| 3
| null | null | null | null |
UTF-8
|
Python
| false
| false
| 175
|
py
|
n,k=map(int,input().split())
l=list(map(int,input().split()))
m=0
if k in l:
print(k)
else:
for i in l:
if i>m and i <k:
m=i
print(m)
|
[
"noreply@github.com"
] |
samrithasudhagar.noreply@github.com
|
10e10f48b83650d04791cae5f7b32d3932dabb2d
|
aaf21aa2d251f52e6276bb71f4749a3d1b9e49e9
|
/service/urls.py
|
7e21f058963443c57bce091dd52ec82676bbef18
|
[] |
no_license
|
ClonedOne/parkado_server
|
e6ce7347219cfcbff612c1a1cbef2bbfac63c8c6
|
038f91717765c95173e2cfc28a727cb6d30b37c8
|
refs/heads/master
| 2021-08-31T20:17:30.456706
| 2017-02-08T23:54:47
| 2017-02-08T23:54:47
| 115,138,431
| 0
| 0
| null | null | null | null |
UTF-8
|
Python
| false
| false
| 318
|
py
|
from rest_framework.urlpatterns import format_suffix_patterns
from django.conf.urls import url
from service import views
urlpatterns = [
url(r'^parkings/$', views.ParkingList.as_view()),
url(r'^parkings/(?P<pk>[0-9]+)/$', views.ParkingDetail.as_view()),
]
urlpatterns = format_suffix_patterns(urlpatterns)
|
[
"severi.giorgio@gmail.com"
] |
severi.giorgio@gmail.com
|
6ece01cf76bcfb1cd3fda00de8835a03e0b68a78
|
d1ba73dd406f680342a453439c6105bf5161e470
|
/src/show_results.py
|
f697a4b7df0af5c91195d6ccb3e1fc5add8be696
|
[
"MIT"
] |
permissive
|
yixchen/holistic_scene_human
|
737b5546529bb7aeed4eda7e4f0d61a43c8f1364
|
338b1cea172864b03451fd2f8e347ab751f5e70a
|
refs/heads/master
| 2021-07-10T14:40:00.210618
| 2020-11-24T03:44:24
| 2020-11-24T03:44:24
| 209,427,745
| 46
| 3
| null | null | null | null |
UTF-8
|
Python
| false
| false
| 526
|
py
|
import os
import pickle
from visulization import vis_all
ROOT_DIR = '../data'
IMAGE_PATH = ROOT_DIR + '/image/'
RESULT_PATH = ROOT_DIR + '/result/'
all_image_path = os.listdir(IMAGE_PATH)
for im_name in all_image_path:
image_name = im_name.split('.')[0]
result_path = os.path.join(RESULT_PATH, image_name)
with open(os.path.join(result_path, 'sampler.pickle'), 'rb') as f:
sampler = pickle.load(f, encoding='latin1')
vis_all(sampler, IMAGE_PATH + im_name, result_path, save_image=False)
pass
|
[
"ethanchen@g.ucla.edu"
] |
ethanchen@g.ucla.edu
|
446132c9623e63a2d548c95694992161a693fcff
|
99241f023f2f256c7e6cbce6c6fe3de48627c9f8
|
/build/art-pathplanning-2014/catkin_generated/pkg.develspace.context.pc.py
|
7e3242213bfa7985b084d241f402bb7169f0c3cb
|
[] |
no_license
|
mohsenpour/art-meta
|
02acf40cc00468798113fccf18da59fb82aec698
|
3cb95fc16ae780362d2db52903c818d60e3bfa66
|
refs/heads/master
| 2021-07-25T01:04:38.540640
| 2017-11-04T20:26:31
| 2017-11-04T20:26:31
| null | 0
| 0
| null | null | null | null |
UTF-8
|
Python
| false
| false
| 370
|
py
|
# generated from catkin/cmake/template/pkg.context.pc.in
CATKIN_PACKAGE_PREFIX = ""
PROJECT_PKG_CONFIG_INCLUDE_DIRS = "".split(';') if "" != "" else []
PROJECT_CATKIN_DEPENDS = "".replace(';', ' ')
PKG_CONFIG_LIBRARIES_WITH_PREFIX = "".split(';') if "" != "" else []
PROJECT_NAME = "path_planner"
PROJECT_SPACE_DIR = "/home/art/art-meta/devel"
PROJECT_VERSION = "0.0.0"
|
[
"mo@Mohammads-MacBook-Pro.local"
] |
mo@Mohammads-MacBook-Pro.local
|
34ea96fab7aa2f5a03931e3a87d652ab5f3e629e
|
f6b5f0d72f3e5deb8a913d0a6d541ef3ad5445cb
|
/braintree/transaction.py
|
5c5a315ac43e7831732e7dec0f4ba96100d19ceb
|
[
"MIT"
] |
permissive
|
hathawsh/braintree_python
|
bf056a4d2b8c8b8094f2c876cea4782dc92c715a
|
4ec0f3696438b8c2117f5917834e67ddbf3ebdc7
|
refs/heads/master
| 2021-01-16T00:04:08.883102
| 2013-07-23T22:11:32
| 2013-07-23T22:11:32
| null | 0
| 0
| null | null | null | null |
UTF-8
|
Python
| false
| false
| 15,913
|
py
|
import braintree
import urllib
import warnings
from decimal import Decimal
from braintree.add_on import AddOn
from braintree.disbursement_detail import DisbursementDetail
from braintree.discount import Discount
from braintree.successful_result import SuccessfulResult
from braintree.status_event import StatusEvent
from braintree.error_result import ErrorResult
from braintree.resource import Resource
from braintree.address import Address
from braintree.configuration import Configuration
from braintree.credit_card import CreditCard
from braintree.customer import Customer
from braintree.subscription_details import SubscriptionDetails
from braintree.resource_collection import ResourceCollection
from braintree.transparent_redirect import TransparentRedirect
from braintree.exceptions.not_found_error import NotFoundError
from braintree.descriptor import Descriptor
class Transaction(Resource):
"""
A class representing Braintree Transaction objects.
An example of creating an sale transaction with all available fields::
result = Transaction.sale({
"amount": "100.00",
"order_id": "123",
"channel": "MyShoppingCartProvider",
"credit_card": {
"number": "5105105105105100",
"expiration_date": "05/2011",
"cvv": "123"
},
"customer": {
"first_name": "Dan",
"last_name": "Smith",
"company": "Braintree Payment Solutions",
"email": "dan@example.com",
"phone": "419-555-1234",
"fax": "419-555-1235",
"website": "https://www.braintreepayments.com"
},
"billing": {
"first_name": "Carl",
"last_name": "Jones",
"company": "Braintree",
"street_address": "123 E Main St",
"extended_address": "Suite 403",
"locality": "Chicago",
"region": "IL",
"postal_code": "60622",
"country_name": "United States of America"
},
"shipping": {
"first_name": "Andrew",
"last_name": "Mason",
"company": "Braintree",
"street_address": "456 W Main St",
"extended_address": "Apt 2F",
"locality": "Bartlett",
"region": "IL",
"postal_code": "60103",
"country_name": "United States of America"
}
})
print(result.transaction.amount)
print(result.transaction.order_id)
For more information on Transactions, see https://www.braintreepayments.com/docs/python/transactions/create
"""
def __repr__(self):
detail_list = ["amount", "credit_card", "payment_method_token", "customer_id"]
return super(Transaction, self).__repr__(detail_list)
class CreatedUsing(object):
"""
Constants representing how the transaction was created. Available types are:
* braintree.Transaction.CreatedUsing.FullInformation
* braintree.Transaction.CreatedUsing.Token
"""
FullInformation = "full_information"
Token = "token"
class GatewayRejectionReason(object):
"""
Constants representing gateway rejection reasons. Available types are:
* braintree.Transaction.GatewayRejectionReason.Avs
* braintree.Transaction.GatewayRejectionReason.AvsAndCvv
* braintree.Transaction.GatewayRejectionReason.Cvv
* braintree.Transaction.GatewayRejectionReason.Duplicate
"""
Avs = "avs"
AvsAndCvv = "avs_and_cvv"
Cvv = "cvv"
Duplicate = "duplicate"
class Source(object):
Api = "api"
ControlPanel = "control_panel"
Recurring = "recurring"
class Status(object):
"""
Constants representing transaction statuses. Available statuses are:
* braintree.Transaction.Status.Authorized
* braintree.Transaction.Status.Authorizing
* braintree.Transaction.Status.Failed
* braintree.Transaction.Status.GatewayRejected
* braintree.Transaction.Status.ProcessorDeclined
* braintree.Transaction.Status.Settled
* braintree.Transaction.Status.SettlementFailed
* braintree.Transaction.Status.Settling
* braintree.Transaction.Status.SubmittedForSettlement
* braintree.Transaction.Status.Void
"""
AuthorizationExpired = "authorization_expired"
Authorized = "authorized"
Authorizing = "authorizing"
Failed = "failed"
GatewayRejected = "gateway_rejected"
ProcessorDeclined = "processor_declined"
Settled = "settled"
SettlementFailed = "settlement_failed"
Settling = "settling"
SubmittedForSettlement = "submitted_for_settlement"
Voided = "voided"
class Type(object):
"""
Constants representing transaction types. Available types are:
* braintree.Transaction.Type.Credit
* braintree.Transaction.Type.Sale
"""
Credit = "credit"
Sale = "sale"
@staticmethod
def clone_transaction(transaction_id, params):
return Configuration.gateway().transaction.clone_transaction(transaction_id, params)
@staticmethod
def confirm_transparent_redirect(query_string):
"""
Confirms a transparent redirect request. It expects the query string from the
redirect request. The query string should _not_ include the leading "?" character. ::
result = braintree.Transaction.confirm_transparent_redirect_request("foo=bar&id=12345")
"""
warnings.warn("Please use TransparentRedirect.confirm instead", DeprecationWarning)
return Configuration.gateway().transaction.confirm_transparent_redirect(query_string)
@staticmethod
def credit(params={}):
"""
Creates a transaction of type Credit.
Amount is required. Also, a credit card,
customer_id or payment_method_token is required. ::
result = braintree.Transaction.credit({
"amount": "100.00",
"payment_method_token": "my_token"
})
result = braintree.Transaction.credit({
"amount": "100.00",
"credit_card": {
"number": "4111111111111111",
"expiration_date": "12/2012"
}
})
result = braintree.Transaction.credit({
"amount": "100.00",
"customer_id": "my_customer_id"
})
"""
params["type"] = Transaction.Type.Credit
return Transaction.create(params)
@staticmethod
def find(transaction_id):
"""
Find a transaction, given a transaction_id. This does not return
a result object. This will raise a :class:`NotFoundError <braintree.exceptions.not_found_error.NotFoundError>` if the provided
credit_card_id is not found. ::
transaction = braintree.Transaction.find("my_transaction_id")
"""
return Configuration.gateway().transaction.find(transaction_id)
@staticmethod
def refund(transaction_id, amount=None):
"""
Refunds an existing transaction.
It expects a transaction_id.::
result = braintree.Transaction.refund("my_transaction_id")
"""
return Configuration.gateway().transaction.refund(transaction_id, amount)
@staticmethod
def sale(params={}):
"""
Creates a transaction of type Sale. Amount is required. Also, a credit card,
customer_id or payment_method_token is required. ::
result = braintree.Transaction.sale({
"amount": "100.00",
"payment_method_token": "my_token"
})
result = braintree.Transaction.sale({
"amount": "100.00",
"credit_card": {
"number": "4111111111111111",
"expiration_date": "12/2012"
}
})
result = braintree.Transaction.sale({
"amount": "100.00",
"customer_id": "my_customer_id"
})
"""
params["type"] = Transaction.Type.Sale
return Transaction.create(params)
@staticmethod
def search(*query):
return Configuration.gateway().transaction.search(*query)
@staticmethod
def submit_for_settlement(transaction_id, amount=None):
"""
Submits an authorized transaction for settlement.
Requires the transaction id::
result = braintree.Transaction.submit_for_settlement("my_transaction_id")
"""
return Configuration.gateway().transaction.submit_for_settlement(transaction_id, amount)
@staticmethod
def tr_data_for_credit(tr_data, redirect_url):
"""
Builds tr_data for a Transaction of type Credit
"""
return Configuration.gateway().transaction.tr_data_for_credit(tr_data, redirect_url)
@staticmethod
def tr_data_for_sale(tr_data, redirect_url):
"""
Builds tr_data for a Transaction of type Sale
"""
return Configuration.gateway().transaction.tr_data_for_sale(tr_data, redirect_url)
@staticmethod
def transparent_redirect_create_url():
"""
Returns the url to be used for creating Transactions through transparent redirect.
"""
warnings.warn("Please use TransparentRedirect.url instead", DeprecationWarning)
return Configuration.gateway().transaction.transparent_redirect_create_url()
@staticmethod
def void(transaction_id):
"""
Voids an existing transaction.
It expects a transaction_id.::
result = braintree.Transaction.void("my_transaction_id")
"""
return Configuration.gateway().transaction.void(transaction_id)
@staticmethod
def create(params):
"""
Creates a transaction. Amount and type are required. Also, a credit card,
customer_id or payment_method_token is required. ::
result = braintree.Transaction.sale({
"type": braintree.Transaction.Type.Sale,
"amount": "100.00",
"payment_method_token": "my_token"
})
result = braintree.Transaction.sale({
"type": braintree.Transaction.Type.Sale,
"amount": "100.00",
"credit_card": {
"number": "4111111111111111",
"expiration_date": "12/2012"
}
})
result = braintree.Transaction.sale({
"type": braintree.Transaction.Type.Sale,
"amount": "100.00",
"customer_id": "my_customer_id"
})
"""
return Configuration.gateway().transaction.create(params)
@staticmethod
def clone_signature():
return ["amount", "channel", {"options": ["submit_for_settlement"]}]
@staticmethod
def create_signature():
return [
"amount", "customer_id", "device_session_id", "merchant_account_id", "order_id", "channel",
"payment_method_token", "purchase_order_number", "recurring", "shipping_address_id",
"tax_amount", "tax_exempt", "type", "venmo_sdk_payment_method_code",
"device_data",
{
"credit_card": [
"token", "cardholder_name", "cvv", "expiration_date", "expiration_month", "expiration_year", "number"
]
},
{
"customer": [
"id", "company", "email", "fax", "first_name", "last_name", "phone", "website"
]
},
{
"billing": [
"first_name", "last_name", "company", "country_code_alpha2", "country_code_alpha3",
"country_code_numeric", "country_name", "extended_address", "locality",
"postal_code", "region", "street_address"
]
},
{
"shipping": [
"first_name", "last_name", "company", "country_code_alpha2", "country_code_alpha3",
"country_code_numeric", "country_name", "extended_address", "locality",
"postal_code", "region", "street_address"
]
},
{
"options": [
"store_in_vault", "store_in_vault_on_success", "submit_for_settlement",
"add_billing_address_to_payment_method", "store_shipping_address_in_vault",
"venmo_sdk_session"
]
},
{"custom_fields": ["__any_key__"]},
{"descriptor": ["name", "phone"]}
]
def __init__(self, gateway, attributes):
if "refund_id" in attributes.keys():
self._refund_id = attributes["refund_id"]
del(attributes["refund_id"])
else:
self._refund_id = None
Resource.__init__(self, gateway, attributes)
self.amount = Decimal(self.amount)
if self.tax_amount:
self.tax_amount = Decimal(self.tax_amount)
if "billing" in attributes:
self.billing_details = Address(gateway, attributes.pop("billing"))
if "credit_card" in attributes:
self.credit_card_details = CreditCard(gateway, attributes.pop("credit_card"))
if "customer" in attributes:
self.customer_details = Customer(gateway, attributes.pop("customer"))
if "shipping" in attributes:
self.shipping_details = Address(gateway, attributes.pop("shipping"))
if "add_ons" in attributes:
self.add_ons = [AddOn(gateway, add_on) for add_on in self.add_ons]
if "discounts" in attributes:
self.discounts = [Discount(gateway, discount) for discount in self.discounts]
if "status_history" in attributes:
self.status_history = [StatusEvent(gateway, status_event) for status_event in self.status_history]
if "subscription" in attributes:
self.subscription_details = SubscriptionDetails(attributes.pop("subscription"))
if "descriptor" in attributes:
self.descriptor = Descriptor(gateway, attributes.pop("descriptor"))
if "disbursement_details" in attributes:
self.disbursement_details = DisbursementDetail(attributes.pop("disbursement_details"))
@property
def refund_id(self):
warnings.warn("Please use Transaction.refund_ids instead", DeprecationWarning)
return self._refund_id
@property
def vault_billing_address(self):
"""
The vault billing address associated with this transaction
"""
return self.gateway.address.find(self.customer_details.id, self.billing_details.id)
@property
def vault_credit_card(self):
"""
The vault credit card associated with this transaction
"""
if self.credit_card_details.token is None:
return None
return self.gateway.credit_card.find(self.credit_card_details.token)
@property
def vault_customer(self):
"""
The vault customer associated with this transaction
"""
if self.customer_details.id is None:
return None
return self.gateway.customer.find(self.customer_details.id)
@property
def is_disbursed(self):
return self.disbursement_details.is_valid
|
[
"code@getbraintree.com"
] |
code@getbraintree.com
|
8c86c0ef0e47e5a09d544c09f18d6faa5632dd4e
|
f6f5d82eacd022235b3a85892554fbd9415e3bee
|
/src/saltext/vmware/modules/vmware_vsan.py
|
743ab40200464a7dddb1562166121ce43b7d7fe0
|
[
"Apache-2.0"
] |
permissive
|
garethgreenaway/saltext-vmware
|
a3dd81d3df90ae711178c2232fcdc16b9142d69f
|
f0cd63fc3454be43229658fc4ff825455fc9e691
|
refs/heads/main
| 2023-03-29T05:15:08.409668
| 2021-04-01T00:33:54
| 2021-04-01T00:33:54
| 348,489,162
| 0
| 0
|
Apache-2.0
| 2021-03-25T18:45:29
| 2021-03-16T20:55:30
|
Python
|
UTF-8
|
Python
| false
| false
| 21,601
|
py
|
import logging
import sys
import saltext.vmware.utils.vmware
from salt.utils.decorators import depends, ignores_kwargs
log = logging.getLogger(__name__)
try:
# pylint: disable=no-name-in-module
from pyVmomi import (
vim,
vmodl,
pbm,
VmomiSupport,
)
# pylint: enable=no-name-in-module
# We check the supported vim versions to infer the pyVmomi version
if (
"vim25/6.0" in VmomiSupport.versionMap
and sys.version_info > (2, 7)
and sys.version_info < (2, 7, 9)
):
log.debug(
"pyVmomi not loaded: Incompatible versions " "of Python. See Issue #29537."
)
raise ImportError()
HAS_PYVMOMI = True
except ImportError:
HAS_PYVMOMI = False
__virtualname__ = "vmware_vsan"
def __virtual__():
return __virtualname__
def _get_vsan_eligible_disks(service_instance, host, host_names):
"""
Helper function that returns a dictionary of host_name keys with either a list of eligible
disks that can be added to VSAN or either an 'Error' message or a message saying no
eligible disks were found. Possible keys/values look like:
return = {'host_1': {'Error': 'VSAN System Config Manager is unset ...'},
'host_2': {'Eligible': 'The host xxx does not have any VSAN eligible disks.'},
'host_3': {'Eligible': [disk1, disk2, disk3, disk4],
'host_4': {'Eligible': []}}
"""
ret = {}
for host_name in host_names:
# Get VSAN System Config Manager, if available.
host_ref = _get_host_ref(service_instance, host, host_name=host_name)
vsan_system = host_ref.configManager.vsanSystem
if vsan_system is None:
msg = (
"VSAN System Config Manager is unset for host '{}'. "
"VSAN configuration cannot be changed without a configured "
"VSAN System.".format(host_name)
)
log.debug(msg)
ret.update({host_name: {"Error": msg}})
continue
# Get all VSAN suitable disks for this host.
suitable_disks = []
query = vsan_system.QueryDisksForVsan()
for item in query:
if item.state == "eligible":
suitable_disks.append(item)
# No suitable disks were found to add. Warn and move on.
# This isn't an error as the state may run repeatedly after all eligible disks are added.
if not suitable_disks:
msg = "The host '{}' does not have any VSAN eligible disks.".format(host_name)
log.warning(msg)
ret.update({host_name: {"Eligible": msg}})
continue
# Get disks for host and combine into one list of Disk Objects
disks = _get_host_ssds(host_ref) + _get_host_non_ssds(host_ref)
# Get disks that are in both the disks list and suitable_disks lists.
matching = []
for disk in disks:
for suitable_disk in suitable_disks:
if disk.canonicalName == suitable_disk.disk.canonicalName:
matching.append(disk)
ret.update({host_name: {"Eligible": matching}})
return ret
@depends(HAS_PYVMOMI)
@ignores_kwargs("credstore")
def vsan_add_disks(
host, username, password, protocol=None, port=None, host_names=None, verify_ssl=True
):
"""
Add any VSAN-eligible disks to the VSAN System for the given host or list of host_names.
host
The location of the host.
username
The username used to login to the host, such as ``root``.
password
The password used to login to the host.
protocol
Optionally set to alternate protocol if the host is not using the default
protocol. Default protocol is ``https``.
port
Optionally set to alternate port if the host is not using the default
port. Default port is ``443``.
host_names
List of ESXi host names. When the host, username, and password credentials
are provided for a vCenter Server, the host_names argument is required to
tell vCenter which hosts need to add any VSAN-eligible disks to the host's
VSAN system.
If host_names is not provided, VSAN-eligible disks will be added to the hosts's
VSAN system for the ``host`` location instead. This is useful for when service
instance connection information is used for a single ESXi host.
verify_ssl
Verify the SSL certificate. Default: True
CLI Example:
.. code-block:: bash
# Used for single ESXi host connection information
salt '*' vsphere.vsan_add_disks my.esxi.host root bad-password
# Used for connecting to a vCenter Server
salt '*' vsphere.vsan_add_disks my.vcenter.location root bad-password \
host_names='[esxi-1.host.com, esxi-2.host.com]'
"""
service_instance = saltext.vmware.utils.vmware.get_service_instance(
host=host,
username=username,
password=password,
protocol=protocol,
port=port,
verify_ssl=verify_ssl,
)
host_names = _check_hosts(service_instance, host, host_names)
response = _get_vsan_eligible_disks(service_instance, host, host_names)
ret = {}
for host_name, value in response.items():
host_ref = _get_host_ref(service_instance, host, host_name=host_name)
vsan_system = host_ref.configManager.vsanSystem
# We must have a VSAN Config in place before we can manipulate it.
if vsan_system is None:
msg = (
"VSAN System Config Manager is unset for host '{}'. "
"VSAN configuration cannot be changed without a configured "
"VSAN System.".format(host_name)
)
log.debug(msg)
ret.update({host_name: {"Error": msg}})
else:
eligible = value.get("Eligible")
error = value.get("Error")
if eligible and isinstance(eligible, list):
# If we have eligible, matching disks, add them to VSAN.
try:
task = vsan_system.AddDisks(eligible)
saltext.vmware.utils.vmware.wait_for_task(
task, host_name, "Adding disks to VSAN", sleep_seconds=3
)
except vim.fault.InsufficientDisks as err:
log.debug(err.msg)
ret.update({host_name: {"Error": err.msg}})
continue
except Exception as err: # pylint: disable=broad-except
msg = "'vsphere.vsan_add_disks' failed for host {}: {}".format(
host_name, err
)
log.debug(msg)
ret.update({host_name: {"Error": msg}})
continue
log.debug(
"Successfully added disks to the VSAN system for host '{}'.".format(
host_name
)
)
# We need to return ONLY the disk names, otherwise Message Pack can't deserialize the disk objects.
disk_names = []
for disk in eligible:
disk_names.append(disk.canonicalName)
ret.update({host_name: {"Disks Added": disk_names}})
elif eligible and isinstance(eligible, str):
# If we have a string type in the eligible value, we don't
# have any VSAN-eligible disks. Pull the message through.
ret.update({host_name: {"Disks Added": eligible}})
elif error:
# If we hit an error, populate the Error return dict for state functions.
ret.update({host_name: {"Error": error}})
else:
# If we made it this far, we somehow have eligible disks, but they didn't
# match the disk list and just got an empty list of matching disks.
ret.update(
{
host_name: {
"Disks Added": "No new VSAN-eligible disks were found to add."
}
}
)
return ret
@depends(HAS_PYVMOMI)
@ignores_kwargs("credstore")
def vsan_disable(
host, username, password, protocol=None, port=None, host_names=None, verify_ssl=True
):
"""
Disable VSAN for a given host or list of host_names.
host
The location of the host.
username
The username used to login to the host, such as ``root``.
password
The password used to login to the host.
protocol
Optionally set to alternate protocol if the host is not using the default
protocol. Default protocol is ``https``.
port
Optionally set to alternate port if the host is not using the default
port. Default port is ``443``.
host_names
List of ESXi host names. When the host, username, and password credentials
are provided for a vCenter Server, the host_names argument is required to
tell vCenter which hosts should disable VSAN.
If host_names is not provided, VSAN will be disabled for the ``host``
location instead. This is useful for when service instance connection
information is used for a single ESXi host.
verify_ssl
Verify the SSL certificate. Default: True
CLI Example:
.. code-block:: bash
# Used for single ESXi host connection information
salt '*' vsphere.vsan_disable my.esxi.host root bad-password
# Used for connecting to a vCenter Server
salt '*' vsphere.vsan_disable my.vcenter.location root bad-password \
host_names='[esxi-1.host.com, esxi-2.host.com]'
"""
service_instance = saltext.vmware.utils.vmware.get_service_instance(
host=host,
username=username,
password=password,
protocol=protocol,
port=port,
verify_ssl=verify_ssl,
)
# Create a VSAN Configuration Object and set the enabled attribute to True
vsan_config = vim.vsan.host.ConfigInfo()
vsan_config.enabled = False
host_names = _check_hosts(service_instance, host, host_names)
ret = {}
for host_name in host_names:
host_ref = _get_host_ref(service_instance, host, host_name=host_name)
vsan_system = host_ref.configManager.vsanSystem
# We must have a VSAN Config in place before we can manipulate it.
if vsan_system is None:
msg = (
"VSAN System Config Manager is unset for host '{}'. "
"VSAN configuration cannot be changed without a configured "
"VSAN System.".format(host_name)
)
log.debug(msg)
ret.update({host_name: {"Error": msg}})
else:
try:
# Disable vsan on the host
task = vsan_system.UpdateVsan_Task(vsan_config)
saltext.vmware.utils.vmware.wait_for_task(
task, host_name, "Disabling VSAN", sleep_seconds=3
)
except vmodl.fault.SystemError as err:
log.debug(err.msg)
ret.update({host_name: {"Error": err.msg}})
continue
except Exception as err: # pylint: disable=broad-except
msg = "'vsphere.vsan_disable' failed for host {}: {}".format(
host_name, err
)
log.debug(msg)
ret.update({host_name: {"Error": msg}})
continue
ret.update({host_name: {"VSAN Disabled": True}})
return ret
@depends(HAS_PYVMOMI)
@ignores_kwargs("credstore")
def vsan_enable(
host, username, password, protocol=None, port=None, host_names=None, verify_ssl=True
):
"""
Enable VSAN for a given host or list of host_names.
host
The location of the host.
username
The username used to login to the host, such as ``root``.
password
The password used to login to the host.
protocol
Optionally set to alternate protocol if the host is not using the default
protocol. Default protocol is ``https``.
port
Optionally set to alternate port if the host is not using the default
port. Default port is ``443``.
host_names
List of ESXi host names. When the host, username, and password credentials
are provided for a vCenter Server, the host_names argument is required to
tell vCenter which hosts should enable VSAN.
If host_names is not provided, VSAN will be enabled for the ``host``
location instead. This is useful for when service instance connection
information is used for a single ESXi host.
verify_ssl
Verify the SSL certificate. Default: True
CLI Example:
.. code-block:: bash
# Used for single ESXi host connection information
salt '*' vsphere.vsan_enable my.esxi.host root bad-password
# Used for connecting to a vCenter Server
salt '*' vsphere.vsan_enable my.vcenter.location root bad-password \
host_names='[esxi-1.host.com, esxi-2.host.com]'
"""
service_instance = saltext.vmware.utils.vmware.get_service_instance(
host=host,
username=username,
password=password,
protocol=protocol,
port=port,
verify_ssl=verify_ssl,
)
# Create a VSAN Configuration Object and set the enabled attribute to True
vsan_config = vim.vsan.host.ConfigInfo()
vsan_config.enabled = True
host_names = _check_hosts(service_instance, host, host_names)
ret = {}
for host_name in host_names:
host_ref = _get_host_ref(service_instance, host, host_name=host_name)
vsan_system = host_ref.configManager.vsanSystem
# We must have a VSAN Config in place before we can manipulate it.
if vsan_system is None:
msg = (
"VSAN System Config Manager is unset for host '{}'. "
"VSAN configuration cannot be changed without a configured "
"VSAN System.".format(host_name)
)
log.debug(msg)
ret.update({host_name: {"Error": msg}})
else:
try:
# Enable vsan on the host
task = vsan_system.UpdateVsan_Task(vsan_config)
saltext.vmware.utils.vmware.wait_for_task(
task, host_name, "Enabling VSAN", sleep_seconds=3
)
except vmodl.fault.SystemError as err:
log.debug(err.msg)
ret.update({host_name: {"Error": err.msg}})
continue
except vim.fault.VsanFault as err:
msg = "'vsphere.vsan_enable' failed for host {}: {}".format(
host_name, err
)
log.debug(msg)
ret.update({host_name: {"Error": msg}})
continue
ret.update({host_name: {"VSAN Enabled": True}})
return ret
@depends(HAS_PYVMOMI)
@ignores_kwargs("credstore")
def get_vsan_enabled(
host,
username,
password,
protocol=None,
port=None,
host_names=None,
verify_ssl=True,
):
"""
Get the VSAN enabled status for a given host or a list of host_names. Returns ``True``
if VSAN is enabled, ``False`` if it is not enabled, and ``None`` if a VSAN Host Config
is unset, per host.
host
The location of the host.
username
The username used to login to the host, such as ``root``.
password
The password used to login to the host.
protocol
Optionally set to alternate protocol if the host is not using the default
protocol. Default protocol is ``https``.
port
Optionally set to alternate port if the host is not using the default
port. Default port is ``443``.
host_names
List of ESXi host names. When the host, username, and password credentials
are provided for a vCenter Server, the host_names argument is required to
tell vCenter which hosts to check if VSAN enabled.
If host_names is not provided, the VSAN status will be retrieved for the
``host`` location instead. This is useful for when service instance
connection information is used for a single ESXi host.
verify_ssl
Verify the SSL certificate. Default: True
CLI Example:
.. code-block:: bash
# Used for single ESXi host connection information
salt '*' vsphere.get_vsan_enabled my.esxi.host root bad-password
# Used for connecting to a vCenter Server
salt '*' vsphere.get_vsan_enabled my.vcenter.location root bad-password \
host_names='[esxi-1.host.com, esxi-2.host.com]'
"""
service_instance = saltext.vmware.utils.vmware.get_service_instance(
host=host,
username=username,
password=password,
protocol=protocol,
port=port,
verify_ssl=verify_ssl,
)
host_names = _check_hosts(service_instance, host, host_names)
ret = {}
for host_name in host_names:
host_ref = _get_host_ref(service_instance, host, host_name=host_name)
vsan_config = host_ref.config.vsanHostConfig
# We must have a VSAN Config in place get information about VSAN state.
if vsan_config is None:
msg = "VSAN System Config Manager is unset for host '{}'.".format(host_name)
log.debug(msg)
ret.update({host_name: {"Error": msg}})
else:
ret.update({host_name: {"VSAN Enabled": vsan_config.enabled}})
return ret
@depends(HAS_PYVMOMI)
@ignores_kwargs("credstore")
def get_vsan_eligible_disks(
host,
username,
password,
protocol=None,
port=None,
host_names=None,
verify_ssl=True,
):
"""
Returns a list of VSAN-eligible disks for a given host or list of host_names.
host
The location of the host.
username
The username used to login to the host, such as ``root``.
password
The password used to login to the host.
protocol
Optionally set to alternate protocol if the host is not using the default
protocol. Default protocol is ``https``.
port
Optionally set to alternate port if the host is not using the default
port. Default port is ``443``.
host_names
List of ESXi host names. When the host, username, and password credentials
are provided for a vCenter Server, the host_names argument is required to
tell vCenter which hosts to check if any VSAN-eligible disks are available.
If host_names is not provided, the VSAN-eligible disks will be retrieved
for the ``host`` location instead. This is useful for when service instance
connection information is used for a single ESXi host.
verify_ssl
Verify the SSL certificate. Default: True
CLI Example:
.. code-block:: bash
# Used for single ESXi host connection information
salt '*' vsphere.get_vsan_eligible_disks my.esxi.host root bad-password
# Used for connecting to a vCenter Server
salt '*' vsphere.get_vsan_eligible_disks my.vcenter.location root bad-password \
host_names='[esxi-1.host.com, esxi-2.host.com]'
"""
service_instance = saltext.vmware.utils.vmware.get_service_instance(
host=host,
username=username,
password=password,
protocol=protocol,
port=port,
verify_ssl=verify_ssl,
)
host_names = _check_hosts(service_instance, host, host_names)
response = _get_vsan_eligible_disks(service_instance, host, host_names)
ret = {}
for host_name, value in response.items():
error = value.get("Error")
if error:
ret.update({host_name: {"Error": error}})
continue
disks = value.get("Eligible")
# If we have eligible disks, it will be a list of disk objects
if disks and isinstance(disks, list):
disk_names = []
# We need to return ONLY the disk names, otherwise
# MessagePack can't deserialize the disk objects.
for disk in disks:
disk_names.append(disk.canonicalName)
ret.update({host_name: {"Eligible": disk_names}})
else:
# If we have disks, but it's not a list, it's actually a
# string message that we're passing along.
ret.update({host_name: {"Eligible": disks}})
return ret
@depends(HAS_PYVMOMI)
@_supports_proxies("esxdatacenter", "vcenter")
@_gets_service_instance_via_proxy
def list_default_vsan_policy(service_instance=None):
"""
Returns the default vsan storage policy.
service_instance
Service instance (vim.ServiceInstance) of the vCenter.
Default is None.
.. code-block:: bash
salt '*' vsphere.list_storage_policies
salt '*' vsphere.list_storage_policy policy_names=[policy_name]
"""
profile_manager = salt.utils.pbm.get_profile_manager(service_instance)
policies = salt.utils.pbm.get_storage_policies(profile_manager, get_all_policies=True)
def_policies = [p for p in policies if p.systemCreatedProfileType == "VsanDefaultProfile"]
if not def_policies:
raise VMwareObjectRetrievalError("Default VSAN policy was not " "retrieved")
return _get_policy_dict(def_policies[0])
|
[
"gareth@saltstack.com"
] |
gareth@saltstack.com
|
49adb1a0d02abd33be4f5345e463f2839479722a
|
b2d3bd39b2de8bcc3b0f05f4800c2fabf83d3c6a
|
/examples/pwr_run/checkpointing/short/max_pwr/job18.py
|
8d86f952590173e5884246a93a2efab702b53071
|
[
"MIT"
] |
permissive
|
boringlee24/keras_old
|
3bf7e3ef455dd4262e41248f13c04c071039270e
|
1e1176c45c4952ba1b9b9e58e9cc4df027ab111d
|
refs/heads/master
| 2021-11-21T03:03:13.656700
| 2021-11-11T21:57:54
| 2021-11-11T21:57:54
| 198,494,579
| 0
| 0
| null | null | null | null |
UTF-8
|
Python
| false
| false
| 7,216
|
py
|
"""
#Trains a ResNet on the CIFAR10 dataset.
"""
from __future__ import print_function
import keras
from keras.layers import Dense, Conv2D, BatchNormalization, Activation
from keras.layers import AveragePooling2D, Input, Flatten
from keras.optimizers import Adam
from keras.callbacks import ModelCheckpoint, LearningRateScheduler
from keras.callbacks import ReduceLROnPlateau, TensorBoard
from keras.preprocessing.image import ImageDataGenerator
from keras.regularizers import l2
from keras import backend as K
from keras.models import Model
from keras.datasets import cifar10
from keras.applications.densenet import DenseNet121, DenseNet169, DenseNet201
from keras import models, layers, optimizers
from datetime import datetime
import tensorflow as tf
import numpy as np
import os
import pdb
import sys
import argparse
import time
import signal
import glob
import json
parser = argparse.ArgumentParser(description='Tensorflow Cifar10 Training')
parser.add_argument('--tc', metavar='TESTCASE', type=str, help='specific testcase name')
parser.add_argument('--resume', dest='resume', action='store_true', help='if True, resume training from a checkpoint')
parser.add_argument('--gpu_num', metavar='GPU_NUMBER', type=str, help='select which gpu to use')
parser.set_defaults(resume=False)
args = parser.parse_args()
os.environ["CUDA_DEVICE_ORDER"]="PCI_BUS_ID"
os.environ["CUDA_VISIBLE_DEVICES"]=args.gpu_num
# Training parameters
batch_size = 256
args_lr = 0.007
args_model = 'densenet121'
epoch_begin_time = 0
job_name = sys.argv[0].split('.')[0]
save_files = '/scratch/li.baol/checkpoint_max_pwr/' + job_name + '*'
total_epochs = 19
starting_epoch = 0
# first step is to update the PID
pid_dict = {}
with open('pid_lock.json', 'r') as fp:
pid_dict = json.load(fp)
pid_dict[job_name] = os.getpid()
json_file = json.dumps(pid_dict)
with open('pid_lock.json', 'w') as fp:
fp.write(json_file)
os.rename('pid_lock.json', 'pid.json')
if args.resume:
save_file = glob.glob(save_files)[0]
# epochs = int(save_file.split('/')[4].split('_')[1].split('.')[0])
starting_epoch = int(save_file.split('/')[4].split('.')[0].split('_')[-1])
data_augmentation = True
num_classes = 10
# Subtracting pixel mean improves accuracy
subtract_pixel_mean = True
n = 3
# Model name, depth and version
model_type = args.tc #'P100_resnet50_he_256_1'
# Load the CIFAR10 data.
(x_train, y_train), (x_test, y_test) = cifar10.load_data()
# Normalize data.
x_train = x_train.astype('float32') / 255
x_test = x_test.astype('float32') / 255
# If subtract pixel mean is enabled
if subtract_pixel_mean:
x_train_mean = np.mean(x_train, axis=0)
x_train -= x_train_mean
x_test -= x_train_mean
print('x_train shape:', x_train.shape)
print(x_train.shape[0], 'train samples')
print(x_test.shape[0], 'test samples')
print('y_train shape:', y_train.shape)
# Convert class vectors to binary class matrices.
y_train = keras.utils.to_categorical(y_train, num_classes)
y_test = keras.utils.to_categorical(y_test, num_classes)
if args.resume:
print('resume from checkpoint')
model = keras.models.load_model(save_file)
else:
print('train from start')
model = models.Sequential()
if '121' in args_model:
base_model = DenseNet121(weights=None, include_top=False, input_shape=(32, 32, 3), pooling='avg')
elif '169' in args_model:
base_model = DenseNet169(weights=None, include_top=False, input_shape=(32, 32, 3), pooling='avg')
elif '201' in args_model:
base_model = DenseNet201(weights=None, include_top=False, input_shape=(32, 32, 3), pooling='avg')
model.add(base_model)
#model.add(layers.Flatten())
#model.add(layers.BatchNormalization())
#model.add(layers.Dense(128, activation='relu'))
#model.add(layers.Dropout(0.5))
#model.add(layers.BatchNormalization())
#model.add(layers.Dense(64, activation='relu'))
#model.add(layers.Dropout(0.5))
#model.add(layers.BatchNormalization())
model.add(layers.Dense(10, activation='softmax'))#, kernel_initializer='he_uniform'))
model.compile(loss='categorical_crossentropy',
optimizer=Adam(lr=args_lr),
metrics=['accuracy'])
#model.summary()
print(model_type)
#pdb.set_trace()
current_epoch = 0
################### connects interrupt signal to the process #####################
def terminateProcess(signalNumber, frame):
# first record the wasted epoch time
global epoch_begin_time
epoch_waste_time = int(time.time() - epoch_begin_time)
epoch_waste_dict = {}
with open('epoch_waste.json', 'r') as fp:
epoch_waste_dict = json.load(fp)
epoch_waste_dict[job_name] += epoch_waste_time
json_file3 = json.dumps(epoch_waste_dict)
with open('epoch_waste.json', 'w') as fp:
fp.write(json_file3)
print('checkpointing the model triggered by kill -15 signal')
# delete whatever checkpoint that already exists
for f in glob.glob(save_files):
os.remove(f)
model.save('/scratch/li.baol/checkpoint_max_pwr/' + job_name + '_' + str(current_epoch) + '.h5')
print ('(SIGTERM) terminating the process')
checkpoint_dict = {}
with open('checkpoint.json', 'r') as fp:
checkpoint_dict = json.load(fp)
checkpoint_dict[job_name] = 1
json_file3 = json.dumps(checkpoint_dict)
with open('checkpoint.json', 'w') as fp:
fp.write(json_file3)
sys.exit()
signal.signal(signal.SIGTERM, terminateProcess)
#################################################################################
logdir = '/scratch/li.baol/tsrbrd_log/job_runs/' + model_type + '/' + job_name
tensorboard_callback = TensorBoard(log_dir=logdir)#, update_freq='batch')
class PrintEpoch(keras.callbacks.Callback):
def on_epoch_begin(self, epoch, logs=None):
global current_epoch
#remaining_epochs = epochs - epoch
current_epoch = epoch
print('current epoch ' + str(current_epoch))
global epoch_begin_time
epoch_begin_time = time.time()
my_callback = PrintEpoch()
callbacks = [tensorboard_callback, my_callback]
#[checkpoint, lr_reducer, lr_scheduler, tensorboard_callback]
# Run training
# creates an file if job qualified for checkpoint
open('ckpt_qual/' + job_name + '.txt', 'a').close()
model.fit(x_train, y_train,
batch_size=batch_size,
epochs=round(total_epochs/2),
validation_data=(x_test, y_test),
shuffle=True,
callbacks=callbacks,
initial_epoch=starting_epoch,
verbose=1
)
# Score trained model.
scores = model.evaluate(x_test, y_test, verbose=1)
print('Test loss:', scores[0])
print('Test accuracy:', scores[1])
finish_dict = {}
while True:
if os.path.exists('finish.json'):
try:
os.rename('finish.json', 'finish_lock.json')
break
except Exception:
pass
else:
time.sleep(1)
with open('finish_lock.json', 'r') as fp:
finish_dict = json.load(fp)
finish_dict[job_name] = 1
json_file2 = json.dumps(finish_dict)
with open('finish_lock.json', 'w') as fp:
fp.write(json_file2)
os.rename('finish_lock.json', 'finish.json')
|
[
"baolin.li1994@gmail.com"
] |
baolin.li1994@gmail.com
|
38e2ca7ad8bd0105e127dde9200cf8a028208f9b
|
de998d0170e81d61982cef9c3181c41fc926bac3
|
/Hardware/PCB/KiCadScript/KiAuto/KiPcb.py
|
3b58c4d5c530bf82be8c8e8d499bea6348ffbf05
|
[
"BSD-3-Clause"
] |
permissive
|
ymei/TMSPlane
|
f100294e039bb73e77964cf9c38930160d5f86d9
|
3e30749440b1a8a0fd81a848b368cfbce10dfb86
|
refs/heads/master
| 2021-01-23T07:44:23.156705
| 2018-03-25T03:16:43
| 2018-03-25T03:16:43
| 86,438,489
| 1
| 2
| null | null | null | null |
UTF-8
|
Python
| false
| false
| 6,380
|
py
|
from __future__ import print_function
import pcbnew
from .util import *
## Class operating on an already established board
#
class KiPcbOp(object):
## @param[in] board Already open and established board object.
def __init__(self, board):
self._board = board
self._io = pcbnew.PCB_IO()
self.layerI = {pcbnew.BOARD_GetStandardLayerName(n):n for n in range(pcbnew.PCB_LAYER_ID_COUNT)}
self.layerN = {s:n for n, s in self.layerI.iteritems()}
# generate a LUT with shape integers to a string
self.padShapes = {
pcbnew.PAD_SHAPE_CIRCLE: "PAD_SHAPE_CIRCLE",
pcbnew.PAD_SHAPE_OVAL: "PAD_SHAPE_OVAL",
pcbnew.PAD_SHAPE_RECT: "PAD_SHAPE_RECT",
pcbnew.PAD_SHAPE_TRAPEZOID: "PAD_SHAPE_TRAPEZOID"
}
if hasattr(pcbnew, 'PAD_SHAPE_ROUNDRECT'):
self.padShapes[pcbnew.PAD_SHAPE_ROUNDRECT] = "PAD_SHAPE_ROUNDRECT"
# returns a dictionary netcode:netinfo_item
self.netCodes = self._board.GetNetsByNetcode()
## called by __init__()
def compute_layer_table(self):
self.layerTable = {}
for i in range(pcbnew.PCB_LAYER_ID_COUNT):
self.layerTable[i] = self._board.GetLayerName(i)
return self.layerTable
def layer_id_to_name(self, id):
return self.layerN[id]
def layer_name_to_id(self, name):
return self.layerI[name]
def print_list_of_nets(self):
for netcode, net in self.netCodes.items():
print("netcode: {}, name: {}".format(netcode, net.GetNetname()))
def get_board_boundary(self):
rect = None
for d in self._board.GetDrawings():
if (d.GetLayerName() != "Edge.Cuts"):
continue
if (rect == None):
rect = d.GetBoundingBox()
else:
rect.Merge(d.GetBoundingBox())
if rect:
return [pcbnew.wxPoint(rect.Centre().x-rect.GetWidth()/2, rect.Centre().y-rect.GetHeight()/2),
pcbnew.wxPoint(rect.Centre().x-rect.GetWidth()/2, rect.Centre().y+rect.GetHeight()/2),
pcbnew.wxPoint(rect.Centre().x+rect.GetWidth()/2, rect.Centre().y+rect.GetHeight()/2),
pcbnew.wxPoint(rect.Centre().x+rect.GetWidth()/2, rect.Centre().y-rect.GetHeight()/2)]
else:
return None
## @param[in] layerId consult layerTable, usually 0: F.Cu, 31: B.Cu
def place_footprint(self, lib, name, ref="", loc=(0,0), layerId=0):
mod = self._io.FootprintLoad(lib, name)
p = pcbnew.wxPoint(loc[0], loc[1]) if type(loc) == tuple or type(loc) == list else loc
mod.SetPosition(p)
mod.SetLayer(layerId)
mod.SetReference(ref)
self._board.Add(mod)
def move_footprint(self, ref, loc, rot=0.0, layerId=None, flip=False):
mod = self._board.FindModuleByReference(ref)
if mod == None:
return None
p = pcbnew.wxPoint(loc[0], loc[1]) if type(loc) == tuple or type(loc) == list else loc
mod.SetPosition(p)
if layerId != None:
mod.SetLayer(layerId)
if flip:
mod.Flip(p)
mod.SetOrientation(rot)
def set_footprint_nets(self, ref="", pinNet={1:'/VDD', 2:'/GND'}):
mod = self._board.FindModuleByReference(ref)
if mod == None:
return None
for p in mod.Pads():
netname = pinNet[int(p.GetPadName())]
netcode = self._board.GetNetcodeFromNetname(netname)
print(netname, netcode)
p.SetNetCode(netcode)
return True
def get_fp_pad_pos_netname(self, ref):
mod = self._board.FindModuleByReference(ref)
if mod == None:
return None
ppn = {}
for p in mod.Pads():
pn = int(p.GetPadName())
if pn in ppn:
if len(ppn[pn][0]) > 0:
ppn[pn][0].append(p.GetCenter())
else:
ppn[pn] = ([p.GetCenter()], p.GetNetname())
return ppn
def add_track(self, posList=[[0,0], [1,1]], width=None, layerId=0, netName="/GND"):
netcode = self._board.GetNetcodeFromNetname(netName)
for i in xrange(len(posList)-1):
t = posList[i]
p1 = pcbnew.wxPoint(t[0], t[1]) if type(t) == tuple or type(t) == list else t
t = posList[i+1]
p2 = pcbnew.wxPoint(t[0], t[1]) if type(t) == tuple or type(t) == list else t
if width == None:
width = self._board.GetDesignSettings().GetCurrentTrackWidth()
track = pcbnew.TRACK(self._board)
track.SetStart(p1)
track.SetEnd(p2)
track.SetWidth(width)
track.SetLayer(layerId)
self._board.Add(track)
track.SetNetCode(netcode)
def add_via(self, pos=[0,0], layerIdPair=(0, 31), netName="/GND", size=None, drill=None, vType=pcbnew.VIA_THROUGH):
netcode = self._board.GetNetcodeFromNetname(netName)
if size == None:
size = self._board.GetDesignSettings().GetCurrentViaSize()
if drill == None:
drill = self._board.GetDesignSettings().GetCurrentViaDrill()
if vType == pcbnew.VIA_THROUGH:
via = pcbnew.VIA(self._board)
via.SetWidth(size)
p1 = pcbnew.wxPoint(pos[0], pos[1]) if type(pos) == tuple or type(pos) == list else pos
via.SetStart(p1)
via.SetEnd(p1)
via.SetLayerPair(layerIdPair[0], layerIdPair[1])
via.SetDrill(drill)
via.SetViaType(pcbnew.VIA_THROUGH)
self._board.Add(via)
via.SetNetCode(netcode)
def add_zone(self, corners=None, layerId=0, netName="/GND",
clearance=mil(4), minWidth=mil(4), padConn=pcbnew.PAD_ZONE_CONN_FULL):
netcode = self._board.GetNetcodeFromNetname(netName)
area = self._board.InsertArea(netcode, 0, layerId, int(corners[0][0]), int(corners[0][1]),
pcbnew.CPolyLine.DIAGONAL_EDGE)
area.SetZoneClearance(clearance)
area.SetMinThickness(minWidth)
area.SetPadConnection(padConn)
for p in corners[1:]:
pw = pcbnew.wxPoint(p[0], p[1]) if type(p) == tuple or type(p) == list else p
area.AppendCorner(pw, -1)
area.Hatch()
|
[
"yuan.mei@gmail.com"
] |
yuan.mei@gmail.com
|
8fb33330b1462f23987648fc31eb06140b7e5caa
|
1e03cd80d27d35ffdc8f68f70a36a461eaae4b9d
|
/apps/common/views.py
|
9da3528729470edad1fb1663ca5e9291ee3c0179
|
[] |
no_license
|
paddy375691/flask_zlbbs
|
bee8f15497c58bd5f1f614d6a686b93301f93d0a
|
90fb142b33aecca6ff66013953ecf6e3e39b6139
|
refs/heads/master
| 2023-02-04T07:55:00.798789
| 2020-12-25T08:51:13
| 2020-12-25T08:51:13
| 324,265,174
| 0
| 0
| null | null | null | null |
UTF-8
|
Python
| false
| false
| 144
|
py
|
from flask import Blueprint
bp = Blueprint('common', __name__, url_prefix='/common')
@bp.route('/')
def index():
return 'common index'
|
[
"you@example.com"
] |
you@example.com
|
f2edf4b8d81e383e300f59c74dc202894f3b6498
|
18c0c008de1f1cb4c6d194fd95e114f0c06476fc
|
/listcomprehension/startso.py
|
cab91ee67e82cbce8ffbb284609b470dc5f3934e
|
[] |
no_license
|
betillogalvanfbc/pythoncourse
|
6e18065e463d8b3fde85a205b3a473703fb9a690
|
8e49ec581b1e4f5f477699cf3f3bc030f6478bbd
|
refs/heads/master
| 2022-05-14T12:25:03.342382
| 2020-04-23T23:07:34
| 2020-04-23T23:07:34
| 258,351,481
| 0
| 0
| null | null | null | null |
UTF-8
|
Python
| false
| false
| 226
|
py
|
friends = ["Beto", "Bernanrdo", "Benito"]
starts_s = [friend for friend in friends if friend.startswith("B")]
print(starts_s)
print(friends)
print(friends is starts_s)
print("friends:", id(friends), "starts_s:", id(starts_s))
|
[
"alberto.galvanfbc@gmail.com"
] |
alberto.galvanfbc@gmail.com
|
9be5e10a4b35c6bcf8efca2e17740a47f51645d7
|
1eba9645eb6922a415cf7c08cd76e3b9f489513b
|
/QuoteEngine/QuoteModel.py
|
89d9de7723398ad7a91577c858d42470df3af006
|
[] |
no_license
|
t4ngojuli3tt/UdacityMeme
|
96741e35162730c675af833302a95501bf5fbbef
|
9b710f3421653c1aee61a040d6da3909b28bdd88
|
refs/heads/main
| 2023-03-04T20:52:44.622380
| 2021-02-12T20:52:35
| 2021-02-12T20:52:35
| 338,428,171
| 0
| 0
| null | null | null | null |
UTF-8
|
Python
| false
| false
| 183
|
py
|
class QuoteModel:
def __init__(self, body, author):
self.author = author
self.body = body
def __repr__(self):
return f'"{self.body}" - {self.author}'
|
[
"tomasz.janiczko@gmail.com"
] |
tomasz.janiczko@gmail.com
|
c713a4ef7e8152c9c7d6aad296a20e9a6724c730
|
4909d98c871fbe71daf7538a2a074751333552b8
|
/file_io.py
|
2397248b04b3ebda2316861d725e553b189d0e68
|
[] |
no_license
|
jamathis77/Python-fundamentals
|
68653da4928b97a336888f2aac840ad26f6d4a6e
|
2e35a1ef384671b912dae81ba96827ec45576857
|
refs/heads/master
| 2020-08-14T09:40:09.375318
| 2019-10-14T20:39:04
| 2019-10-14T20:39:04
| 215,143,182
| 0
| 0
| null | null | null | null |
UTF-8
|
Python
| false
| false
| 471
|
py
|
# # Create a python file in python, that itself modifies the testing.txt file to say
# # I'm sorry, I cannot do that.
# python_file = open("./file_io_hub/change_text.py", "x")
# valid_code = """
# to_change = open("./testing.txt", "w")
# to_change.write("I'm sorry, I cannot do that")
# to_change.close()
# """
# python_file.write(valid_code)
# python_file.close()
with open("with_statement.py", "w") as f:
f.write("# Im a comment\nprint('Hello world')")
|
[
"jamathis77@gmail.com"
] |
jamathis77@gmail.com
|
08389cf04636acaaeff77c935cbdfb741525b0e5
|
5c8b1cb941467020ec52c7bea0a03dff696881b9
|
/simorgh/polls/forms.py
|
4edcd15dd5147c4f47f46c13bd5e269bc2bf07ef
|
[] |
no_license
|
amir1370/simorgh
|
43dfaa4d3c7f4a4489cc35e84f6e24a61408649f
|
83788712a8ef9f5001446a9370aa0e0611eb91fc
|
refs/heads/master
| 2020-05-20T02:22:53.473410
| 2019-07-06T20:03:08
| 2019-07-06T20:03:08
| 185,330,230
| 0
| 0
| null | null | null | null |
UTF-8
|
Python
| false
| false
| 673
|
py
|
from django import forms
from .models import Choice, Question
from django.forms import ModelForm, formset_factory
from django.forms.widgets import RadioSelect
class ChoiceForm(ModelForm):
EXCELLENT, GOOD, MEDIUM, BAD = 'EX', 'GO', 'ME', 'BA'
question_choices = (
(EXCELLENT, 'عالی'),
(GOOD, 'خوب'),
(MEDIUM, 'متوسط'),
(BAD, 'ضعیف')
)
choice_text = forms.ChoiceField(choices=question_choices, widget=RadioSelect(attrs={"required":"required"}), required=True)
class Meta:
model = Choice
fields = ['choice_text']
ChoiceFormset = formset_factory(ChoiceForm, extra=Question.objects.count())
|
[
"amirhoseintaherijam@gmail.com"
] |
amirhoseintaherijam@gmail.com
|
2102290e12b10e2260838df9b63b6bbc4466e206
|
83d5439a809aaf029cfbab22d776a99c8ea85bd2
|
/vista1/migrations/0002_auto_20210412_1851.py
|
2b48628f1e6af25952df7855c37cd8adb70bcf23
|
[] |
no_license
|
lapcpc/exame2
|
cca8933ecc7049737d0c327cddf0cc4e2113ddbb
|
9695dcda5c1668c5bc92073c4c4fb1c2d98e43e5
|
refs/heads/main
| 2023-04-10T10:07:03.448832
| 2021-04-13T00:21:52
| 2021-04-13T00:21:52
| 357,371,901
| 0
| 0
| null | null | null | null |
UTF-8
|
Python
| false
| false
| 563
|
py
|
# Generated by Django 2.2.6 on 2021-04-12 23:51
from django.db import migrations, models
class Migration(migrations.Migration):
dependencies = [
('vista1', '0001_initial'),
]
operations = [
migrations.AlterField(
model_name='registro',
name='oxigenacion',
field=models.IntegerField(blank=True, null=True),
),
migrations.AlterField(
model_name='registro',
name='temperatura',
field=models.IntegerField(blank=True, null=True),
),
]
|
[
"lapcpc@hotmail.com"
] |
lapcpc@hotmail.com
|
e037ce0f746846b6294b60c884db7209be1e7464
|
efbc8c73e9ac5cbcb9321518ab06b3965369a5f0
|
/SWEA/D2/1974_스도쿠 검증.py
|
5c70e3a29c30f7e85436a3f0b4edb86ade4466a6
|
[] |
no_license
|
AshOil/APS
|
56b9395dcbb8eeec87a047407d4326b879481612
|
fe5a2cd63448fcc4b11b5e5bc060976234ed8eea
|
refs/heads/master
| 2023-07-15T17:32:20.684742
| 2021-08-23T13:04:05
| 2021-08-23T13:04:05
| 283,709,661
| 0
| 0
| null | null | null | null |
UTF-8
|
Python
| false
| false
| 1,389
|
py
|
import sys
sys.stdin = open('input_data/1974.txt',"r")
num_dict = {}
T = int(input())
for t in range(1, T+1):
for tt in range(1,10):
num_dict[tt] = list(map(int, input() .split()))
result = True
# 가로부터 검사하자
for hori in num_dict.values():
if sorted(hori) != [1, 2, 3, 4, 5, 6, 7, 8, 9]:
result = False
#세로 검사
for num in range(9):
verti_check = []
for verti in num_dict.values():
verti_check.append(verti[num])
verti_result = verti_check
if sorted(verti_check) != [1, 2, 3, 4, 5, 6, 7, 8, 9]:
result = False
#블록검사
line_start = 0
line_end = 3
block_list = list(num_dict.values())
for __ in range(3):
turn_block_list = block_list[line_start:line_end]
block_start = 0
block_end = 3
for _ in range(3):
block_check = []
for turn in range(3):
for block in turn_block_list[turn][block_start:block_end]:
block_check.append(block)
block_start += 3
block_end += 3
if sorted(block_check) != [1, 2, 3, 4, 5, 6, 7, 8, 9]:
result = False
line_start += 3
line_end += 3
if result:
print('#{} 1'.format(t))
else:
print('#{} 0'.format(t))
|
[
"ka0@kakao.com"
] |
ka0@kakao.com
|
88ac908df79f1845c722a579324a1598126bc0dd
|
644984edc7bd12c395f20b204e0201aca068c541
|
/Basic Data Structures/array/leet_154_FindMinimuminRotatedSortedArrayII.py
|
0fae83c285503ec1c802172d4ef714d78f74bca3
|
[
"MIT"
] |
permissive
|
rush2catch/algorithms-leetcode
|
3bd7ad0fc00d303b520ab6b1f0b3ba84cf449115
|
38a5e6aa33d48fa14fe09c50c28a2eaabd736e55
|
refs/heads/master
| 2021-09-06T17:45:20.295678
| 2018-02-09T08:07:54
| 2018-02-09T08:07:54
| 91,420,882
| 0
| 0
| null | null | null | null |
UTF-8
|
Python
| false
| false
| 880
|
py
|
# Problem: Find Minimum in Rotated Sorted Array
# Difficulty: Medium
# Category: Array
# Leetcode 154: https://leetcode.com/problems/find-minimum-in-rotated-sorted-array/description/
"""
Suppose an array sorted in ascending order is rotated at some pivot unknown to you beforehand.
(i.e., 0 1 2 4 5 6 7 might become 4 5 6 7 0 1 2).
Find the minimum element.
You may assume no duplicate exists in the array.
"""
class Solution(object):
def find_mini(self, nums):
if len(nums) == 0:
return None
start = 0
end = len(nums) - 1
while start < end:
mid = (start + end) // 2
if nums[mid] < nums[end]:
end = mid
elif nums[mid] > nums[end]:
start = mid + 1
else:
end -= 1
return nums[start]
obj = Solution()
arr1 = [4, 5, 6, 7, 0, 1, 2]
arr2 = [3, 4, 5, 0, 1, 1, 1, 2]
arr3 = [1]
print(obj.find_mini(arr1), obj.find_mini(arr2), obj.find_mini(arr3))
|
[
"jiangyuan1988@gmail.com"
] |
jiangyuan1988@gmail.com
|
f6afdb1bcdb602520eb1f466922a2c9ea015a1db
|
2da476c3ba893f471760edc3553c5d4a98067370
|
/swapcase.py
|
fb2e8fa940996ddd13131980a5a8088c83233d3b
|
[] |
no_license
|
Yashasvini18/EDUYEAR-PYTHON-20
|
4b0b4e37feca34ad76f7d11599f1243c0ae4083b
|
43900aa11a1c57c3834d21ef94fb383be00e6fa6
|
refs/heads/main
| 2023-04-20T04:42:05.263050
| 2021-05-15T12:59:14
| 2021-05-15T12:59:14
| 364,542,373
| 0
| 0
| null | null | null | null |
UTF-8
|
Python
| false
| false
| 83
|
py
|
input_str=input("Enter the word in alternate case ")
print(input_str.swapcase())
|
[
"noreply@github.com"
] |
Yashasvini18.noreply@github.com
|
3f1fa779ebc4ad9bbf4068ec2d29801aa363b470
|
5061a6089dbf1997e2d25ffc47d1dac9761228a0
|
/bellamira/views.py
|
e26206ef766461ebfc5d62f816d97bf238347a6a
|
[] |
no_license
|
DmitriDergalov/Bellamira
|
49ccc903940ac7de82996e069d0032cafff19b8f
|
2b271ded6c5abdc8b9f39efac7682a4c20a5e5b8
|
refs/heads/master
| 2021-08-19T14:45:57.561253
| 2017-11-26T18:10:41
| 2017-11-26T18:10:41
| 112,104,831
| 0
| 1
| null | null | null | null |
UTF-8
|
Python
| false
| false
| 275
|
py
|
from django.shortcuts import render
from django.views import generic
from .models import Suit
class SuitsView(generic.ListView):
template_name = 'bellamira/suits.html'
context_object_name = 'suits_list'
def get_queryset(self):
return Suit.objects.all()
|
[
"dergalov.mitya2011@yandex.ru"
] |
dergalov.mitya2011@yandex.ru
|
74a8b60c7bd5c37aa81e3159a4faeb777038fa95
|
d099ed0f4f34387d80a10490e5d621cbe8944320
|
/tests/decoder_test1.py
|
095efc613dc505ca5cca76fd24ec39c60e198d41
|
[] |
no_license
|
unfallible/jrpg_engine_experiment
|
c8c3da285cec68e4101b022c591325f7764a6efa
|
7709fb197aada8e991220d2174fac457d314d887
|
refs/heads/master
| 2023-08-28T02:47:53.186911
| 2021-10-13T03:21:07
| 2021-10-13T03:21:07
| 411,520,928
| 0
| 0
| null | null | null | null |
UTF-8
|
Python
| false
| false
| 1,281
|
py
|
from pprint import PrettyPrinter
from JrpgBattle.Attack import VanillaAttack, AttackType
from JrpgBattle.Character import CharacterTemplate
from JrpgBattle.JsonProcessing.JrpgEncoder import decode_jrpg_data, encode_jrpg_data
from fractions import Fraction
import json
def main():
pp = PrettyPrinter(indent=2)
atk1 = VanillaAttack('chocolate', damage=3)
atk2 = VanillaAttack('strawberry', damage=5, stamina_point_cost=150)
atk1_json = json.dumps(atk1, default=encode_jrpg_data)
atk2_json = json.dumps(atk2, default=encode_jrpg_data)
chocolate = json.loads(atk1_json, object_hook=decode_jrpg_data)
strawberry = json.loads(atk2_json, object_hook=decode_jrpg_data)
dude = CharacterTemplate(name='tyler', max_hp=10, offensive_type_affinities={}, defensive_type_affinities={},
attack_list={chocolate, strawberry}, parry_effectiveness=Fraction(3, 4))
encoded_char = json.dumps(dude, default=encode_jrpg_data, indent=2)
print('encoded:')
print(encoded_char)
print('decoded:')
reencoded_char = json.dumps(json.loads(encoded_char, object_hook=decode_jrpg_data), default=encode_jrpg_data, indent=2)
print(reencoded_char)
assert encoded_char == reencoded_char
if __name__ == '__main__':
main()
|
[
"ryn292@mocs.utc.edu"
] |
ryn292@mocs.utc.edu
|
b74265a2d8ce076cbf5577d14a301cccf86b39ff
|
0c25cc9150add4645d83cad35989ff9c8858c58e
|
/image_processing/classify/ex_svm.py
|
2582c596858bfc54be183cecad8d66c959242cc2
|
[] |
no_license
|
SelinaJing/python_proj
|
be0e7e6c1f5da233a87d764df8f5bb5ec40cf44e
|
8c0e4df00a877464a64548e8d3a655ee2c0879f3
|
refs/heads/master
| 2020-09-09T06:59:09.188166
| 2014-05-25T22:18:23
| 2014-05-25T22:18:23
| null | 0
| 0
| null | null | null | null |
UTF-8
|
Python
| false
| false
| 1,153
|
py
|
# -*- coding: utf-8 -*-
"""
Spyder Editor
This temporary script file is located here:
D:\bao\WinPython-64bit-2.7.5.3\settings\.spyder2\.temp.py
"""
import pickle
from svmutil import *
import imtools
import numpy as np
# load 2D example points using Pickle
with open('points_normal.pkl','r') as f:
class_1 = pickle.load(f)
class_2 = pickle.load(f)
labels = pickle.load(f)
#convert to lists for libsvm
class_1 = map(list,class_1)
class_2 = map(list,class_2)
labels = list(labels)
samples = class_1+class_2
#create SVM
prob = svm_problem(labels,samples)
param = svm_parameter('-t 2')
# train SVM on data
m = svm_train(prob,param)
res = svm_predict(labels,samples,m)
with open('points_normal_test.pkl','r') as f:
class_1 = pickle.load(f)
class_2 = pickle.load(f)
labels = pickle.load(f)
#convert to lists for libsvm
class_1 = map(list,class_1)
class_2 = map(list,class_2)
#definefunctionforplotting
def predict(x,y,model=m):
return np.array(svm_predict([0]*len(x),zip(x,y),model)[0])
#plottheclassificationboundary
imtools.plot_2D_boundary([-6,6,-6,6],[np.array(class_1),np.array(class_2)],predict,[-1,1])
show()
|
[
"baoruihan@gmail.com"
] |
baoruihan@gmail.com
|
88b507bb6359e5f3d65c82163e8547cfab8d6b21
|
2c35db3ea04538d4d08a2d62aa4a7dfd295a2173
|
/code_review_plugins/js.py
|
159e9d7f76e3219dd90de4e317229a836fc36933
|
[] |
no_license
|
HugoDelval/pseudo-automated-code-review
|
2ecff25063594cfef6af20aeeaad3724d847599e
|
b04fd7fa0be87a2d88f1609d0167aa7ca1ca55b4
|
refs/heads/master
| 2021-01-12T15:07:06.222743
| 2018-10-01T14:22:05
| 2018-10-01T14:22:05
| 71,704,830
| 1
| 1
| null | 2018-10-01T13:32:09
| 2016-10-23T13:32:14
|
Python
|
UTF-8
|
Python
| false
| false
| 1,433
|
py
|
from code_review import CodeReview
class Js(CodeReview):
def __init__(self, directory_path):
super().__init__(directory_path)
self.exclude += []
self.language = "JS"
def launch_code_review(self):
print("Launching JS review for : " + self.directory_path)
"""
Assuming nodeJS :
- new run_cmd( ...
- exec( ...
- spawn( ...
"""
output_run_cmd = self.launch_command(
'egrep "run_cmd[ |\(]+" ' + self.directory_path + ' -R 2> /dev/null |' +
'grep -v ' + '| grep -v '.join(self.exclude) + " | grep -x '^.\{3,300\}'"
)
output_exec = self.launch_command(
'egrep "exec[ |\(]+" ' + self.directory_path + ' -R 2> /dev/null |' +
'grep -v ' + '| grep -v '.join(self.exclude) + " | grep -x '^.\{3,300\}'"
)
output_spawn = self.launch_command(
'egrep "spawn[ |\(]+" ' + self.directory_path + ' -R 2> /dev/null |' +
'grep -v ' + '| grep -v '.join(self.exclude) + " | grep -x '^.\{3,300\}'"
)
self.audit_results.update({
"outputs": {
"output_run_cmd": output_run_cmd,
"output_exec": output_exec,
"output_spawn": output_spawn,
},
"nb_outputs": len(output_spawn) + len(output_run_cmd) + len(output_exec)
})
return self.audit_results
|
[
"hugodelval@gmail.com"
] |
hugodelval@gmail.com
|
e80c6918df2886ea94e0f9f355d21492dedf1d0e
|
2b3e08faaa4edb548ef9bd386247f35f50a06766
|
/djangoProject7/migrations/0001_initial.py
|
1d639b122724c8bee5e28d6bbede30f6f3f745c0
|
[] |
no_license
|
aliciawill/pythonproject
|
e495d3a9e5d65768a69ea6ac01ff559e7112dd75
|
f6d166c366522f0e3c5c74fdd11ca7a7b5489ee1
|
refs/heads/master
| 2023-07-08T04:39:59.597131
| 2021-08-22T05:15:08
| 2021-08-22T05:15:08
| 360,455,729
| 0
| 0
| null | null | null | null |
UTF-8
|
Python
| false
| false
| 629
|
py
|
# Generated by Django 3.2.6 on 2021-08-22 04:19
from django.db import migrations, models
class Migration(migrations.Migration):
initial = True
dependencies = [
]
operations = [
migrations.CreateModel(
name='Test',
fields=[
('id', models.BigAutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')),
('name', models.CharField(max_length=200)),
('tel', models.CharField(max_length=200)),
('addr', models.CharField(max_length=200)),
],
),
]
|
[
"noreply@github.com"
] |
aliciawill.noreply@github.com
|
3ad52e8c095b3ad50975940c78e30707311ab01e
|
32ba9f1c35ae916d33b121daeeea8e1910a447d7
|
/utils/tag.py
|
99083b27f2ee3987fa2f8e96f2665c4f0533c66f
|
[
"MIT"
] |
permissive
|
rituparna/glTools
|
8b02fa2751e1b997f7a202c7df8a3dd3d3032722
|
c512a96c20ba7a4ee93a123690b626bb408a8fcd
|
refs/heads/master
| 2020-03-19T19:23:47.684580
| 2018-06-10T23:53:58
| 2018-06-10T23:53:58
| 136,853,456
| 0
| 0
| null | 2018-06-10T23:46:54
| 2018-06-10T23:46:53
| null |
UTF-8
|
Python
| false
| false
| 2,893
|
py
|
import maya.cmds as mc
import glTools.tools.namingConvention
class Tag( object ):
def __init__(self):
'''
'''
self.nameTagAttr = 'nameTag'
def addNameTag(self,control,tag):
'''
Set the name tag value for the specified control
'''
# Check control
if not mc.objExists(control): raise Exception('Object '+control+' does not exist!')
# Add Tag attribute
if mc.objExists(control+'.'+self.nameTagAttr):
mc.addAttr(control,ln=self.nameTagAttr,dt='string')
mc.setAttr(control+'.'+self.nameTagAttr,tag,type='string')
def getNameTag(self,control):
'''
Return the name tag value of the specified control
'''
# Check control
if not mc.objExists(control): raise Exception('Object '+control+' does not exist!')
# Check tag attribute
if not mc.objExists(control+'.'+self.nameTagAttr): raise Exception('Object '+control+' does not have a "'+self.nameTagAttr+'" attribute!')
# Return tag string value
return mc.getAttr(control+'.'+self.nameTagAttr)
def guessNameTag(self,control,side=True,part=True,optSide=True,subPart=True,node=False):
'''
Return a best guess name tag based on a controls current name.
Uses name element comparison to our naming convention module.
'''
tag = ''
# Get naming convention dictionaries
nameConvention = glTools.tools.namingConvention.NamingConvention()
sideDict = dict((value, key) for key, value in nameConvention.side.iteritems())
partDict = dict((value, key) for key, value in nameConvention.part.iteritems())
subPartDict = dict((value, key) for key, value in nameConvention.subPart.iteritems())
nodeDict = dict((value, key) for key, value in nameConvention.node.iteritems())
# Get name elements
controlElem = control.split(nameConvention.delineator)
controlElemCnt = len(controlElem)
controlElemInd = 0
# Check number of elements
if controlElemCnt < 3: print 'Warning: Name pattern does not match naming convention'
# Get side
if side and sideDict.has_key(controlElem[controlElemInd]):
if controlElem[controlElemInd] != nameConvention.side['center']:
tag += sideDict[controlElem[controlElemInd]].capitalize()
controlElemInd += 1
else: return
# Get part
if part and partDict.has_key(controlElem[controlElemInd][0:-2]):
tag += partDict[controlElem[controlElemInd][0:-2]].capitalize()
controlElemInd += 1
else: return
# Get optional side
if optSide and sideDict.has_key(controlElem[controlElemInd][0:-2]):
tag += sideDict[controlElem[controlElemInd][0:-2]].capitalize()
controlElemInd += 1
# Get sub-part
if subPart and subPartDict.has_key(controlElem[controlElemInd][0:-2]):
tag += subPartDict[controlElem[controlElemInd][0:-2]].capitalize()
controlElemInd += 1
# Get type
if node and nodeDict.has_key(controlElem[controlElemInd]):
tag += nodeDict[controlElem[controlElemInd]].capitalize()
return tag
|
[
"grant@bungnoid.com"
] |
grant@bungnoid.com
|
f2fb7b797424d3ba6ad396d5b0583a42bb65355b
|
8f3c35b663f554d62d259ccbbd0c972b22dbb382
|
/one_cycle.py
|
5eef32e001aa5794757799e8de06188822047d52
|
[
"MIT"
] |
permissive
|
WenjinSun/kaggle-google-quest
|
29f3875b224dfa4f9748480afbf9c6a12998dcff
|
69f84bb2c4ef83b300b4c7f8378c43f7dd84f7c6
|
refs/heads/master
| 2021-01-09T12:50:03.775453
| 2020-02-13T12:55:02
| 2020-02-13T12:55:02
| null | 0
| 0
| null | null | null | null |
UTF-8
|
Python
| false
| false
| 2,244
|
py
|
import math
import numpy as np
import matplotlib.pyplot as plt
from torch.optim.lr_scheduler import LambdaLR
from utils.torch import set_optimizer_mom
def cosine_annealing(it, n_iter, start_val, end_val):
cos_inner = math.pi * (it % n_iter) / n_iter
return ((start_val - end_val) * (math.cos(cos_inner) + 1) / 2) + end_val
def cosine_annealing_range(n_iter, start_val, end_val):
return [cosine_annealing(i, n_iter, start_val, end_val)
for i in range(n_iter)]
class OneCycleLR(LambdaLR):
def __init__(self, optimizer, lr_div_factor=25, warmup_frac=0.3,
mom_range=(0.95, 0.85), n_epochs=10, n_batches=None,
start_epoch=0):
n_batches = 1 if n_batches is None else n_batches
self.n_epochs, self.n_iter = n_epochs, (n_epochs * n_batches) + 1
self.start_it = -1 if start_epoch==0 else start_epoch * n_batches
self._build_schedules(lr_div_factor, mom_range, warmup_frac)
super().__init__(optimizer, self.lr_lambda, last_epoch=self.start_it)
def _build_schedules(self, lr_div_factor, mom_range, warmup_frac):
n_warmup = int(self.n_iter * warmup_frac)
n_decay = self.n_iter - n_warmup
self.lrs = cosine_annealing_range(n_warmup, 1/lr_div_factor, 1)
self.lrs += cosine_annealing_range(n_decay, 1, 1/lr_div_factor)
self.lr_lambda = lambda i: self.lrs[i]
self.moms = cosine_annealing_range(n_warmup, *mom_range)
self.moms += cosine_annealing_range(n_decay, *mom_range[::-1])
self.mom_lambda = lambda i: self.moms[i]
def get_mom(self):
return self.mom_lambda(self.last_epoch)
def step(self, epoch=None):
super().step(epoch)
set_optimizer_mom(self.optimizer, self.get_mom())
def plot_schedules(self):
x = np.linspace(0, self.n_epochs, self.n_iter)
_, ax = plt.subplots(1, 2, figsize=(15, 4))
ax[0].set_title('LR Schedule')
ax[0].set_ylabel('lr')
ax[0].set_xlabel('epoch')
ax[0].plot(x, self.lrs)
ax[1].set_title('Momentum Schedule')
ax[1].set_ylabel('momentum')
ax[1].set_xlabel('epoch')
ax[1].plot(x, self.moms)
|
[
"48721401+robinniesert@users.noreply.github.com"
] |
48721401+robinniesert@users.noreply.github.com
|
8085959284cded2def530eb0f5a6b7c2e6f21c18
|
68cc98398c537d783a0e30dab1d09d81f0f15f01
|
/grammar/description_parser.py
|
8949960f8cc9b7dec6557b7c2b5a2290471c481d
|
[] |
no_license
|
Ulitochka/medicine
|
9bef0ed66dae1dd6bff2c821667a20cbc36fd9c3
|
49e55e73db0868592b33a715201e8a5f5ea6926d
|
refs/heads/master
| 2020-04-16T11:34:19.674655
| 2019-01-18T16:14:53
| 2019-01-18T16:14:53
| 165,541,659
| 1
| 0
| null | null | null | null |
UTF-8
|
Python
| false
| false
| 5,357
|
py
|
from tools.utils import Utils
class DescriptionParser:
"""
Класс реализующий rule-based подход для выявления симпотомов в описаниях состояний, составленных самими пациентами.
"""
def __init__(self, *, config_data):
self.config_data = config_data
self.utils = Utils()
self.max_ngrams_size = self.config_data.get("max_ngrams_size")
def preprocessed(self, description):
"""
Метод осуществляющий препроцессинг данных:
исправление ошибок
токенизация
лемматизация
:param description:
:return:
"""
correction = self.utils.spell_checking(description)
tokens = self.utils.tokenization(description)
if correction:
tokens = [correction.get(token, token) for token in tokens]
morpho_info = [self.utils.get_pos(tokens[index], index) for index in range(len(tokens))]
return morpho_info
def parse(self, description, symptoms_block, ngrams, pattern_type='keywords'):
"""
Метод принимающий на вход сгенерированные из описания нграммы, список ключевых слов для определения симптома,
разделенных на несколько категорий: object - объект, feel - ощущения, place - место на теле человека, operators
- операторы, в нашем случае это будут различные служебные части речи (предлоги, союзы).
Суть метода в нахождении пересечений различных слов из нграммы со словами из категорий симптома.
В конце у нас получается бинарный вектор для каждой нграммы. Если все компоненты этого вектора (то есть пересе-
чения с определенными ключевыми леммами из категорий) совпадают с переменной kw_status (так как у какого то
симптома может не быть ключевых слов для какой-либо категории), то мы можем отнести симптом к описанию.
Таким образом, каждый набор symptoms_block представляет собой правило для нахождения определенного типа симптома.
Различные значения аргумента pattern_type.
Существует два вида паттернов: keywords - ключевые слова разных категорий; delimiters - так как в разметке
описаний много противоречивых моментов, то они помогают решать некоторые противоречивые случаи, когда по одним
и тем же словам можно отнести разные симпотомы. По сути это keywords, они используются следующим образом: так
как правила применяются независимо, хотя интересно было бы сделать иерархию, то симптом присваивается только
в следующем случае: если есть совпадение с keywords и нет совпадений с delimiters, то симптом присваивается;
:param description:
:param symptoms_block:
:param ngrams:
:param pattern_type:
:return:
"""
key_word_ngrams = []
kw_status = {
"object": True if symptoms_block[pattern_type]['object'] else False,
"feel": True if symptoms_block[pattern_type]['feel'] else False,
"place": True if symptoms_block[pattern_type]['place'] else False,
"operators": True if symptoms_block[pattern_type]['operators'] else False
}
if [s for s in kw_status if kw_status[s]]:
if not ngrams:
ngrams = [self.utils.ngrams(description, n) for n in range(2, self.max_ngrams_size)]
n_grams = [
{"ngrams": [t['normal_form'] for t in ngrams],
"object": False,
"feel": False,
"place": False,
"operators": False
} for ngram_variant in ngrams for ngrams in ngram_variant]
for collocation in n_grams:
for kw in symptoms_block[pattern_type]:
for word in symptoms_block[pattern_type][kw]:
if word in collocation['ngrams']:
collocation[kw] = True
checking_status = [collocation[kw] == kw_status[kw] for kw in kw_status]
if all(checking_status):
key_word_ngrams.append(collocation['ngrams'])
return key_word_ngrams
|
[
"m.domrachev.scientist@gmail.com"
] |
m.domrachev.scientist@gmail.com
|
bdb8ce5f430ef2154c672fd3a39dab1855522a76
|
72dd2e55d8c52390940548587c032201f3976a8d
|
/src/train_test.py
|
dd6468ba58f797591279d999bb0d06ff027c9e69
|
[] |
no_license
|
phber/dd2404_bioinformatics
|
f2143266092663829b326ba6bfcb7326e3a4c05f
|
da39e5c8987e92ee8971b3fbfb1215399b0d5d98
|
refs/heads/master
| 2021-09-02T20:26:13.374377
| 2018-01-03T21:47:57
| 2018-01-03T21:47:57
| 115,660,308
| 0
| 0
| null | null | null | null |
UTF-8
|
Python
| false
| false
| 6,203
|
py
|
import pandas as pd
from sklearn.model_selection import KFold
import numpy as np
from Bio.HMM import MarkovModel, Trainer
from sklearn.metrics import confusion_matrix
from sklearn.utils import shuffle
from sklearn.metrics import matthews_corrcoef
from read_write import seq_alphabet, state_alphabet_pos, state_alphabet_neg
from read_write import load_dir, load_signalp, fill_train_df, store_result
from Bio.Seq import Seq
"""Train a HMM using ML-estimation"""
def fit_model(train_df, positive = True):
training_seqs = []
if positive:
builder = MarkovModel.MarkovModelBuilder(state_alphabet_pos, seq_alphabet)
builder.allow_transition('n', 'n')
builder.allow_transition('n', 'h')
builder.allow_transition('h', 'h')
builder.allow_transition('h', 'c')
builder.allow_transition('c', 'c')
builder.allow_transition('c', 'C')
builder.allow_transition('C', 'O')
builder.allow_transition('O', 'O')
builder.allow_transition('C', 'o')
builder.allow_transition('o', 'o')
builder.allow_transition('o', 'M')
builder.allow_transition('M', 'M')
builder.allow_transition('M', 'i')
builder.allow_transition('i', 'i')
builder.set_random_probabilities()
builder.set_initial_probabilities({'n' : 1})
else:
builder = MarkovModel.MarkovModelBuilder(state_alphabet_neg, seq_alphabet)
builder.allow_all_transitions()
builder.destroy_transition('O', 'i')
builder.destroy_transition('O', 'M')
builder.destroy_transition('O', 'o')
builder.set_initial_probabilities({'o' : 0.1, 'i' : 0.4, 'O': 0.5})
training_seqs = []
for i, row in train_df.iterrows():
ann = row['ann']
ann.alphabet = state_alphabet_pos if positive else state_alphabet_neg
train_seq = Trainer.TrainingSequence(row['seq'], ann)
training_seqs.append(train_seq)
model = builder.get_markov_model()
trainer = Trainer.KnownStateTrainer(model)
return trainer.train(training_seqs)
"""Prediction scores"""
def perf_measure(y_true, y_pred):
tn, fp, fn, tp = confusion_matrix(y_true, y_pred).ravel()
rec = 1.0*tp/(tp+fn)
prec = 1.0*tp/(tp+fp)
f = 2*prec*rec/(prec+rec)
mcc = matthews_corrcoef(y_true, y_pred)
return (prec, rec, f, mcc)
"""Computes test scores for a df with known labels"""
def hmm_test(pos_model, neg_model, test_df):
predictions = []
for i, row in test_df.iterrows():
seq = str(row['seq'])
seq = seq.replace('U', '')
pos_seq, pos_score = pos_model.viterbi(seq, state_alphabet_pos)
neg_sec, neg_score = neg_model.viterbi(seq, state_alphabet_neg)
if abs(pos_score - neg_score) < 2:
predictions.append(0)
elif pos_score > neg_score:
predictions.append(1)
else:
predictions.append(0)
print np.sum(predictions)
return perf_measure(list(test_df['label']), predictions)
"""Helper function for splitting dataframe by binary group"""
def split_df(df, group='label'):
#Split training into negative and positive
gb = df.groupby(group)
groups = [gb.get_group(x) for x in gb.groups]
train_df_neg = groups[0]
train_df_pos = groups[1]
return train_df_neg, train_df_pos
"""Get cross validation scores for training data"""
def cross_validate(df, write = True, validations = 5):
if len(df) == 0:
raise ValueError('No training data was found.')
kf = KFold(n_splits = validations, shuffle = True, random_state = 13)
scores_all = []
scores_non_tm = []
scores_tm = []
for train_index, test_index in kf.split(df):
print 'Running k-fold...'
# Split into train and test
train_df = df.iloc[train_index]
test_df = df.iloc[test_index]
#Split training into negative and positive
train_df_neg, train_df_pos = split_df(train_df, 'label')
#Train models
pos_model = fit_model(train_df_pos, True)
neg_model = fit_model(train_df_neg, False)
#Run tests
test_df_non_tm, test_df_tm = split_df(test_df, 'tm')
scores_all.append(hmm_test(pos_model, neg_model, test_df))
scores_tm.append(hmm_test(pos_model, neg_model, test_df_tm))
scores_non_tm.append(hmm_test(pos_model, neg_model, test_df_non_tm))
if write:
store_result(scores_all, scores_tm, scores_non_tm, 'CROSS VALIDATION', validations)
else:
print 'ALL TESTDATA', scores_all
"""Stats for proteom df"""
def stats(df):
test_df_non_tm, test_df_tm = split_df(df, 'tm')
print 'Non-TM count, TM count:'
print len(test_df_non_tm), len(test_df_tm)
print 'Negative TM count, Positive TM count:'
t1, t2 = split_df(test_df_tm)
print len(t1), len(t2)
print 'Negative Non-TM count, Positive Non-TM count:'
t1, t2 = split_df(test_df_non_tm)
print len(t1), len(t2)
"""Test a model HMM on a proteom with signalP resulsts"""
def test_proteom(train_df, proteom, stopidx, startidx = 1, write = True):
#Load proteom
prot_df = load_signalp(proteom, stopidx, startidx)
if proteom != 'bacillus':
test_df_non_tm, test_df_tm = split_df(prot_df, 'tm')
stats(prot_df)
else:
print len(prot_df), np.sum(prot_df['label'])
scores_all = []
scores_non_tm = []
scores_tm = []
train_df_neg, train_df_pos = split_df(train_df)
pos_model = fit_model(train_df_pos, True)
neg_model = fit_model(train_df_neg, False)
scores_all.append(hmm_test(pos_model, neg_model, prot_df))
if proteom != 'bacillus':
scores_tm.append(hmm_test(pos_model, neg_model, test_df_tm))
scores_non_tm.append(hmm_test(pos_model, neg_model, test_df_non_tm))
if write:
print scores_all
store_result(scores_all, scores_tm, scores_non_tm, 'PROTEOM of ' + proteom, 1)
else:
print 'ALL TESTDATA', np.mean(scores_all)
print 'TM TESTDATA', np.mean(scores_tm)
print 'NON TM TESTDATA', np.mean(scores_non_tm)
df = fill_train_df()
#test_proteom(df, 'bacillus', stopidx = 3, write = True)
#cross_validate(df, validations = 2, write = True)
|
[
"philip@Philips-MacBook.local"
] |
philip@Philips-MacBook.local
|
eb74add9cc0461ba97666dc459e71a6df598d5c1
|
3f7bc519f83609c7aac1ca24615710019f35e20a
|
/pomodoro.py
|
a18ad7133fbabc883a36accbfbf26a447f080c79
|
[
"MIT"
] |
permissive
|
blipml/bliptimer
|
2077f18cb0db7209a27a5fb62aeb7fd746037961
|
f0ec6b1de4ac797b4f3ee98240719851a23b9f81
|
refs/heads/main
| 2023-03-29T12:52:08.841306
| 2021-04-12T06:50:47
| 2021-04-12T06:50:47
| 356,160,883
| 0
| 0
| null | null | null | null |
UTF-8
|
Python
| false
| false
| 2,666
|
py
|
#!/usr/bin/python3
# -*- coding: utf-8 -*-
from datetime import datetime as dt, timedelta
from time import time, sleep
import sys
import asyncio
SLEEP_TIME=1
class TimeParseException(Exception):
pass
class Timer(object):
def __init__(
self, target=0, sleep_time=SLEEP_TIME):
self.target = Timer.parse_time(sys.argv[1])
self.elapsed = 0
print(self.target)
self.sleep_time = sleep_time
@staticmethod
def parse_time(string):
mul = [60*60, 60, 1][::-1]
components = string.split(':')[::-1]
if len(components) > len(mul):
raise TimeParseException()
elif len(components) != len(mul):
mul = mul[:len(components)-len(mul)]
_sum = 0
for m, c in zip(mul, components):
_sum += m*int(c)
return _sum
def start_timer(self):
last = now = time()
while(self.elapsed < self.target):
self.elapsed = now-last
sleep(self.sleep_time)
now = time()
self.tick_callback()
self.end_callback()
def tick_callback(self):
pass
def end_callback(self):
pass
class AsyncTimer(Timer):
# implement coroutines for callbacks?
def __init__(self, *args, **kwargs):
super(AsyncTimer, self).__init__(*args, **kwargs)
self.task = None # ???
class PreemptAsyncTimer(AsyncTimer):
# do we want to be able to somehow pre-empt timers?
def __init__(self, *args, **kwargs):
super(PreemptAsyncTimer, self).__init__(*args, **kwargs)
def cancel(self):
# maybe we don't want this here :)
# self.task.cancel()
pass
class Pomodowo(Timer):
def __init__(self, *args, **kwargs):
super(Pomodowo, self).__init__(*args, **kwargs)
def tick_callback(self):
"""
we can access tick level information here
"""
outfile = './discordtimer'
header = 'pomod🍅w🍅'
output = "{}\n\r◇{}◇".format(header, str(
timedelta(seconds=round(
self.target-self.elapsed))))
with open(outfile, 'w') as f:
f.write(output)
print(output)
class WaterTimer(Timer):
def __init__(self,*args, **kwargs):
super(WaterTimer, self).__init__(*args, **kwargs)
def end_callback(self):
# control OBS scene?
pass
if __name__ == '__main__':
if len(sys.argv) < 2:
print("Pass me a target for the timer format [HH]:[mm]:ss")
exit()
t = Pomodowo(sys.argv[1])
t.start_timer()
|
[
"ocean.kuzuri@gmail.com"
] |
ocean.kuzuri@gmail.com
|
fafb61cff6256896c850cd738ec16f4c6209143c
|
279ed6bd82f604f40fc7d13a189b99b9f6c70c8b
|
/subtask1.py
|
4d8ddcfba87cc5b42c8101e575ba336ba945a7b0
|
[] |
no_license
|
AtarioGitHub/Sub-task-8.1-and-8.2
|
bbf8044d5bfc7f2b1c0c99a743948247a8304151
|
27430455d049f2bb04b6cb8f45f23ad7869100e6
|
refs/heads/main
| 2023-09-03T21:37:37.945018
| 2021-11-14T22:19:14
| 2021-11-14T22:19:14
| 428,054,172
| 0
| 0
| null | null | null | null |
UTF-8
|
Python
| false
| false
| 339
|
py
|
import math
n = input('Enter your number which we will call n = ')
print('n = ',n)
DecimalNumberofN = float(n) # We use float so that we can also use a decimal value
TheSquareRootofThatNumber = math.sqrt(DecimalNumberofN)
xyz = int(TheSquareRootofThatNumber) # Let xyz be an integer value
q=pow(xyz,2)
print('q = ', q)
|
[
"noreply@github.com"
] |
AtarioGitHub.noreply@github.com
|
7ece9319493f54ae20dca0009b5fd429c73c203f
|
6fdda8566a04114d506a09237f6f2d69a5ff9435
|
/ai/Codes/Project/first/pp.py
|
91dbc9f1d63440ab44f01705dc85d3c4d1dcdf84
|
[] |
no_license
|
vgswn/AI
|
d0cff1ae37b02b9359682e35bef5528be9775575
|
0c6d8b1918f53eda400552c440d24192ace0b284
|
refs/heads/master
| 2021-03-16T07:54:23.506696
| 2018-01-22T09:07:15
| 2018-01-22T09:07:15
| 107,943,151
| 1
| 0
| null | null | null | null |
UTF-8
|
Python
| false
| false
| 2,881
|
py
|
import random
import queue as q
def RULE_MATCH(matrix, x, y, count, n, m, visited_cell):
d = q.PriorityQueue()
if x-1 >=0 and visited_cell[x-1][y]==0:
if matrix[x-1][y]=='Dirty':
d.put([1+(-5),'UP'])
else :
d.put([1,'UP'])
if x+1< n and visited_cell[x+1][y]==0 :
if matrix[x+1][y]=='Dirty':
d.put([2+(-5),'DOWN'])
else :
d.put([2,'DOWN'])
if y-1 >=0 and visited_cell[x][y-1]==0 :
if matrix[x][y-1]=='Dirty':
d.put([3+(-5),'LEFT'])
else :
d.put([3,'LEFT'])
if y+1 < m and visited_cell[x][y+1]==0:
if matrix[x][y+1]=='Dirty':
d.put([4+(-5),'RIGHT'])
else :
d.put([4,'RIGHT'])
if d.empty() :
flag = 0
while flag == 0:
r = random.randrange(4)
if r == 0 and x - 1 >= 0:
flag = 1
return 'UP'
elif r == 1 and x + 1 <= n:
flag = 1
return 'DOWN'
elif r == 2 and y - 1 >= 0:
flag = 1
return 'LEFT'
elif r == 3 and y + 1 <= m:
flag = 1
return 'RIGHT'
x=d.get()
return x[1]
def SIMPLE_REFLEX_AGENT(matrix, x, y, count, n, m, visited_cell):
if matrix[x][y] == 'Dirty':
matrix[x][y] = 'Clean'
print('Sucked at ', x, y)
return RULE_MATCH(matrix, x, y, count, n, m, visited_cell)
def vacuum_cleaner(matrix, x, y, count, n, m, visited_cell):
if count == (m + 1) * (n + 1):
return
if visited_cell[x][y] == 0:
count = count + 1
visited_cell[x][y] = 1
action = SIMPLE_REFLEX_AGENT(matrix, x, y, count, n, m, visited_cell)
if action == 'UP':
print('going up from ', x, y, ' to ', x - 1, y)
vacuum_cleaner(matrix, x - 1, y, count, n, m, visited_cell)
elif action == 'DOWN':
print('going down from ', x, y, ' to ', x + 1, y)
vacuum_cleaner(matrix, x + 1, y, count, n, m, visited_cell)
elif action == 'LEFT':
print('going left from ', x, y, ' to ', x, y - 1)
vacuum_cleaner(matrix, x, y - 1, count, n, m, visited_cell)
elif action == 'RIGHT':
print('going right from ', x, y, ' to ', x, y + 1)
vacuum_cleaner(matrix, x, y + 1, count, n, m, visited_cell)
n, m = input().split(' ')
n, m = (int(n), int(m))
matrix = []
visited_cell = []
for i in range(n):
matrix.append([])
visited_cell.append([])
for j in range(m):
x = random.randrange(2)
visited_cell[i].append(0)
if x == 0:
matrix[i].append('Clean')
else:
matrix[i].append('Dirty')
print(matrix)
x = random.randrange(n)
y = random.randrange(m)
count = 0
vacuum_cleaner(matrix, x, y, count, n - 1, m - 1, visited_cell)
print(visited_cell)
|
[
"iit2015038@iiita.ac.in"
] |
iit2015038@iiita.ac.in
|
fe5e0fe2bf92a1ef8907a67d02628bc6953df529
|
c180719c80c8078e689d1add2e1ae0ab62095d67
|
/website/settings.py
|
44f8b0bb47975a133c01d60909fb24be87e73608
|
[] |
no_license
|
SeekerHub/Django_webapp
|
2ba1cf05e93f229cdd773eb9db870edf2389e805
|
342537921cb53e6f071aadf62f65b16d784a56f3
|
refs/heads/master
| 2020-06-05T20:01:28.880496
| 2019-09-27T14:48:35
| 2019-09-27T14:48:35
| null | 0
| 0
| null | null | null | null |
UTF-8
|
Python
| false
| false
| 3,639
|
py
|
"""
Django settings for website project.
Generated by 'django-admin startproject' using Django 2.1.5.
For more information on this file, see
https://docs.djangoproject.com/en/2.1/topics/settings/
For the full list of settings and their values, see
https://docs.djangoproject.com/en/2.1/ref/settings/
"""
import os
# Build paths inside the project like this: os.path.join(BASE_DIR, ...)
BASE_DIR = os.path.dirname(os.path.dirname(os.path.abspath(__file__)))
# Quick-start development settings - unsuitable for production
# See https://docs.djangoproject.com/en/2.1/howto/deployment/checklist/
# SECURITY WARNING: keep the secret key used in production secret!
SECRET_KEY = 'f#kdl=s37%p=e-wex_tim2viw^qb_9=o%%1guy0^qk3-2!-=8g'
# SECURITY WARNING: don't run with debug turned on in production!
DEBUG = True
ALLOWED_HOSTS = ['127.0.0.1', '.pythonanywhere.com']
# Application definition
INSTALLED_APPS = [
'users.apps.UsersConfig',
'crispy_forms', #The new app created by us
'blog.apps.BlogConfig', #The new app created by us
'django.contrib.admin',
'django.contrib.auth',
'django.contrib.contenttypes',
'django.contrib.sessions',
'django.contrib.messages',
'django.contrib.staticfiles',
]
MIDDLEWARE = [
'django.middleware.security.SecurityMiddleware',
'django.contrib.sessions.middleware.SessionMiddleware',
'django.middleware.common.CommonMiddleware',
'django.middleware.csrf.CsrfViewMiddleware',
'django.contrib.auth.middleware.AuthenticationMiddleware',
'django.contrib.messages.middleware.MessageMiddleware',
'django.middleware.clickjacking.XFrameOptionsMiddleware',
]
ROOT_URLCONF = 'website.urls'
TEMPLATES = [
{
'BACKEND': 'django.template.backends.django.DjangoTemplates',
'DIRS': [],
'APP_DIRS': True,
'OPTIONS': {
'context_processors': [
'django.template.context_processors.debug',
'django.template.context_processors.request',
'django.contrib.auth.context_processors.auth',
'django.contrib.messages.context_processors.messages',
],
},
},
]
WSGI_APPLICATION = 'website.wsgi.application'
# Database
# https://docs.djangoproject.com/en/2.1/ref/settings/#databases
DATABASES = {
'default': {
'ENGINE': 'django.db.backends.sqlite3',
'NAME': os.path.join(BASE_DIR, 'db.sqlite3'),
}
}
# Password validation
# https://docs.djangoproject.com/en/2.1/ref/settings/#auth-password-validators
AUTH_PASSWORD_VALIDATORS = [
{
'NAME': 'django.contrib.auth.password_validation.UserAttributeSimilarityValidator',
},
{
'NAME': 'django.contrib.auth.password_validation.MinimumLengthValidator',
},
{
'NAME': 'django.contrib.auth.password_validation.CommonPasswordValidator',
},
{
'NAME': 'django.contrib.auth.password_validation.NumericPasswordValidator',
},
]
# Internationalization
# https://docs.djangoproject.com/en/2.1/topics/i18n/
LANGUAGE_CODE = 'en-us'
TIME_ZONE = 'UTC'
USE_I18N = True
USE_L10N = True
USE_TZ = True
# Static files (CSS, JavaScript, Images)
# https://docs.djangoproject.com/en/2.1/howto/static-files/
STATIC_URL = '/static/'
MEDIA_ROOT = os.path.join(BASE_DIR, 'media') # where our uploaded files will be located in the file system it will create a media folder into which it will save the foles
MEDIA_URL = '/media/' #how we are going to acces in image
CRISPY_TEMPLATE_PACK = 'bootstrap4' #important to add when you add any kind of template
LOGIN_REDIRECT_URL = 'blog-home'
LOGIN_URL = 'login'
|
[
"bhaveshgandhi1999@gmail.com"
] |
bhaveshgandhi1999@gmail.com
|
38cb295c9d71a9429c4d97b4d12140241d2a185c
|
3efc463fb20d9f7532bd0da732f23f37ad425a54
|
/main.py
|
a43f499159f67c42b9990ea8ce9b90f694d445f8
|
[] |
no_license
|
workfunction/demosacing_py
|
0d44488a555e932fe7f8159d64571c7cdab1b141
|
b345c2f4e0db953ee581666989f8dbbe8f7579ba
|
refs/heads/master
| 2023-06-27T20:09:31.158870
| 2020-02-03T05:58:22
| 2020-02-03T05:58:22
| 245,328,695
| 0
| 0
| null | null | null | null |
UTF-8
|
Python
| false
| false
| 1,687
|
py
|
from mepi import MEPI
from mepi_delta import MEPR
from mosaic import Mosaic
from MGBI5 import MGBI_5
import numpy as np
from PIL import Image
from matplotlib import pyplot as plt
import math
import tensorflow as tf
import sys
import os
def read_img(img):
return tf.convert_to_tensor(img, dtype=np.uint8)
@tf.function
def do_psnr(tf_img1, tf_img2):
return tf.image.psnr(tf_img1, tf_img2, max_val=255)
def main():
if len(sys.argv) != 3:
print("No input file!")
return -1
path = sys.argv[1]
if not os.path.isfile(path):
print("File '" + path + "' does not exist!")
return -1
oim = Image.open(path)
oimage = np.array(oim, dtype=np.uint8)
im = oim.resize((int(oim.width/2), int(oim.height/2)), Image.BICUBIC)
image = np.array(im, dtype=np.uint8)
mos = Mosaic(image, im.width, im.height)
mimage = mos.Algorithm()
epi = MEPR(mimage, 2)
out = epi.Algorithm()
#plt.imshow(out, cmap='gray', vmin=0, vmax=255)
#mos = MGBI_5(out)
#out = mos.Algorithm()
#out[:3, :, :] = oimage[:3, :, :]
#out[:, :3, :] = oimage[:, :3, :]
#out[-3:, :, :] = oimage[-3:, :, :]
#out[:, -3:, :] = oimage[:, -3:, :]
plt.imshow(out[...,1], cmap='gray', vmin=0, vmax=255)
#plt.imshow(out)
plt.show()
im = Image.fromarray(out)
filename, ext = os.path.splitext(path)
im.save("result/" + os.path.basename(filename) + "_2x" + ext)
p = do_psnr(read_img(oimage), read_img(out))
print(p)
#im.save("result/kodim01.png")
f = open("demofile2.txt", "a")
f.write(os.path.basename(filename) + ": " + str(p) + "\n")
f.close()
if __name__ == "__main__":
main()
|
[
"jason840507@gmail.com"
] |
jason840507@gmail.com
|
0185767e3d02ccdf330cda6d1f5e806dad9a8b00
|
aa0ab3eaee3a04eb39f1819cb411ce9fa2062c14
|
/scripts/driver_messaging/list_difference.py
|
1a1e13d62c6c560bce2fa535225742ac83bce60e
|
[] |
no_license
|
gilkra/tweet_proofer
|
30fd99dd11306e805526044a155a9c34dffc0713
|
04ead63aeb2cb8f0e2a92cc39a731ba926d9b617
|
refs/heads/master
| 2021-01-12T06:19:58.967602
| 2016-12-25T21:11:27
| 2016-12-25T21:11:27
| 77,342,680
| 0
| 0
| null | null | null | null |
UTF-8
|
Python
| false
| false
| 310
|
py
|
import pandas as pd
import numpy as np
uuids = pd.read_csv('calc_didnt_drive.csv')
all_id = uuids['All']
not_paid = uuids['Not_Paid']
paid = uuids['Paid']
new_list = list(set(all_id)-set(not_paid)-set(paid))
new_list = pd.DataFrame(new_list)
new_list.to_csv('didnt_drive.csv', index = False, header=False)
|
[
"kazimirovg@gmail.com"
] |
kazimirovg@gmail.com
|
74578cccd156b76ee3c9c58c8d8df9ba7a5fea79
|
d5e4c3f0d1ae6591d403804fc452a330a62d365d
|
/utils/analysis.py
|
be7fc4ea5d6a59545811ff213818a9c960112f00
|
[
"MIT"
] |
permissive
|
mtasende/Machine-Learning-Nanodegree-Capstone
|
e5ffb435b2c7a19e1e416c98d122f080a2a3ad0c
|
69c5ca499fa52122d51131a0607bc199ec054cfd
|
refs/heads/master
| 2021-05-15T16:12:20.252910
| 2017-10-22T23:42:42
| 2017-10-22T23:42:42
| 107,441,206
| 1
| 0
| null | null | null | null |
UTF-8
|
Python
| false
| false
| 5,844
|
py
|
import pandas as pd
import matplotlib.pyplot as plt
import numpy as np
CLOSE_COL_NAME = 'Close'
def compute_portfolio_stats(portfolioValue, rfr=0.0, sf=252.0):
cumRet = (portfolioValue.iloc[-1]/portfolioValue.iloc[0]) - 1
dailyRets = get_daily(portfolioValue)
# Statistics below don't take into account the first day value
averageDReturn = dailyRets[1:].mean()
sdDReturn = dailyRets[1:].std()
# Sharpe ratio calculations
if rfr==0.0:
daily_rfr = 0.0
else:
daily_rfr = ((1.0+rfr)**(1.0/sf)) -1
sharpeRatio = np.sqrt(sf)*(dailyRets[1:] - daily_rfr).mean()/(dailyRets[1:]-daily_rfr).std()
return cumRet, averageDReturn, sdDReturn, sharpeRatio
def assess_portfolio(start_date,
end_date,
symbols,
allocations,
initial_capital=1000,
risk_free_rate=0.0,
sampling_frequency=252.0,
data=None,
gen_plot=False,
verbose=False):
"""
This function returns some statistics about a portfolio.
:param start_date: Starting Date; Type = dt.datetime
:param end_date: Ending Date; Type = dt.datetime
:param symbols: A list of ticker symbols; Type: list of strings
:param allocations: A list with the fraction of allocations to each symbol; Type: list of float
:param initial_capital: Starting Value of the portfolio; Type: float/int
:param risk_free_rate: Free rate of interest; Type: float
:param sampling_frequency: Sampling frequency per year (252.0 = daily)
:param data: A dataframe with the data of the s&p500 stocks
:param gen_plot: if True create a plot
:param verbose: if True, print the output data
:returns cumulative_ret: Cumulative Return; Type: float
:returns average_daily_ret: Average Daily Return. If 'sf' is different from 252,
then it is the average in the sampling period instead of "daily"; Type: float
:returns sd_daily_ret: Standard deviation of daily return; Type: float
:returns sharpe_ratio: Sharpe Ratio; Type: float
:returns end_value: End value of portfolio; Type: float/int
"""
adj_close = data.xs(CLOSE_COL_NAME, level='feature').loc[start_date:end_date,symbols]
adj_close /= adj_close.iloc[0] # Normalize to the first day
norm_value = adj_close.dot(allocations) # Get the normalized total value
portfolio_value = pd.DataFrame(norm_value * initial_capital)
# Compute statistics from the total portfolio value
cumulative_ret, \
average_daily_ret, \
sd_daily_ret, \
sharpe_ratio = compute_portfolio_stats(portfolio_value, risk_free_rate, sampling_frequency)
end_value = portfolio_value.iloc[-1]
if gen_plot:
adj_close_SPY = data.xs(CLOSE_COL_NAME, level='feature').loc[start_date:end_date,'SPY']
adj_close_SPY /= adj_close_SPY.iloc[0]
ax = adj_close_SPY.plot(color='g', label='SPY')
ax.plot(norm_value, color='b', label='Portfolio')
plt.title('Daily portfolio value and SPY')
plt.xlabel('Date')
plt.ylabel('Normalized price')
plt.legend(loc='upper left')
plt.show()
if verbose:
print('Start Date: ' + str(start_date))
print('End Date: ' + str(end_date))
print('Symbols: ' + str(symbols))
print('Allocations ' + str(allocations))
print('Sharpe Ratio: %.15f' % sharpe_ratio)
print('Volatility (stdev of daily returns): %.15f' % sd_daily_ret)
print('Average Daily Return: %.15f' % average_daily_ret)
print('Cumulative Return: %.15f' % cumulative_ret)
return float(cumulative_ret), float(average_daily_ret), float(sd_daily_ret), float(sharpe_ratio), float(end_value)
def value_eval(value_df,
risk_free_rate=0.0,
sampling_frequency=252.0,
verbose=False,
graph=False,
data_df=None):
""" This function takes a value of portfolio series, returns some statistics, and shows some plots"""
cumulative_ret = (value_df.iloc[-1]/value_df.iloc[0]) - 1
daily_rets = get_daily(value_df)
# Statistics below don't take into account the first day value
average_daily_ret = daily_rets[1:].mean()
sd_daily_ret = daily_rets[1:].std()
# Sharpe ratio calculations
if risk_free_rate == 0.0:
daily_rfr = 0.0
else:
daily_rfr = ((1.0 + risk_free_rate)**(1.0/sampling_frequency)) -1
sharpe_ratio = np.sqrt(sampling_frequency)*(daily_rets[1:] - daily_rfr).mean()/(daily_rets[1:]-daily_rfr).std()
if verbose:
print('sharpeRatio = %f' % sharpe_ratio)
print('cumRet = %f' % cumulative_ret)
print('sdDReturn = %f' % sd_daily_ret)
print('averageDReturn = %f' % average_daily_ret)
print('Final Value: %f' % value_df.iloc[-1])
if graph:
if data_df is not None:
value_df = value_df \
.join(data_df.xs('Close', level='feature').loc[:, 'SPY'], how='left')
value_df = value_df / value_df.iloc[0]
value_df.plot()
return sharpe_ratio.values[0], \
cumulative_ret.values[0], \
average_daily_ret.values[0], \
sd_daily_ret.values[0], \
value_df.iloc[-1].values[0]
# Returns the daily change of a variable
def get_daily(data):
daily = data.copy()
daily[1:] = (data[1:] / data[:-1].values)-1
daily.ix[0,:] = np.zeros(len(data.columns))
return daily
# Printing some basic data about a ticker's attribute data
def basic_data(ticker,attr_data):
print('Ticker name: '+ticker)
print(' Mean: %f'% attr_data[ticker].mean())
print(' Std: %f'% attr_data[ticker].std())
print(' Kurtosis: %f'% attr_data[ticker].kurtosis())
print('')
|
[
"mtasendebracco@antel.com.uy"
] |
mtasendebracco@antel.com.uy
|
aee404230ad4f577d5f611ed575b11337a8eb407
|
3ba478d3d52fba786c5421f59693fb094ad4364c
|
/week10/spiders/spiders/items.py
|
e08eb780639382d2863faee497f4ddd751171508
|
[] |
no_license
|
stingh711/Python001-class01
|
c5cfb91e3dcad4994cdf37251f06034e91748e50
|
51a320ada66b70ade4427bb352e28b84794d8503
|
refs/heads/master
| 2022-12-06T20:56:11.103688
| 2020-09-06T02:07:48
| 2020-09-06T02:07:48
| 273,141,919
| 0
| 0
| null | 2020-06-18T04:33:35
| 2020-06-18T04:33:34
| null |
UTF-8
|
Python
| false
| false
| 282
|
py
|
# -*- coding: utf-8 -*-
# Define here the models for your scraped items
#
# See documentation in:
# https://docs.scrapy.org/en/latest/topics/items.html
import scrapy
class GoodsItem(scrapy.Item):
name = scrapy.Field()
link = scrapy.Field()
comments = scrapy.Field()
|
[
"lihong.hu@siemens.com"
] |
lihong.hu@siemens.com
|
741cdf5634fbc3cb461a36e5e68dc1ac744f172f
|
68e0b34c7c749331d0d660c1b5721ea82617205f
|
/fortune_server_data_management/daily_call_for_vincent.py
|
096999fd8e80bb808e5915a94925932a5f4bfffb
|
[] |
no_license
|
TheIllusion/TheIllusionsLibraries
|
ffeebe3a6ddfd635b69d07c54058d3e8bf74c337
|
0adf5a0a5d4c2e4de41faac6fcc75700104c2b53
|
refs/heads/master
| 2022-05-08T18:16:00.883192
| 2022-04-14T01:08:26
| 2022-04-14T01:08:26
| 63,536,801
| 1
| 0
| null | null | null | null |
UTF-8
|
Python
| false
| false
| 305
|
py
|
from subprocess import call
import datetime
import time
#Get the date of yesterday
timestamp = datetime.datetime.fromtimestamp(time.time() - 24*60*60).strftime('%Y_%m_%d')
print 'date = ', timestamp
call(["/home/nhnent/H1/users/rklee/Data/server_data/server_hand_data/get_data_from_svc.sh", timestamp])
|
[
"gunboy83@gmail.com"
] |
gunboy83@gmail.com
|
78f015df88e7a0ff2590c9378ea411fd33cbebbf
|
83ff6674ad1fc4ac1d9523216149d5dae82f8bbd
|
/trainer/migrations/0009_trainer_image.py
|
50deec7ba1e9345665e98a31775ed109aae6c5ba
|
[] |
no_license
|
4802852/Act-Agora
|
8a75c4622b0e306a514096b0a4aaa3e3360ec26e
|
7603444e32cebd6d5ae2d3a6f8f3a349373120d2
|
refs/heads/master
| 2023-04-28T19:49:18.498810
| 2021-05-10T13:44:03
| 2021-05-10T13:44:03
| 351,899,676
| 0
| 0
| null | null | null | null |
UTF-8
|
Python
| false
| false
| 463
|
py
|
# Generated by Django 3.1.7 on 2021-02-28 20:04
from django.db import migrations
import imagekit.models.fields
class Migration(migrations.Migration):
dependencies = [
('trainer', '0008_auto_20210228_1822'),
]
operations = [
migrations.AddField(
model_name='trainer',
name='image',
field=imagekit.models.fields.ProcessedImageField(blank=True, null=True, upload_to='profiles/'),
),
]
|
[
"4802852@gmail.com"
] |
4802852@gmail.com
|
d73f79ef6d4d5880eb4ee7c8c1972b1d25ee4cc0
|
594d059453f90a0b6d43009a630d162bf0ffaa46
|
/company_blog_flask/company_blog/models.py
|
a9b231749d0877d3a577c589061d55afb14b0315
|
[] |
no_license
|
sarfrajobject/Company_blog_post_flask
|
a707ba444b0205b07c69484d509d3a9c578c3100
|
3c0996178ad3eee35dde416bd1542c8b1560fe55
|
refs/heads/master
| 2020-09-08T23:22:43.423191
| 2019-11-12T17:43:01
| 2019-11-12T17:43:01
| 221,273,916
| 0
| 0
| null | null | null | null |
UTF-8
|
Python
| false
| false
| 1,869
|
py
|
# models.py
from company_blog import db, login_manager
from werkzeug.security import generate_password_hash, check_password_hash
from flask_login import UserMixin
from datetime import datetime
@login_manager.user_loader
def load_user(user_id):
return User.query.get(user_id)
class User(db.Model,UserMixin):
__tablename__ = 'users'
id = db.Column(db.Integer, primary_key=True)
profile_image = db.Column(db.String(64),
nullable=False,
default='default_profile.png')
email = db.Column(db.String(),unique=True,
index=True)
username = db.Column(db.String(64),unique=True,
index=True)
password_hash = db.Column(db.String(128))
posts = db.relationship('BlogPost', backref='author', lazy=True)
def __init__(self,email,username,password):
self.email = email
self.username = username
self.password_hash = generate_password_hash(password)
def check_password(self,password):
return check_password_hash(self.password_hash,password)
def __repr__(self):
return 'Username {self.username}'
#########################################################
class BlogPost(db.Model,UserMixin):
users = db.relationship(User)
id = db.Column(db.Integer, primary_key=True)
user_id = db.Column(db.Integer,db.ForeignKey('users.id'),
nullable=False)
date = db.Column(db.DateTime, nullable=False,
default=datetime.utcnow)
title = db.Column(db.String(140), nullable=False)
text = db.Column(db.Text, nullable=False)
def __init__(self,title,text,user_id):
self.title = title
self.text = text
self.user_id = user_id
def __repr__(self):
return 'POST_ID: {self.id} : DATE {self.date} : TITLE {self.title}'
|
[
"sarfraj.object@gmail.com"
] |
sarfraj.object@gmail.com
|
569012129e554d8d356b12f5bb351e3f331f6621
|
d93529bee8c03312d376f7e0cc3d42aa538fc0d4
|
/travello/views.py
|
0ec5884194d205d78ff13645de4e84dbd03ba7f9
|
[] |
no_license
|
AL-NOMA/django-projects
|
962a9caaf4b75e090731a279f890c8b4bb0e7361
|
230e701b2890ac87995ec5b0dacfbf52badd788a
|
refs/heads/master
| 2023-06-03T11:51:20.187755
| 2021-06-16T12:10:00
| 2021-06-16T12:10:00
| 375,130,119
| 0
| 0
| null | null | null | null |
UTF-8
|
Python
| false
| false
| 443
|
py
|
from django.shortcuts import render, get_object_or_404
from .models import Place
# Create your views here.
def index(request):
places = Place.objects.all()
context = {'places': places}
return render(request, 'travello/homepage.html', context)
def destination(request, place_id):
place = get_object_or_404(Place, pk=place_id)
context = {'place': place}
return render(request, 'travello/destination.html', context)
|
[
"alexisnomanyo@gmail.com"
] |
alexisnomanyo@gmail.com
|
b332b14aa6f86a6013de1b87f5c0920504478890
|
4be56098894a95da5964622fc4102b69e4530ab6
|
/题库/870.矩阵中的幻方.py
|
00fdbac5b41f5978d6246140777b7c174e01c850
|
[] |
no_license
|
ACENDER/LeetCode
|
7c7c7ecc8d0cc52215272f47ec34638637fae7ac
|
3383b09ab1246651b1d7b56ab426a456f56a4ece
|
refs/heads/master
| 2023-03-13T19:19:07.084141
| 2021-03-15T09:29:21
| 2021-03-15T09:29:21
| 299,332,864
| 0
| 0
| null | null | null | null |
UTF-8
|
Python
| false
| false
| 85
|
py
|
# !/usr/bin/env python3
# -*- coding: utf-8 -*-
# @File : 870.矩阵中的幻方.py
|
[
"1641429327@qq.com"
] |
1641429327@qq.com
|
b2b58656fafea5b3228a70403034f42c9a4cc4c7
|
7bb3a08e8628780d2bdc86ed84690100acbb8771
|
/sphere.py
|
82d278a2b81dc077f503c61ffe86f0b09e6b1064
|
[] |
no_license
|
javamak/python_raytrace_tut
|
2e84b2a757f429925b7af3f8542f96a1ffc857a9
|
4db3e0650d2cd414d8f74b1b25f18657a8e6f320
|
refs/heads/main
| 2023-04-26T15:56:22.477737
| 2021-05-03T14:36:26
| 2021-05-03T14:36:26
| 362,014,469
| 0
| 0
| null | null | null | null |
UTF-8
|
Python
| false
| false
| 700
|
py
|
from math import sqrt
class Sphere:
def __init__(self, center, radius, material):
self.center = center
self.radius = radius
self.material = material
def intesects(self, ray):
sphere_to_ray = ray.origin - self.center
#a = 1
b = 2 * ray.direction.dot_product(sphere_to_ray)
c = sphere_to_ray.dot_product(sphere_to_ray) - self.radius * self.radius
discriminant = b * b - 4 * c
if discriminant >= 0:
dist = (-b - sqrt(discriminant)) / 2
if dist > 0:
return dist
return None
def normal(self, surface_point):
return (surface_point - self.center).normalize()
|
[
"javamak@gmail.com"
] |
javamak@gmail.com
|
1bd246b511bcb25535f008e43dec5d7633a97690
|
2c112f781016f2022dc7ff1c616b1f57185fe8f8
|
/tests/conftest.py
|
34a2935ba0d6d69229c0b0455e16b60a8fcb1f85
|
[] |
no_license
|
dominicgs/Website
|
c15312a5b081b42db880b99df6811c8c04777824
|
fc3587daacff20ec3ab590df121c9f693f09a8ce
|
refs/heads/master
| 2020-03-21T16:49:07.492309
| 2018-06-26T21:41:50
| 2018-06-26T21:41:50
| 64,015,414
| 0
| 1
| null | 2016-07-23T12:33:16
| 2016-07-23T12:33:16
| null |
UTF-8
|
Python
| false
| false
| 2,265
|
py
|
" PyTest Config. This contains global-level pytest fixtures. "
import os
import os.path
import pytest
import shutil
from models.user import User
from main import create_app, db as db_obj, Mail
from utils import CreateBankAccounts, CreateTickets
@pytest.fixture(scope="module")
def app():
""" Fixture to provide an instance of the app.
This will also create a Flask app_context and tear it down.
This fixture is scoped to the module level to avoid too much
Postgres teardown/creation activity which is slow.
"""
if 'SETTINGS_FILE' not in os.environ:
root = os.path.join(os.path.dirname(os.path.abspath(__file__)), '..')
os.environ['SETTINGS_FILE'] = os.path.join(root, 'config', 'test.cfg')
tmpdir = os.environ.get('TMPDIR', '/tmp')
prometheus_dir = os.path.join(tmpdir, 'emf_test_prometheus')
os.environ['prometheus_multiproc_dir'] = prometheus_dir
if os.path.exists(prometheus_dir):
shutil.rmtree(prometheus_dir)
if not os.path.exists(prometheus_dir):
os.mkdir(prometheus_dir)
app = create_app()
with app.app_context():
try:
db_obj.session.close()
except:
pass
db_obj.drop_all()
db_obj.create_all()
CreateBankAccounts().run()
CreateTickets().run()
yield app
db_obj.session.close()
db_obj.drop_all()
@pytest.fixture
def client(app):
" Yield a test HTTP client for the app "
yield app.test_client()
@pytest.fixture
def db(app):
" Yield the DB object "
yield db_obj
@pytest.fixture
def request_context(app):
" Run the test in an app request context "
with app.test_request_context('/') as c:
yield c
@pytest.fixture
def user(db):
" Yield a test user. Note that this user will be identical across all tests in a module. "
email = 'test_user@test.invalid'
user = User.query.filter(User.email == email).one_or_none()
if not user:
user = User(email, 'Test User')
db.session.add(user)
db.session.commit()
yield user
@pytest.fixture
def outbox(app):
" Capture mail and yield the outbox. "
mail_obj = Mail()
with mail_obj.record_messages() as outbox:
yield outbox
|
[
"russ@garrett.co.uk"
] |
russ@garrett.co.uk
|
efaf672a163e3201ff01ff94b0c00aa263fe5a66
|
7896f35d698e4896be25eb879ac546ee9b46b69c
|
/main.py
|
a608e718e883d43b5c7a9cd940e98ac5e6dc179f
|
[] |
no_license
|
SajedNahian/GraphicsWork05
|
51297fe44e5072765af82ab7d6e837b40698089d
|
b2099746416a9470622bb496a788f4c9873e2467
|
refs/heads/master
| 2020-04-30T12:30:41.408764
| 2019-03-20T22:48:19
| 2019-03-20T22:48:19
| 176,828,348
| 0
| 0
| null | null | null | null |
UTF-8
|
Python
| false
| false
| 489
|
py
|
from display import *
from draw import *
from scriptParser import *
from matrix import *
import math
screen = new_screen()
color = [ 0, 255, 0 ]
edges = []
transform = new_matrix()
# print_matrix( make_translate(3, 4, 5) )
# print
# print_matrix( make_scale(3, 4, 5) )
# print
# print_matrix( make_rotX(math.pi/4) )
# print
# print_matrix( make_rotY(math.pi/4) )
# print
# print_matrix( make_rotZ(math.pi/4) )
parse_file( 'script', edges, transform, screen, color )
|
[
"noreply@github.com"
] |
SajedNahian.noreply@github.com
|
9c7ade8ee0b8e0940518d647d405439a209f33bc
|
3af9515966c7d1c074a0923ebab8e60787c8eafa
|
/oop5.py
|
9a39b6de0ef278a5b85ebb57262dfa4900b143e0
|
[] |
no_license
|
suzaana749/basicpython
|
8dbc4bcf903655143ed616f751a128d0ba8f1b67
|
035d67fe08c4bb5a4bb864c0b6da2d79bd21ff78
|
refs/heads/master
| 2020-04-13T11:15:40.938938
| 2018-12-31T11:45:49
| 2018-12-31T11:45:49
| 163,168,865
| 0
| 0
| null | null | null | null |
UTF-8
|
Python
| false
| false
| 635
|
py
|
from __future__ import print_function
import math
import random
class circle:
def calCircum(self):
return 2*math.pi*self.radius
def calDiameter(self):
return self.radius *2
def calArea(self):
return math.pi * (self.radius ** 2)
def __init__(self): # constructor function called
self.radius - random.uniform(1.1,10.5)
circles = []
for i in range(0,10):
c = circle()
#c.radius = random.uniform(1.1,10.5)
circles.append(c)
for c in circles:
print("radius",round(c.radius,2),"circum",round(c.calCircum(),2),"diameter",round(c.calDiameter(),2),"area",round(c.calArea(),2))
|
[
"mgrsuzaana@gmail.com"
] |
mgrsuzaana@gmail.com
|
a7a68c783777b2cd112ffd422219f4293e1aee15
|
ac329340f6376bdcf76a893e65641304faf85d86
|
/Des_Test.py
|
f8a0a13e2181550e3c1f2db84df317262b894fe8
|
[] |
no_license
|
JimShu716/Modified_MNIST_Classification
|
b9df06ba23eadc40f5958cea4024d6830e1785e3
|
ed17261d8579e1808a8839c7b883e7b38f9a3c5f
|
refs/heads/master
| 2020-08-30T14:00:34.464475
| 2020-02-11T00:22:38
| 2020-02-11T00:22:38
| 218,401,968
| 1
| 0
| null | null | null | null |
UTF-8
|
Python
| false
| false
| 2,291
|
py
|
import pandas as pd
import numpy as np
import torch
import torch.nn as nn
import torch.utils.data as Data
import torch.nn.functional as F
import torchvision.models as model
import matplotlib.pyplot as plt
# Hyperparameters
EPOCH = 6
BATCH_SIZE = 64
LR = 0.0001
SPLIT_RATIO = 1
# 文件路径自己去改~~~~~~~~~~~~~~~~~~~~~~~~~~~~ :) :) :) :< :< :<
train_img = pd.read_pickle('/Users/tylerliu/GitHub/Proj3_source/train_max_x')
test_img = pd.read_pickle('/Users/tylerliu/GitHub/Proj3_source/test_max_x')
train_out = pd.read_csv('./data/train_max_y.csv').to_numpy().astype('int32')[:, -1]
train_img = torch.Tensor(train_img)
train_out = torch.tensor(train_out, dtype=torch.int64)
test_img = torch.Tensor(test_img)
# 自己改测试用的大小 50000改成别的->?
x = torch.unsqueeze(train_img, dim=1)[:SPLIT_RATIO*50000]/255.
y = train_out[:50000]
# mini-sample for testing
x_t = torch.unsqueeze(train_img, dim=1)[:300]/255.
y_t = train_out[:300]
my_dataset = Data.TensorDataset(x, y)
train_loader = Data.DataLoader(dataset=my_dataset, batch_size=BATCH_SIZE, shuffle=True)
# vgg16 = model.vgg16(pretrained=True)
des = model.DenseNet(num_init_features=32,num_classes=10)
cnn = nn.Sequential(
nn.Conv2d(1,3,5,1,2,bias=True),
des,
# nn.Conv2d(1,3,5,1,2,bias=True),
# vgg16,
# nn.Dropout(0.5),
# nn.Linear(1000,10),
# nn.Softmax()
)
# print(cnn)
# optimizer = torch.optim.Adam(cnn.parameters(), lr = LR)
# optimizer = torch.optim.Adam(cnn.parameters(), lr = 1e-4)
optimizer = torch.optim.Adam(cnn.parameters(), lr = 0.0001)
loss_fun = nn.CrossEntropyLoss()
def train():
for epoch in range(EPOCH):
for step, (x, y) in enumerate(train_loader):
output = cnn(x)
loss = loss_fun(output, y)
optimizer.zero_grad()
loss.backward()
optimizer.step()
if step % 10 == 0:
x_t_out = cnn(x_t)
y_pred = torch.max(x_t_out, dim=1)[1].data.numpy()
accuracy = float((y_pred == y_t.data.numpy()).astype(int).sum()) / float(y_t.size(0))
print('Epoch ', epoch, 'train loss: %.4f' % loss.data.numpy(), '| test accuracy: %.3f' % accuracy)
# if step % 30 == 0
# LR 做一个修正
train()
|
[
"tylerliu260@gmail.com"
] |
tylerliu260@gmail.com
|
933cecb82f03c9b88ce4e634c0b0e9c7bf32ba2a
|
1fc6d2baa958992adecd42d57e89f979972dc3c5
|
/random/py/05.level_order_bfs.py
|
89bcc6c4ae80c9181ea30c396c217251888dbe28
|
[] |
no_license
|
Ritapravo/cpp
|
b5c56d3b649b0fd2698482e960e39217d547e8fb
|
0510f41b5ff5c59126461f4c36b3d05c9b1a362e
|
refs/heads/master
| 2021-08-20T06:30:49.176592
| 2021-07-12T17:43:43
| 2021-07-12T17:43:43
| 216,426,587
| 5
| 6
| null | null | null | null |
UTF-8
|
Python
| false
| false
| 1,040
|
py
|
#level node tree using bfs
#https://www.geeksforgeeks.org/level-node-tree-source-node-using-bfs/
from collections import deque as queue
def makegraph():
n = int(input())
graph = [[] for i in range(n)]
for i in range(n-1):
x,y = map(int, input().split())
graph[x].append(y)
graph[y].append(x)
return graph
def bfs(graph, u):
n = len(graph)
vis = [False]*n
vis[u] = True
levels = []
q = queue()
q.append(u)
k = 1
q_size = 1
while(not q_size==0):
k -= 1
if(k==0):
k = q_size
levels.append(list(q))
v = q.popleft()
q_size -= 1
for i in graph[v]:
if not vis[i]:
vis[i] = True
q.append(i)
q_size += 1
return levels
if __name__ == "__main__":
#graph = [[1, 2], [0, 3, 4, 5], [0, 6], [1], [1], [1], [2, 7], [6]]
graph = [[],[2,3],[1,4,5],[1,6,7],[2,8],[2,8],[3,8],[3,8],[4,5,6,7]]
levels = bfs(graph, 1)
print(levels)
|
[
"ritopravosarker@gmail.com"
] |
ritopravosarker@gmail.com
|
1fc74891fa1324f804b07585e2b154d9b49afdf6
|
de681ebfa95a07c04fbb1280bf722847b06ee548
|
/migrations/versions/3fd0d7bc25ea_create_tag_model.py
|
c5297e733152e5145f95a89eca64b85173b984bb
|
[] |
no_license
|
Dzhoker/flask-lessons
|
156957ed29a674df474cfc6b8cdca12adae021d7
|
590e436516dbd8a3a9af4ad33aafbc854088a6aa
|
refs/heads/master
| 2023-03-18T09:44:19.431920
| 2021-02-18T02:52:07
| 2021-02-18T02:52:07
| null | 0
| 0
| null | null | null | null |
UTF-8
|
Python
| false
| false
| 761
|
py
|
"""create Tag model
Revision ID: 3fd0d7bc25ea
Revises: 1b2fd89e61b5
Create Date: 2021-01-12 07:40:03.728879
"""
from alembic import op
import sqlalchemy as sa
# revision identifiers, used by Alembic.
revision = '3fd0d7bc25ea'
down_revision = '1b2fd89e61b5'
branch_labels = None
depends_on = None
def upgrade():
# ### commands auto generated by Alembic - please adjust! ###
op.create_table('tag',
sa.Column('id', sa.Integer(), nullable=False),
sa.Column('name', sa.String(length=32), server_default='', nullable=False),
sa.PrimaryKeyConstraint('id')
)
# ### end Alembic commands ###
def downgrade():
# ### commands auto generated by Alembic - please adjust! ###
op.drop_table('tag')
# ### end Alembic commands ###
|
[
"surenkhorenyan@gmail.com"
] |
surenkhorenyan@gmail.com
|
ca408825b28434521ba7bb95bb73b324467a56b2
|
c2c45ff6987a161e945b34d3b377ce11a9c9cf3b
|
/python-flask-server/swagger_server/models/expression.py
|
cf1a5cf4c0e286dca3231a5f5dbecd00572320c2
|
[] |
no_license
|
JeSuisArina/Python-CGIHTTPServer
|
089cabad6b6dcfa77ae6fef380b235711fa89e2d
|
60853297b9337f27286916de55abfc7c39ad19a5
|
refs/heads/master
| 2020-03-17T03:31:37.175275
| 2018-06-06T07:51:16
| 2018-06-06T07:51:16
| 133,239,195
| 0
| 0
| null | null | null | null |
UTF-8
|
Python
| false
| false
| 2,294
|
py
|
# coding: utf-8
from __future__ import absolute_import
from datetime import date, datetime # noqa: F401
from typing import List, Dict # noqa: F401
from swagger_server.models.base_model_ import Model
from swagger_server.models.expression_result import ExpressionResult # noqa: F401,E501
from swagger_server import util
class Expression(Model):
"""NOTE: This class is auto generated by the swagger code generator program.
Do not edit the class manually.
"""
def __init__(self, num: List[int]=None, result: ExpressionResult=None): # noqa: E501
"""Expression - a model defined in Swagger
:param num: The num of this Expression. # noqa: E501
:type num: List[int]
:param result: The result of this Expression. # noqa: E501
:type result: ExpressionResult
"""
self.swagger_types = {
'num': List[int],
'result': ExpressionResult
}
self.attribute_map = {
'num': 'num',
'result': 'result'
}
self._num = num
self._result = result
@classmethod
def from_dict(cls, dikt) -> 'Expression':
"""Returns the dict as a model
:param dikt: A dict.
:type: dict
:return: The Expression of this Expression. # noqa: E501
:rtype: Expression
"""
return util.deserialize_model(dikt, cls)
@property
def num(self) -> List[int]:
"""Gets the num of this Expression.
:return: The num of this Expression.
:rtype: List[int]
"""
return self._num
@num.setter
def num(self, num: List[int]):
"""Sets the num of this Expression.
:param num: The num of this Expression.
:type num: List[int]
"""
self._num = num
@property
def result(self) -> ExpressionResult:
"""Gets the result of this Expression.
:return: The result of this Expression.
:rtype: ExpressionResult
"""
return self._result
@result.setter
def result(self, result: ExpressionResult):
"""Sets the result of this Expression.
:param result: The result of this Expression.
:type result: ExpressionResult
"""
self._result = result
|
[
"noreply@github.com"
] |
JeSuisArina.noreply@github.com
|
5ebf743396b7da4369c2afb4a306a720519eb571
|
46645262901ac3117e8959b9a334617825def109
|
/python/couler/couler/pyfunc.py
|
c13509fdc5e68b39bf61472a6167618c496d987b
|
[
"Apache-2.0"
] |
permissive
|
ggaaooppeenngg/sqlflow
|
e8dc3c641db6aa3ba91f201f3f46d0d4bb11b83c
|
e2f04c6482d722aa952707c19f62670305a2bc3c
|
refs/heads/develop
| 2020-09-17T03:57:43.047158
| 2019-11-25T09:59:49
| 2019-11-25T09:59:49
| 223,980,975
| 0
| 0
|
Apache-2.0
| 2019-12-19T11:08:19
| 2019-11-25T15:28:07
| null |
UTF-8
|
Python
| false
| false
| 5,408
|
py
|
# Copyright 2019 The SQLFlow Authors. All rights reserved.
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
import base64
import importlib
import inspect
import os
import re
import textwrap
def _argo_safe_name(name):
"""Some names are to be used in the Argo YAML file. For example,
the generateName and template name in
https://github.com/argoproj/argo/blob/master/examples/hello-world.yaml. As
Argo is to use the YAML as part of Kubernetes job description
YAML, these names must follow Kubernetes's convention -- no
period or underscore. This function replaces these prohibited
characters into dashes.
"""
if name is None:
return None
# '_' and '.' are not allowed
return re.sub(r"_|\.", "-", name)
def invocation_location():
"""If a function A in file B calls function C, which in turn calls
invocation_location(), the call returns information about the invocation,
in particular, the caller's name "A" and the line number where A
calls C. Return (B + line_number) as function_name if A doesn't exist,
where users directly calls C in file B.
:return: a tuple of (function_name, invocation_line)
"""
stack = inspect.stack()
if len(stack) < 4:
line_number = stack[len(stack) - 1][2]
func_name = "%s-%d" % (
_argo_safe_name(workflow_filename()),
line_number,
)
else:
func_name = _argo_safe_name(stack[2][3])
line_number = stack[3][2]
return func_name, line_number
def body(func_obj):
"""If a function A calls body(), the call returns the Python source code of
the function definition body (not including the signature) of A.
"""
if func_obj is None:
return None
code = inspect.getsource(func_obj)
# Remove function signature
code = code[code.find(":") + 1 :] # noqa: E203
# Function might be defined in some indented scope
# (e.g. in another function).
# We need to handle this and properly dedent the function source code
return textwrap.dedent(code)
def workflow_filename():
"""Return the Python file that defines the workflow.
"""
stacks = inspect.stack()
frame = inspect.stack()[len(stacks) - 1]
full_path = frame[0].f_code.co_filename
filename, _ = os.path.splitext(os.path.basename(full_path))
filename = _argo_safe_name(filename)
return filename
def input_parameter(function_name, var_pos):
"""Generate parameter name for using as template input parameter names
in Argo YAML. For example, the parameter name "message" in the
container template print-message in
https://github.com/argoproj/argo/tree/master/examples#output-parameters.
"""
return "para-%s-%s" % (function_name, var_pos)
def container_output(function_name, caller_line, output_id):
"""Generate output name from an Argo container template. For example,
"{{steps.generate-parameter.outputs.parameters.hello-param}}" used in
https://github.com/argoproj/argo/tree/master/examples#output-parameters.
"""
function_id = invocation_name(function_name, caller_line)
return "couler.%s.%s.outputs.parameters.%s" % (
function_name,
function_id,
output_id,
)
def script_output(function_name, caller_line):
"""Generate output name from an Argo script template. For example,
"{{steps.generate.outputs.result}}" in
https://github.com/argoproj/argo/tree/master/examples#scripts--results
"""
function_id = invocation_name(function_name, caller_line)
return "couler.%s.%s.outputs.result" % (function_name, function_id)
def invocation_name(function_name, caller_line):
"""Argo YAML requires that each step, which is an invocation to a
template, has a name. For example, hello1, hello2a, and hello2b
in https://github.com/argoproj/argo/tree/master/examples#steps.
However, in Python programs, there are no names for function
invocations. So we hype a name by the callee and where it is
called.
"""
return "%s-%s" % (function_name, caller_line)
def load_cluster_config():
"""Load user provided cluster specification file. For example,
config file for Sigma EU95 cluster is placed at 'couler/clusters/eu95.py'.
"""
module_file = os.getenv("couler_cluster_config")
if module_file is None:
return None
spec = importlib.util.spec_from_file_location(module_file, module_file)
module = importlib.util.module_from_spec(spec)
spec.loader.exec_module(module)
return module.cluster
def encode_base64(s):
"""
Encode a string using base64 and return a binary string.
This function is used in Secret creation.
For example, the secrets for Argo YAML:
https://github.com/argoproj/argo/blob/master/examples/README.md#secrets
"""
bencode = base64.b64encode(s.encode("utf-8"))
return str(bencode, "utf-8")
|
[
"noreply@github.com"
] |
ggaaooppeenngg.noreply@github.com
|
c96137e00332c4e6faaaafb5e3bf3bf35180ad73
|
e9caa9eeff9ec8443e492e825997f94d2a3a80d3
|
/pythonProject/HRank/max_sum_con.py
|
0890e412988dfd51dede4143d1d7daa92482f64a
|
[] |
no_license
|
saurabhpiyush1187/webapp_api_framework
|
1fe7270d660f2bd59914d449d649686b7a2a0d17
|
8c5c60720c7481d238a0cdac9c5aace356829a20
|
refs/heads/master
| 2023-03-02T05:13:46.960326
| 2021-02-13T13:33:23
| 2021-02-13T13:33:23
| 334,638,494
| 0
| 0
| null | null | null | null |
UTF-8
|
Python
| false
| false
| 716
|
py
|
# This is a sample Python script.
# Press Shift+F10 to execute it or replace it with your code.
# Press Double Shift to search everywhere for classes, files, tool windows, actions, and settings.
def print_hi(name):
# Use a breakpoint in the code line below to debug your script.
print(f'Hi, {name}') # Press Ctrl+F8 to toggle the breakpoint.
lst_num = [2,3,4,1,90,4,98]
n=3
lst_max = []
i=0
while(i+n<=len(lst_num)):
Sum = sum(lst_num[i:i+n])
lst_max.append(Sum)
i+=1
print(lst_max)
# Press the green button in the gutter to run the script.
if __name__ == '__main__':
print_hi('PyCharm')
# See PyCharm help at https://www.jetbrains.com/help/pycharm/
|
[
"saurabhpiyush@spglobal.com"
] |
saurabhpiyush@spglobal.com
|
8786f9605a9ca3d5d09b16b60f4fb0c77fcd8b81
|
e365e230ad8a889107264b39a5eda71d68172f6f
|
/archive/combine_rev_aws_transcripts.py
|
7cb4923f0956dd8ea9ec05a340c1f00c2773b29f
|
[] |
no_license
|
wkevwang/Intuition-scribe
|
b4efd0584703a76f208c1a7b099782e26bb17607
|
969973103f13bb5e89d35cfe316527db8cd59cb1
|
refs/heads/master
| 2023-06-23T20:18:39.999554
| 2021-07-28T22:23:58
| 2021-07-28T22:23:58
| 280,757,348
| 1
| 0
| null | null | null | null |
UTF-8
|
Python
| false
| false
| 3,356
|
py
|
"""
Patch up the low confidence parts of the Rev.ai transcipt using the
AWS Transcribe transcript, and return an edited version of the Rev.ai
transcript.
"""
import argparse
import os
import json
from rev_diarization import parse_transcript_elements as parse_rev_transcript_elements
def find_elements_in_time_bound(elements, start, end, margin=0.1):
"""
Find the elements in "elements" that have start and end times within
"start" and "end", with an additional margin time of "margin"
If an element in "elements" is used, mark it as used so it is not
used again.
"""
elements_in_bound = []
start_bound = start - margin
end_bound = end + margin
for e in elements:
if "ts" not in e: # Skip punctuation since it does not have timestamps
continue
if (start_bound <= e["ts"]) and (e["end_ts"] <= end_bound) and (not e.get("used", False)):
elements_in_bound.append(e)
e["used"] = True
return elements_in_bound
if __name__ == "__main__":
parser = argparse.ArgumentParser()
parser.add_argument("--audio_file", type=str, required=True) # Used for getting the original filename
parser.add_argument("--rev_transcript", type=str, required=True)
parser.add_argument("--aws_transcript", type=str, required=True)
parser.add_argument("--output_folder", type=str, required=True)
args = parser.parse_args()
with open(args.rev_transcript, 'r') as f:
rev_transcript = json.load(f)
with open(args.aws_transcript, 'r') as f:
aws_transcript = json.load(f)
# Form a single list of the elements for the transcript
# Each element has "value" and for non-punct elements, "ts" and "end_ts"
rev_transcript_elements = parse_rev_transcript_elements(rev_transcript)
aws_transcript_items = aws_transcript["results"]["items"]
aws_transcript_elements = []
for item in aws_transcript_items:
element = {
"value": item["alternatives"][0]["content"],
}
if "start_time" in item:
element["ts"] = float(item["start_time"])
element["end_ts"] = float(item["end_time"])
element["confidence"] = float(item["alternatives"][0]["confidence"])
aws_transcript_elements.append(element)
# Build list of final elements based on timestamps
# Use Rev element if confidence is high. Otherwise, use AWS Transcribe element
final_elements = []
for rev_e in rev_transcript_elements:
if (rev_e["type"] == "punct") or (rev_e["confidence"] > 0.95):
final_elements.append(rev_e)
else:
aws_elems = find_elements_in_time_bound(
aws_transcript_elements, rev_e["ts"], rev_e["end_ts"])
if len(aws_elems) > 0:
final_elements += aws_elems
else:
final_elements.append(rev_e)
filename_prefix = os.path.splitext(os.path.basename(args.audio_file))[0]
filename = filename_prefix + '_rev_aws_combined_transcript.json'
with open(os.path.join(args.output_folder, filename), 'w') as f:
# Write in Rev format (with a single speaker to meet the expected format)
json.dump({
"monologues": [{
"speaker": 0,
"elements": final_elements
}]
}, f, indent=4)
|
[
"weixiang@ualberta.ca"
] |
weixiang@ualberta.ca
|
12180bacf865e009ab34847ca2fc32c7d48d7f9b
|
531c47c15b97cbcb263ec86821d7f258c81c0aaf
|
/sdk/network/azure-mgmt-network/azure/mgmt/network/v2019_12_01/operations/_ddos_custom_policies_operations.py
|
9bdb5268477680f052fb35e51b998db264bd5970
|
[
"LicenseRef-scancode-generic-cla",
"LGPL-2.1-or-later",
"MIT"
] |
permissive
|
YijunXieMS/azure-sdk-for-python
|
be364d3b88204fd3c7d223df23756386ff7a3361
|
f779de8e53dbec033f98f976284e6d9491fd60b3
|
refs/heads/master
| 2021-07-15T18:06:28.748507
| 2020-09-04T15:48:52
| 2020-09-04T15:48:52
| 205,457,088
| 1
| 2
|
MIT
| 2020-06-16T16:38:15
| 2019-08-30T21:08:55
|
Python
|
UTF-8
|
Python
| false
| false
| 19,277
|
py
|
# coding=utf-8
# --------------------------------------------------------------------------
# Copyright (c) Microsoft Corporation. All rights reserved.
# Licensed under the MIT License. See License.txt in the project root for license information.
# Code generated by Microsoft (R) AutoRest Code Generator.
# Changes may cause incorrect behavior and will be lost if the code is regenerated.
# --------------------------------------------------------------------------
from typing import TYPE_CHECKING
import warnings
from azure.core.exceptions import HttpResponseError, ResourceExistsError, ResourceNotFoundError, map_error
from azure.core.pipeline import PipelineResponse
from azure.core.pipeline.transport import HttpRequest, HttpResponse
from azure.core.polling import LROPoller, NoPolling, PollingMethod
from azure.mgmt.core.exceptions import ARMErrorFormat
from azure.mgmt.core.polling.arm_polling import ARMPolling
from .. import models
if TYPE_CHECKING:
# pylint: disable=unused-import,ungrouped-imports
from typing import Any, Callable, Dict, Generic, Optional, TypeVar, Union
T = TypeVar('T')
ClsType = Optional[Callable[[PipelineResponse[HttpRequest, HttpResponse], T, Dict[str, Any]], Any]]
class DdosCustomPoliciesOperations(object):
"""DdosCustomPoliciesOperations operations.
You should not instantiate this class directly. Instead, you should create a Client instance that
instantiates it for you and attaches it as an attribute.
:ivar models: Alias to model classes used in this operation group.
:type models: ~azure.mgmt.network.v2019_12_01.models
:param client: Client for service requests.
:param config: Configuration of service client.
:param serializer: An object model serializer.
:param deserializer: An object model deserializer.
"""
models = models
def __init__(self, client, config, serializer, deserializer):
self._client = client
self._serialize = serializer
self._deserialize = deserializer
self._config = config
def _delete_initial(
self,
resource_group_name, # type: str
ddos_custom_policy_name, # type: str
**kwargs # type: Any
):
# type: (...) -> None
cls = kwargs.pop('cls', None) # type: ClsType[None]
error_map = {404: ResourceNotFoundError, 409: ResourceExistsError}
error_map.update(kwargs.pop('error_map', {}))
api_version = "2019-12-01"
# Construct URL
url = self._delete_initial.metadata['url'] # type: ignore
path_format_arguments = {
'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str'),
'ddosCustomPolicyName': self._serialize.url("ddos_custom_policy_name", ddos_custom_policy_name, 'str'),
'subscriptionId': self._serialize.url("self._config.subscription_id", self._config.subscription_id, 'str'),
}
url = self._client.format_url(url, **path_format_arguments)
# Construct parameters
query_parameters = {} # type: Dict[str, Any]
query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str')
# Construct headers
header_parameters = {} # type: Dict[str, Any]
# Construct and send request
request = self._client.delete(url, query_parameters, header_parameters)
pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs)
response = pipeline_response.http_response
if response.status_code not in [200, 202, 204]:
map_error(status_code=response.status_code, response=response, error_map=error_map)
raise HttpResponseError(response=response, error_format=ARMErrorFormat)
if cls:
return cls(pipeline_response, None, {})
_delete_initial.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Network/ddosCustomPolicies/{ddosCustomPolicyName}'} # type: ignore
def begin_delete(
self,
resource_group_name, # type: str
ddos_custom_policy_name, # type: str
**kwargs # type: Any
):
# type: (...) -> LROPoller
"""Deletes the specified DDoS custom policy.
:param resource_group_name: The name of the resource group.
:type resource_group_name: str
:param ddos_custom_policy_name: The name of the DDoS custom policy.
:type ddos_custom_policy_name: str
:keyword callable cls: A custom type or function that will be passed the direct response
:keyword str continuation_token: A continuation token to restart a poller from a saved state.
:keyword polling: True for ARMPolling, False for no polling, or a
polling object for personal polling strategy
:paramtype polling: bool or ~azure.core.polling.PollingMethod
:keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present.
:return: An instance of LROPoller that returns either None or the result of cls(response)
:rtype: ~azure.core.polling.LROPoller[None]
:raises ~azure.core.exceptions.HttpResponseError:
"""
polling = kwargs.pop('polling', True) # type: Union[bool, PollingMethod]
cls = kwargs.pop('cls', None) # type: ClsType[None]
lro_delay = kwargs.pop(
'polling_interval',
self._config.polling_interval
)
cont_token = kwargs.pop('continuation_token', None) # type: Optional[str]
if cont_token is None:
raw_result = self._delete_initial(
resource_group_name=resource_group_name,
ddos_custom_policy_name=ddos_custom_policy_name,
cls=lambda x,y,z: x,
**kwargs
)
kwargs.pop('error_map', None)
kwargs.pop('content_type', None)
def get_long_running_output(pipeline_response):
if cls:
return cls(pipeline_response, None, {})
if polling is True: polling_method = ARMPolling(lro_delay, lro_options={'final-state-via': 'location'}, **kwargs)
elif polling is False: polling_method = NoPolling()
else: polling_method = polling
if cont_token:
return LROPoller.from_continuation_token(
polling_method=polling_method,
continuation_token=cont_token,
client=self._client,
deserialization_callback=get_long_running_output
)
else:
return LROPoller(self._client, raw_result, get_long_running_output, polling_method)
begin_delete.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Network/ddosCustomPolicies/{ddosCustomPolicyName}'} # type: ignore
def get(
self,
resource_group_name, # type: str
ddos_custom_policy_name, # type: str
**kwargs # type: Any
):
# type: (...) -> "models.DdosCustomPolicy"
"""Gets information about the specified DDoS custom policy.
:param resource_group_name: The name of the resource group.
:type resource_group_name: str
:param ddos_custom_policy_name: The name of the DDoS custom policy.
:type ddos_custom_policy_name: str
:keyword callable cls: A custom type or function that will be passed the direct response
:return: DdosCustomPolicy, or the result of cls(response)
:rtype: ~azure.mgmt.network.v2019_12_01.models.DdosCustomPolicy
:raises: ~azure.core.exceptions.HttpResponseError
"""
cls = kwargs.pop('cls', None) # type: ClsType["models.DdosCustomPolicy"]
error_map = {404: ResourceNotFoundError, 409: ResourceExistsError}
error_map.update(kwargs.pop('error_map', {}))
api_version = "2019-12-01"
# Construct URL
url = self.get.metadata['url'] # type: ignore
path_format_arguments = {
'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str'),
'ddosCustomPolicyName': self._serialize.url("ddos_custom_policy_name", ddos_custom_policy_name, 'str'),
'subscriptionId': self._serialize.url("self._config.subscription_id", self._config.subscription_id, 'str'),
}
url = self._client.format_url(url, **path_format_arguments)
# Construct parameters
query_parameters = {} # type: Dict[str, Any]
query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str')
# Construct headers
header_parameters = {} # type: Dict[str, Any]
header_parameters['Accept'] = 'application/json'
# Construct and send request
request = self._client.get(url, query_parameters, header_parameters)
pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs)
response = pipeline_response.http_response
if response.status_code not in [200]:
map_error(status_code=response.status_code, response=response, error_map=error_map)
raise HttpResponseError(response=response, error_format=ARMErrorFormat)
deserialized = self._deserialize('DdosCustomPolicy', pipeline_response)
if cls:
return cls(pipeline_response, deserialized, {})
return deserialized
get.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Network/ddosCustomPolicies/{ddosCustomPolicyName}'} # type: ignore
def _create_or_update_initial(
self,
resource_group_name, # type: str
ddos_custom_policy_name, # type: str
parameters, # type: "models.DdosCustomPolicy"
**kwargs # type: Any
):
# type: (...) -> "models.DdosCustomPolicy"
cls = kwargs.pop('cls', None) # type: ClsType["models.DdosCustomPolicy"]
error_map = {404: ResourceNotFoundError, 409: ResourceExistsError}
error_map.update(kwargs.pop('error_map', {}))
api_version = "2019-12-01"
content_type = kwargs.pop("content_type", "application/json")
# Construct URL
url = self._create_or_update_initial.metadata['url'] # type: ignore
path_format_arguments = {
'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str'),
'ddosCustomPolicyName': self._serialize.url("ddos_custom_policy_name", ddos_custom_policy_name, 'str'),
'subscriptionId': self._serialize.url("self._config.subscription_id", self._config.subscription_id, 'str'),
}
url = self._client.format_url(url, **path_format_arguments)
# Construct parameters
query_parameters = {} # type: Dict[str, Any]
query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str')
# Construct headers
header_parameters = {} # type: Dict[str, Any]
header_parameters['Content-Type'] = self._serialize.header("content_type", content_type, 'str')
header_parameters['Accept'] = 'application/json'
# Construct and send request
body_content_kwargs = {} # type: Dict[str, Any]
body_content = self._serialize.body(parameters, 'DdosCustomPolicy')
body_content_kwargs['content'] = body_content
request = self._client.put(url, query_parameters, header_parameters, **body_content_kwargs)
pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs)
response = pipeline_response.http_response
if response.status_code not in [200, 201]:
map_error(status_code=response.status_code, response=response, error_map=error_map)
raise HttpResponseError(response=response, error_format=ARMErrorFormat)
deserialized = None
if response.status_code == 200:
deserialized = self._deserialize('DdosCustomPolicy', pipeline_response)
if response.status_code == 201:
deserialized = self._deserialize('DdosCustomPolicy', pipeline_response)
if cls:
return cls(pipeline_response, deserialized, {})
return deserialized
_create_or_update_initial.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Network/ddosCustomPolicies/{ddosCustomPolicyName}'} # type: ignore
def begin_create_or_update(
self,
resource_group_name, # type: str
ddos_custom_policy_name, # type: str
parameters, # type: "models.DdosCustomPolicy"
**kwargs # type: Any
):
# type: (...) -> LROPoller
"""Creates or updates a DDoS custom policy.
:param resource_group_name: The name of the resource group.
:type resource_group_name: str
:param ddos_custom_policy_name: The name of the DDoS custom policy.
:type ddos_custom_policy_name: str
:param parameters: Parameters supplied to the create or update operation.
:type parameters: ~azure.mgmt.network.v2019_12_01.models.DdosCustomPolicy
:keyword callable cls: A custom type or function that will be passed the direct response
:keyword str continuation_token: A continuation token to restart a poller from a saved state.
:keyword polling: True for ARMPolling, False for no polling, or a
polling object for personal polling strategy
:paramtype polling: bool or ~azure.core.polling.PollingMethod
:keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present.
:return: An instance of LROPoller that returns either DdosCustomPolicy or the result of cls(response)
:rtype: ~azure.core.polling.LROPoller[~azure.mgmt.network.v2019_12_01.models.DdosCustomPolicy]
:raises ~azure.core.exceptions.HttpResponseError:
"""
polling = kwargs.pop('polling', True) # type: Union[bool, PollingMethod]
cls = kwargs.pop('cls', None) # type: ClsType["models.DdosCustomPolicy"]
lro_delay = kwargs.pop(
'polling_interval',
self._config.polling_interval
)
cont_token = kwargs.pop('continuation_token', None) # type: Optional[str]
if cont_token is None:
raw_result = self._create_or_update_initial(
resource_group_name=resource_group_name,
ddos_custom_policy_name=ddos_custom_policy_name,
parameters=parameters,
cls=lambda x,y,z: x,
**kwargs
)
kwargs.pop('error_map', None)
kwargs.pop('content_type', None)
def get_long_running_output(pipeline_response):
deserialized = self._deserialize('DdosCustomPolicy', pipeline_response)
if cls:
return cls(pipeline_response, deserialized, {})
return deserialized
if polling is True: polling_method = ARMPolling(lro_delay, lro_options={'final-state-via': 'azure-async-operation'}, **kwargs)
elif polling is False: polling_method = NoPolling()
else: polling_method = polling
if cont_token:
return LROPoller.from_continuation_token(
polling_method=polling_method,
continuation_token=cont_token,
client=self._client,
deserialization_callback=get_long_running_output
)
else:
return LROPoller(self._client, raw_result, get_long_running_output, polling_method)
begin_create_or_update.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Network/ddosCustomPolicies/{ddosCustomPolicyName}'} # type: ignore
def update_tags(
self,
resource_group_name, # type: str
ddos_custom_policy_name, # type: str
parameters, # type: "models.TagsObject"
**kwargs # type: Any
):
# type: (...) -> "models.DdosCustomPolicy"
"""Update a DDoS custom policy tags.
:param resource_group_name: The name of the resource group.
:type resource_group_name: str
:param ddos_custom_policy_name: The name of the DDoS custom policy.
:type ddos_custom_policy_name: str
:param parameters: Parameters supplied to update DDoS custom policy resource tags.
:type parameters: ~azure.mgmt.network.v2019_12_01.models.TagsObject
:keyword callable cls: A custom type or function that will be passed the direct response
:return: DdosCustomPolicy, or the result of cls(response)
:rtype: ~azure.mgmt.network.v2019_12_01.models.DdosCustomPolicy
:raises: ~azure.core.exceptions.HttpResponseError
"""
cls = kwargs.pop('cls', None) # type: ClsType["models.DdosCustomPolicy"]
error_map = {404: ResourceNotFoundError, 409: ResourceExistsError}
error_map.update(kwargs.pop('error_map', {}))
api_version = "2019-12-01"
content_type = kwargs.pop("content_type", "application/json")
# Construct URL
url = self.update_tags.metadata['url'] # type: ignore
path_format_arguments = {
'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str'),
'ddosCustomPolicyName': self._serialize.url("ddos_custom_policy_name", ddos_custom_policy_name, 'str'),
'subscriptionId': self._serialize.url("self._config.subscription_id", self._config.subscription_id, 'str'),
}
url = self._client.format_url(url, **path_format_arguments)
# Construct parameters
query_parameters = {} # type: Dict[str, Any]
query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str')
# Construct headers
header_parameters = {} # type: Dict[str, Any]
header_parameters['Content-Type'] = self._serialize.header("content_type", content_type, 'str')
header_parameters['Accept'] = 'application/json'
# Construct and send request
body_content_kwargs = {} # type: Dict[str, Any]
body_content = self._serialize.body(parameters, 'TagsObject')
body_content_kwargs['content'] = body_content
request = self._client.patch(url, query_parameters, header_parameters, **body_content_kwargs)
pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs)
response = pipeline_response.http_response
if response.status_code not in [200]:
map_error(status_code=response.status_code, response=response, error_map=error_map)
raise HttpResponseError(response=response, error_format=ARMErrorFormat)
deserialized = self._deserialize('DdosCustomPolicy', pipeline_response)
if cls:
return cls(pipeline_response, deserialized, {})
return deserialized
update_tags.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Network/ddosCustomPolicies/{ddosCustomPolicyName}'} # type: ignore
|
[
"noreply@github.com"
] |
YijunXieMS.noreply@github.com
|
6844ce56ffa18f4d971b348110a9f410a1502c7e
|
a3c662a5eda4e269a8c81c99e229879b946a76f6
|
/.venv/lib/python3.7/site-packages/pylint/test/input/func_noerror_yield_assign_py25.py
|
f40d8d96e837e9022fc2596b23ce8733990a450c
|
[
"MIT"
] |
permissive
|
ahmadreza-smdi/ms-shop
|
0c29da82c58b243507575672bbc94fb6e8068aeb
|
65ba3f3061e2ac5c63115b08dadfe7d67f645fb6
|
refs/heads/master
| 2023-04-27T19:51:34.858182
| 2019-11-24T20:57:59
| 2019-11-24T20:57:59
| 223,616,552
| 6
| 2
|
MIT
| 2023-04-21T20:51:21
| 2019-11-23T16:09:03
|
Python
|
UTF-8
|
Python
| false
| false
| 387
|
py
|
"""http://www.logilab.org/ticket/8771"""
from __future__ import print_function
def generator():
"""yield as assignment"""
yield 45
xxxx = yield 123
print(xxxx)
def generator_fp1(seq):
"""W0631 false positive"""
for val in seq:
pass
for val in seq:
yield val
def generator_fp2():
"""E0601 false positive"""
xxxx = 12
yield xxxx
|
[
"ahmadreza.smdi@gmail.com"
] |
ahmadreza.smdi@gmail.com
|
e1f046bf1125c305df03d5d353029498f0cbe882
|
56d41bbc6b5d831ba699ad4a44f5880ba3d195c8
|
/thread_sync.py
|
bcd0c1730eb098adb4c5bb1a0e3dc4d595662b6d
|
[] |
no_license
|
pigliangliang/2018-07-05-08
|
4635e4dc1926f3f17eae7f607a0b188f6aaf9f43
|
ba95331f4b0cc0316377a5c67f86d03e8cc257b8
|
refs/heads/master
| 2020-03-22T08:51:08.209718
| 2018-07-08T07:58:06
| 2018-07-08T07:58:06
| 139,795,833
| 1
| 0
| null | null | null | null |
UTF-8
|
Python
| false
| false
| 606
|
py
|
#author_by zhuxiaoliang
#2018-07-05 下午3:39
"""
A. Semaphore(信号量)
在多线程编程中,为了防止不同的线程同时对一个公用的资源(比如全部变量)进行修改,需要进行同时访问的数量(通常是1)的限制。信号量同步基于内部计数器,每调用一次acquire(),计数器减1;每调用一次release(),计数器加1.当计数器为0时,acquire()调用被阻塞。
"""
import time
from random import random
from threading import Thread,Semaphore,enumerate
sema = Semaphore(3)
def foo(tid):
with sema:
print()
|
[
"836904717@qq.com"
] |
836904717@qq.com
|
7fd6bd8c1c7690c327cccbd16ff1cfdb292ee885
|
478cf410ccf4b1511e7d31d99aec68a6c5499792
|
/17-CS-Python/01-Intro-Python-Overview/src/15_classes.py
|
e10767a5ae1aa5e9d3d93c51dce62a9fee5720c5
|
[
"MIT"
] |
permissive
|
felipe0216/data-science-journal
|
8f4bf9d572c194044c4c738fd39f4816ea542bdc
|
1ea6208e7abdd514ffbac3f53d2279aac2d4e493
|
refs/heads/master
| 2020-07-13T18:52:19.916794
| 2019-08-23T18:34:05
| 2019-08-23T18:34:05
| 205,134,360
| 1
| 0
| null | 2019-08-29T10:02:33
| 2019-08-29T10:02:33
| null |
UTF-8
|
Python
| false
| false
| 1,467
|
py
|
# Make a class LatLon that can be passed parameters `lat` and `lon` to the
# constructor
# YOUR CODE HERE
class LatLon(object):
def __init__(self, lat, lon):
self.lat = lat
self.lon = lon
# Make a class Waypoint that can be passed parameters `name`, `lat`, and `lon` to the
# constructor. It should inherit from LatLon. Look up the `super` method.
# YOUR CODE HERE
class Waypoint(LatLon):
def __init__(self, name, lat, lon):
self.name = name
super().__init__(lat, lon)
def __str__(self):
return self.name
# Make a class Geocache that can be passed parameters `name`, `difficulty`,
# `size`, `lat`, and `lon` to the constructor. What should it inherit from?
# YOUR CODE HERE
class Geocache(Waypoint):
def __init__(self, name, difficulty, size, lat, lon):
self.difficulty = difficulty
self.size = size
super().__init__(name, lat, lon)
# Make a new waypoint and print it out: "Catacombs", 41.70505, -121.51521
# YOUR CODE HERE
waypoint = Waypoint("Catacombs", 41.70505, -121.51521)
# Without changing the following line, how can you make it print into something
# more human-readable? Hint: Look up the `object.__str__` method
print(waypoint)
# Make a new geocache "Newberry Views", diff 1.5, size 2, 44.052137, -121.41556
# YOUR CODE HERE
geocache = Geocache("Newberry Views", 1.5, 2, 44.052137, -121.41556)
# Print it--also make this print more nicely
print(geocache)
|
[
"chrisluedtke@gmail.com"
] |
chrisluedtke@gmail.com
|
688072f226f2b49b242f3cc4a00b67dc6eae16bf
|
a393ccc763ee9cb78e7c6e2acd9ecb906bc006bd
|
/tests/conftest.py
|
411e2a0fe30f78b0827f2d421eea59b048341f54
|
[] |
no_license
|
Kilo59/GoT-Char_DB16
|
d2e548886d1c698439ad0b6b948e750f6da25149
|
122f81419b28590e138dd8fb082c7a0dbb9ae00a
|
refs/heads/master
| 2021-09-08T14:38:11.765705
| 2021-09-04T16:22:42
| 2021-09-04T16:22:42
| 65,403,337
| 1
| 0
| null | 2021-09-04T16:22:43
| 2016-08-10T17:37:48
|
Python
|
UTF-8
|
Python
| false
| false
| 657
|
py
|
"""
tests.conftest.py
~~~~~~~~~~~~~~~~~
Shared testing fixtures
"""
import pathlib
import pytest
ROOT = pathlib.Path(__file__).parents[1]
# TODO: don't rely on specific database path or name
SQLITE_DB = ROOT / "database.db"
@pytest.fixture(scope="session")
def monkey_session():
"""Session scoped monkeypatch"""
with pytest.MonkeyPatch.context() as monkeypatch:
yield monkeypatch
@pytest.fixture(scope="session", autouse=True)
def session_setup_teardown():
# setup
if SQLITE_DB.exists():
SQLITE_DB.unlink() # missing_ok py 3.8
# yield
yield
# teardown
if SQLITE_DB.exists():
SQLITE_DB.unlink()
|
[
"noreply@github.com"
] |
Kilo59.noreply@github.com
|
855e3ca79856a91739da6ccf5b04275f51eadda3
|
f1d3bca01c6fd993aef2cc4d6b22f488cdc5e3b5
|
/cat_scraper.py
|
f58e21cbeaaedba8a34c1d4102e93013845bf8c8
|
[] |
no_license
|
HanKruiger/cat-scraper
|
cad1db39de6336321866a52a46f33e9b3153ee87
|
e4cef07e17e763f131a5fdc9aaec0db8ea55cc2b
|
refs/heads/main
| 2023-04-19T23:03:04.510215
| 2021-05-08T14:53:04
| 2021-05-08T14:53:04
| 365,496,083
| 0
| 0
| null | null | null | null |
UTF-8
|
Python
| false
| false
| 1,432
|
py
|
import logging as log
import requests
from bs4 import BeautifulSoup
from urllib.parse import urlsplit
class CatScraper:
def __init__(self, url):
self.url = url
split_url = urlsplit(url)
self.base_path = f'{split_url.scheme}://{split_url.netloc}'
self.cats = None
def scrape(self):
log.info('Scraping %s', self.url)
try:
response = requests.get(self.url)
if not response.ok:
log.warning('Response to %s not OK: %s', self.url, response)
return []
except requests.exceptions.ConnectionError:
log.warning('Could not connect to %s', self.url)
return []
soup = BeautifulSoup(response.content, 'html.parser')
# Select all <div> elements with a 'dieren' class, but WITHOUT the
# 'header' class.
cat_divs = soup.select('div.dieren:not(.header)')
# Convert the <div> elements to cat dictionaries.
self.cats = [self.div_to_cat(cat_div) for cat_div in cat_divs]
def get_cats(self):
if self.cats is None:
self.scrape()
return self.cats
def div_to_cat(self, cat_div):
return {
'url': self.base_path + cat_div.find('a')['href'],
'name': cat_div.find('li', class_='naam').text.split(':')[1],
'status': cat_div.find('li', class_='status').text.split(':')[1],
}
|
[
"489397-hankruiger@users.noreply.gitlab.com"
] |
489397-hankruiger@users.noreply.gitlab.com
|
09517b256fb2ff6d1963ce364542f0f10d77b7c2
|
92e290932cba0e045c86b26de13fcc902e920d83
|
/tart/python/bb/asound.py
|
884d3285fdb5317d935503c9f30cc70377038847
|
[] |
no_license
|
HorizonXP/blackberry-py
|
4f9615f2637c1e9bc81b712172c41d13ce6e1726
|
ee9823b5f1ce1efa6531005df60b246a4147c4a2
|
refs/heads/master
| 2020-05-27T13:08:56.497247
| 2013-08-30T06:56:03
| 2013-08-30T06:56:03
| 7,714,244
| 2
| 1
| null | 2013-08-29T08:57:05
| 2013-01-20T08:27:19
|
Python
|
UTF-8
|
Python
| false
| false
| 74,497
|
py
|
'''Wrappers for asound.h routines'''
import ctypes
from ctypes import (sizeof,
c_bool, c_ubyte, c_byte, c_char,
c_ushort, c_short,
c_uint, c_int,
c_ulonglong, c_longlong,
c_float, c_double,
c_char_p, c_void_p,
POINTER, Structure, Union, CFUNCTYPE)
# TODO: confirm these are correct
size_t = c_uint
ssize_t = c_int
uint8_t = c_ubyte
uint16_t = c_ushort
uint32_t = c_uint
uint64_t = c_ulonglong
int8_t = c_byte
int16_t = c_short
int32_t = c_int
int64_t = c_longlong
from ._wrap import _func, _register_funcs
class timeval(Structure):
_fields_ = [
('tv_secs', uint32_t),
('tv_usec', int32_t),
]
class snd_pcm_t(Structure):
pass
# /****************************************************************************
# * *
# * control.h *
# * Control Interface *
# * *
# ****************************************************************************/
# typedef struct snd_ctl_callbacks {
# void *private_data; /* should be used by an application */
# void (*rebuild) (void *private_data);
# void (*xswitch) (void *private_data, int cmd, int iface, snd_switch_list_item_t *item);
# void *reserved[29]; /* reserved for the future use - must be NULL!!! */
# } snd_ctl_callbacks_t;
class snd_ctl_t(Structure):
pass
class snd_mixer_t(Structure):
pass
SND_PROTOCOL_VERSION = lambda t,a,b,c: (t<<24|a<<16|b<<8|c)
SND_PROTOCOL_INCOMPATIBLE = lambda a,b: (a)!=(b)
# /*****************/
# /*****************/
# /*** SWITCH ***/
# /*****************/
# /*****************/
# #define SND_SW_TYPE_BOOLEAN 1
# #define SND_SW_TYPE_BYTE 2
# #define SND_SW_TYPE_WORD 3
# #define SND_SW_TYPE_DWORD 4
# #define SND_SW_TYPE_LIST 10
# #define SND_SW_TYPE_STRING_11 100+11
# #define SND_SW_SUBTYPE_DEC 0
# #define SND_SW_SUBTYPE_HEXA 1
class snd_switch_list_item_t(Structure):
_fields_ = [
('name', c_char * 32),
('reserved', uint8_t * 128),
]
class snd_switch_list_t(Structure):
_fields_ = [
('iface', int32_t),
('device', int32_t),
('channel', int32_t),
('switches_size', int32_t),
('switches', int32_t),
('switches_over', int32_t),
('pswitches', POINTER(snd_switch_list_item_t)),
('pzero', c_void_p), # align pointers on 64-bits --> point to NULL
('reserved', uint8_t * 128), # must be filled with zero
]
# typedef struct snd_switch
# {
# int32_t iface;
# int32_t device;
# int32_t channel;
# char name[36];
# uint32_t type;
# uint32_t subtype;
# uint32_t zero[2];
# union
# {
# uint32_t enable:1;
# struct
# {
# uint8_t data;
# uint8_t low;
# uint8_t high;
# }
# byte;
# struct
# {
# uint16_t data;
# uint16_t low;
# uint16_t high;
# }
# word;
# struct
# {
# uint32_t data;
# uint32_t low;
# uint32_t high;
# }
# dword;
# struct
# {
# uint32_t data;
# uint32_t items[30];
# uint32_t items_cnt;
# }
# list;
# struct
# {
# uint8_t selection;
# char strings[11][11];
# uint8_t strings_cnt;
# }
# string_11;
# uint8_t raw[32];
# uint8_t reserved[128]; /* must be filled with zero */
# }
# value;
# uint8_t reserved[128]; /* must be filled with zero */
# }
# snd_switch_t;
# /*****************/
# /*****************/
# /*** CONTROL ***/
# /*****************/
# /*****************/
# #define SND_CTL_VERSION SND_PROTOCOL_VERSION('C',3,0,0)
# #define SND_CTL_SW_JOYSTICK "joystick"
# #define SND_CTL_SW_JOYSTICK_ADDRESS "joystick port address"
class snd_ctl_hw_info_t(Structure):
_fields_ = [
('type', uint32_t),
('hwdepdevs', uint32_t),
('pcmdevs', uint32_t),
('mixerdevs', uint32_t),
('mididevs', uint32_t),
('timerdevs', uint32_t),
('id', c_char * 16),
('abbreviation', c_char * 16),
('name', c_char * 32),
('longname', c_char * 80),
('reserved', uint8_t * 128), # must be filled with zero
]
# #define SND_CTL_IFACE_CONTROL 100
# #define SND_CTL_IFACE_MIXER 200
# #define SND_CTL_IFACE_PCM_PLAYBACK 300
# #define SND_CTL_IFACE_PCM_CAPTURE 301
# #define SND_CTL_IFACE_RAWMIDI_OUTPUT 400
# #define SND_CTL_IFACE_RAWMIDI_INPUT 401
# #define SND_CTL_READ_REBUILD 120
# #define SND_CTL_READ_SWITCH_VALUE 121
# #define SND_CTL_READ_SWITCH_CHANGE 122
# #define SND_CTL_READ_SWITCH_ADD 123
# #define SND_CTL_READ_SWITCH_REMOVE 124
# typedef struct snd_ctl_read_s
# {
# int32_t cmd;
# uint8_t zero[4]; /* alignment -- zero fill */
# union
# {
# struct
# {
# int32_t iface;
# uint8_t zero[4]; /* alignment -- zero fill */
# snd_switch_list_item_t switem;
# uint8_t reserved[128]; /* must be filled with zero */
# } sw;
# uint8_t reserved[128]; /* must be filled with zero */
# } data;
# uint8_t reserved[128]; /* must be filled with zero */
# } snd_ctl_read_t;
# #define SND_CTL_IOCTL_DRVR_VER _IOR ('C', 0x00, int)
# #define SND_CTL_IOCTL_PVERSION _IOR ('C', 0x10, int)
# #define SND_CTL_IOCTL_HW_INFO _IOR ('C', 0x20, snd_ctl_hw_info_t)
# #define SND_CTL_IOCTL_SWITCH_LIST _IOWR('C', 0x30, snd_switch_list_t)
# #define SND_CTL_IOCTL_SWITCH_READ _IOWR('C', 0x31, snd_switch_t)
# #define SND_CTL_IOCTL_SWITCH_WRITE _IOWR('C', 0x32, snd_switch_t)
# #define SND_CTL_IOCTL_MIXER_DEVICE _IOW ('C', 0x40, int)
# #define SND_CTL_IOCTL_MIXER_INFO _IOR ('C', 0x41, snd_mixer_info_t)
# #define SND_CTL_IOCTL_MIXER_SWITCH_LIST _IOWR('C', 0x42, snd_switch_list_t)
# #define SND_CTL_IOCTL_MIXER_SWITCH_READ _IOWR('C', 0x43, snd_switch_t)
# #define SND_CTL_IOCTL_MIXER_SWITCH_WRITE _IOWR('C', 0x44, snd_switch_t)
# #define SND_CTL_IOCTL_PCM_CHANNEL _IOW ('C', 0x50, int)
# #define SND_CTL_IOCTL_PCM_CHANNEL_INFO _IOR ('C', 0x51, snd_pcm_channel_info_t)
# #define SND_CTL_IOCTL_PCM_DEVICE _IOW ('C', 0x52, int)
# #define SND_CTL_IOCTL_PCM_INFO _IOR ('C', 0x53, snd_pcm_info_t)
# #define SND_CTL_IOCTL_PCM_PREFER_SUBDEVICE _IOW ('C', 0x54, int)
# #define SND_CTL_IOCTL_PCM_SUBDEVICE _IOW ('C', 0x55, int)
# #define SND_CTL_IOCTL_PCM_SWITCH_LIST _IOWR('C', 0x56, snd_switch_list_t)
# #define SND_CTL_IOCTL_PCM_SWITCH_READ _IOWR('C', 0x57, snd_switch_t)
# #define SND_CTL_IOCTL_PCM_SWITCH_WRITE _IOWR('C', 0x58, snd_switch_t)
# #define SND_CTL_IOCTL_RAWMIDI_CHANNEL _IOW ('C', 0x60, int)
# #define SND_CTL_IOCTL_RAWMIDI_DEVICE _IOW ('C', 0x61, int)
# #define SND_CTL_IOCTL_RAWMIDI_INFO _IOR ('C', 0x62, snd_rawmidi_info_t)
# #define SND_CTL_IOCTL_RAWMIDI_SWITCH_LIST _IOWR('C', 0x63, snd_switch_list_t)
# #define SND_CTL_IOCTL_RAWMIDI_SWITCH_READ _IOWR('C', 0x64, snd_switch_t)
# #define SND_CTL_IOCTL_RAWMIDI_SWITCH_WRITE _IOWR('C', 0x65, snd_switch_t)
# /*****************/
# /*****************/
# /*** MIXER ***/
# /*****************/
# /*****************/
# #define SND_MIXER_VERSION SND_PROTOCOL_VERSION('M',3,0,0)
# #define SND_TYPE_SPEAKER_OUT 0
# #define SND_TYPE_HEADPHONE_OUT 1
# #define SND_TYPE_LINE_OUT 2
# #define SND_TYPE_BLUETOOTH_OUT 3
# #define SND_TYPE_HDMI_OUT 4
# #define SND_TYPE_TOSLINK_OUT 5
# /* PLAYBACK_GROUP_NAMES */
# #define SND_MIXER_AUX_OUT "Aux"
# #define SND_MIXER_CD_OUT "CD"
# #define SND_MIXER_CENTER_OUT "Center"
# #define SND_MIXER_DAC_OUT "DAC"
# #define SND_MIXER_DSP_OUT "DSP"
# #define SND_MIXER_FM_OUT "FM"
# #define SND_MIXER_FRONT_OUT "Front"
# #define SND_MIXER_HEADPHONE_OUT "Headphone"
# #define SND_MIXER_LINE_OUT "Line"
# #define SND_MIXER_MASTER_OUT "Master"
# #define SND_MIXER_MASTER_DIGITAL_OUT "Master Digital"
# #define SND_MIXER_MASTER_MONO_OUT "Master Mono"
# #define SND_MIXER_MIC_OUT "Mic"
# #define SND_MIXER_MONO_OUT "Mono"
# #define SND_MIXER_PCM_OUT "PCM"
# #define SND_MIXER_PCM_OUT_SUBCHN "PCM Subchannel"
# #define SND_MIXER_PCM_OUT_MIXER "PCM Mixer"
# #define SND_MIXER_PCM_OUT_UNIFIED "PCM Unified"
# #define SND_MIXER_PHONE_OUT "Phone"
# #define SND_MIXER_RADIO_OUT "Radio"
# #define SND_MIXER_REAR_OUT "Rear"
# #define SND_MIXER_SIDE_OUT "Side Surr"
# #define SND_MIXER_SPDIF_OUT "S/PDIF"
# #define SND_MIXER_SPEAKER_OUT "PC Speaker"
# #define SND_MIXER_SURROUND_OUT "Surround"
# #define SND_MIXER_SYNTHESIZER_OUT "Synth"
# #define SND_MIXER_VIDEO_OUT "Video"
# #define SND_MIXER_WOOFER_OUT "Woofer"
# /* CAPTURE_GROUP_NAMES */
# #define SND_MIXER_ADC_IN "ADC In"
# #define SND_MIXER_AUX_IN "Aux In"
# #define SND_MIXER_CD_IN "CD In"
# #define SND_MIXER_DSP_IN "DSP In"
# #define SND_MIXER_FM_IN "FM In"
# #define SND_MIXER_LINE_IN "Line In"
# #define SND_MIXER_MIC_IN "Mic In"
# #define SND_MIXER_MONO_IN "Mono In"
# #define SND_MIXER_PCM_IN "PCM In"
# #define SND_MIXER_PCM_IN_SUBCHN "PCM In Subchannel"
# #define SND_MIXER_PHONE_IN "Phone In"
# #define SND_MIXER_RADIO_IN "Radio In"
# #define SND_MIXER_SPDIF_IN "S/PDIF In"
# #define SND_MIXER_SYNTHESIZER_IN "Synth In"
# #define SND_MIXER_VIDEO_IN "Video In"
# #if 1 /* LEGACY GROUP_NAMES from GPL ALSA 0.5.x (DO NOT USE) */
# #define SND_MIXER_IN_AUX SND_MIXER_AUX_OUT
# #define SND_MIXER_IN_CD SND_MIXER_CD_OUT
# #define SND_MIXER_IN_CENTER SND_MIXER_CENTER_OUT
# #define SND_MIXER_IN_DAC SND_MIXER_DAC_OUT
# #define SND_MIXER_IN_DSP SND_MIXER_DSP_OUT
# #define SND_MIXER_IN_FM SND_MIXER_FM_OUT
# #define SND_MIXER_IN_LINE SND_MIXER_LINE_OUT
# #define SND_MIXER_IN_MIC SND_MIXER_MIC_OUT
# #define SND_MIXER_IN_MONO SND_MIXER_MONO_OUT
# #define SND_MIXER_IN_PCM SND_MIXER_PCM_OUT
# #define SND_MIXER_IN_PCM_SUBCHN SND_MIXER_PCM_OUT_SUBCHN
# #define SND_MIXER_IN_PHONE SND_MIXER_PHONE_OUT
# #define SND_MIXER_IN_RADIO SND_MIXER_RADIO_OUT
# #define SND_MIXER_IN_SPDIF SND_MIXER_SPDIF_OUT
# #define SND_MIXER_IN_SPEAKER SND_MIXER_SPEAKER_OUT
# #define SND_MIXER_IN_SURROUND SND_MIXER_SURROUND_OUT
# #define SND_MIXER_IN_SYNTHESIZER SND_MIXER_SYNTHESIZER_OUT
# #define SND_MIXER_IN_VIDEO SND_MIXER_VIDEO_OUT
# #define SND_MIXER_IN_WOOFER SND_MIXER_WOOFER_OUT
# #define SND_MIXER_OUT_CENTER SND_MIXER_CENTER_OUT
# #define SND_MIXER_OUT_DSP SND_MIXER_DSP_OUT
# #define SND_MIXER_OUT_HEADPHONE SND_MIXER_HEADPHONE_OUT
# #define SND_MIXER_OUT_MASTER SND_MIXER_MASTER_OUT
# #define SND_MIXER_OUT_MASTER_DIGITAL SND_MIXER_MASTER_DIGITAL_OUT
# #define SND_MIXER_OUT_MASTER_MONO SND_MIXER_MASTER_MONO_OUT
# #define SND_MIXER_OUT_PHONE SND_MIXER_PHONE_OUT
# #define SND_MIXER_OUT_SURROUND SND_MIXER_SURROUND_OUT
# #define SND_MIXER_OUT_WOOFER SND_MIXER_WOOFER_OUT
# #endif
# /* ELEMENT_NAMES */
# #define SND_MIXER_ELEMENT_ADC "Analog Digital Converter"
# #define SND_MIXER_ELEMENT_CAPTURE "Capture"
# #define SND_MIXER_ELEMENT_DAC "Digital Analog Converter"
# #define SND_MIXER_ELEMENT_PLAYBACK "Playback"
# #define SND_MIXER_ELEMENT_DIGITAL_ACCU "Digital Accumulator"
# #define SND_MIXER_ELEMENT_INPUT_ACCU "Input Accumulator"
# #define SND_MIXER_ELEMENT_MONO_IN_ACCU "Mono In Accumulator"
# #define SND_MIXER_ELEMENT_MONO_OUT_ACCU "Mono Out Accumulator"
# #define SND_MIXER_ELEMENT_OUTPUT_ACCU "Output Accumulator"
# #define SND_MIXER_ELEMENT_INPUT_MUX "Input MUX"
# #define SND_MIXER_ELEMENT_TONE_CONTROL "Tone Control"
# /* SWITCH NAMES */
# #define SND_MIXER_SW_MIC_BOOST "Mic Gain Boost"
# #define SND_MIXER_SW_SIM_STEREO "Simulated Stereo Enhancement"
# #define SND_MIXER_SW_LOUDNESS "Loudness (Bass Boost)"
# #define SND_MIXER_SW_IEC958_INPUT "IEC958 (S/PDIF) Input"
# #define SND_MIXER_SW_IEC958_OUTPUT "IEC958 (S/PDIF) Output"
# /* GROUP NAMES */
# #define SND_MIXER_GRP_ANALOG_LOOPBACK "Analog Loopback"
# #define SND_MIXER_GRP_BASS "Bass"
# #define SND_MIXER_GRP_DIGITAL_LOOPBACK "Digital Loopback"
# #define SND_MIXER_GRP_EFFECT "Effect"
# #define SND_MIXER_GRP_EFFECT_3D "3D Effect"
# #define SND_MIXER_GRP_EQUALIZER "Equalizer"
# #define SND_MIXER_GRP_FADER "Fader"
# #define SND_MIXER_GRP_IGAIN "Input Gain"
# #define SND_MIXER_GRP_MIC_GAIN "Mic Gain"
# #define SND_MIXER_GRP_OGAIN "Output Gain"
# #define SND_MIXER_GRP_TREBLE "Treble"
# #define SND_MIXER_OSS_ALTPCM 1
# #define SND_MIXER_OSS_BASS 2
# #define SND_MIXER_OSS_CD 3
# #define SND_MIXER_OSS_DIGITAL1 4
# #define SND_MIXER_OSS_IGAIN 5
# #define SND_MIXER_OSS_LINE 6
# #define SND_MIXER_OSS_LINE1 7
# #define SND_MIXER_OSS_LINE2 8
# #define SND_MIXER_OSS_LINE3 9
# #define SND_MIXER_OSS_MIC 10
# #define SND_MIXER_OSS_OGAIN 11
# #define SND_MIXER_OSS_PCM 12
# #define SND_MIXER_OSS_PHONEIN 13
# #define SND_MIXER_OSS_PHONEOUT 14
# #define SND_MIXER_OSS_SPEAKER 15
# #define SND_MIXER_OSS_SYNTH 16
# #define SND_MIXER_OSS_TREBLE 17
# #define SND_MIXER_OSS_UNKNOWN 18
# #define SND_MIXER_OSS_VIDEO 19
# #define SND_MIXER_OSS_VOLUME 20
# #define SND_MIXER_ETYPE_INPUT 100
# #define SND_MIXER_ETYPE_ADC 101
# #define SND_MIXER_ETYPE_CAPTURE1 102
# #define SND_MIXER_ETYPE_CAPTURE2 103
# #define SND_MIXER_ETYPE_OUTPUT 104
# #define SND_MIXER_ETYPE_DAC 105
# #define SND_MIXER_ETYPE_PLAYBACK1 106
# #define SND_MIXER_ETYPE_PLAYBACK2 107
# #define SND_MIXER_ETYPE_SWITCH1 200
# #define SND_MIXER_ETYPE_SWITCH2 201
# #define SND_MIXER_ETYPE_SWITCH3 202
# #define SND_MIXER_ETYPE_VOLUME1 203
# #define SND_MIXER_ETYPE_VOLUME2 204
# #define SND_MIXER_ETYPE_ACCU1 205
# #define SND_MIXER_ETYPE_ACCU2 206
# #define SND_MIXER_ETYPE_ACCU3 207
# #define SND_MIXER_ETYPE_MUX1 208
# #define SND_MIXER_ETYPE_MUX2 209
# #define SND_MIXER_ETYPE_TONE_CONTROL1 210
# #define SND_MIXER_ETYPE_3D_EFFECT1 211
# #define SND_MIXER_ETYPE_EQUALIZER1 212
# #define SND_MIXER_ETYPE_PAN_CONTROL1 213
# #define SND_MIXER_ETYPE_PRE_EFFECT1 214
# #define SND_MIXER_VOICE_UNUSED 0
# #define SND_MIXER_VOICE_MONO 1
# #define SND_MIXER_VOICE_LEFT 2
# #define SND_MIXER_VOICE_RIGHT 3
# #define SND_MIXER_VOICE_CENTER 4
# #define SND_MIXER_VOICE_REAR_LEFT 5
# #define SND_MIXER_VOICE_REAR_RIGHT 6
# #define SND_MIXER_VOICE_WOOFER 7
# #define SND_MIXER_VOICE_SURR_LEFT 8
# #define SND_MIXER_VOICE_SURR_RIGHT 9
# typedef struct
# {
# uint16_t voice:15, vindex:1;
# uint8_t reserved[124];
# } snd_mixer_voice_t;
class snd_mixer_eid_t(Structure):
_fields_ = [
('type', int32_t),
('name', c_char * 36),
('index', int32_t),
('reserved', uint8_t * 120), # must be filled with zero
('weight', int32_t), # Reserved used for internal sorting oprations
]
# #define SND_MIXER_EIO_DIGITAL (0x0)
# typedef struct snd_mixer_element_io_info
# {
# uint32_t attrib;
# int32_t voices, voices_over, voices_size;
# snd_mixer_voice_t *pvoices;
# void *pzero; /* align pointers on 64-bits --> point to NULL */
# uint8_t reserved[128]; /* must be filled with zero */
# } snd_mixer_element_io_info;
# typedef struct snd_mixer_element_pcm1_info
# {
# int32_t devices, devices_over, devices_size;
# uint8_t zero[4]; /* align on 64-bits */
# int32_t *pdevices;
# void *pzero; /* align pointers on 64-bits --> point to NULL */
# uint8_t reserved[128]; /* must be filled with zero */
# } snd_mixer_element_pcm1_info;
# typedef struct snd_mixer_element_pcm2_info
# {
# int32_t device, subdevice;
# uint8_t reserved[128]; /* must be filled with zero */
# } snd_mixer_element_pcm2_info;
# typedef struct snd_mixer_element_converter_info
# {
# uint32_t resolution;
# uint8_t reserved[124]; /* must be filled with zero */
# } snd_mixer_element_converter_info;
# #define SND_SW_TYPE_BOOLEAN 1 /* 0 or 1 (enable) */
# typedef struct snd_mixer_element_switch1
# {
# int32_t sw, sw_over, sw_size;
# uint8_t zero[4]; /* align on 64-bits */
# uint32_t *psw;
# void *pzero; /* align pointers on 64-bits --> point to NULL */
# uint8_t reserved[128]; /* must be filled with zero */
# } snd_mixer_element_switch1;
# typedef struct snd_mixer_element_switch2
# {
# uint32_t sw:1;
# uint8_t reserved[124]; /* must be filled with zero */
# } snd_mixer_element_switch2;
# #define SND_MIXER_SWITCH3_FULL_FEATURED 1
# #define SND_MIXER_SWITCH3_ALWAYS_DESTINATION 2
# #define SND_MIXER_SWITCH3_ALWAYS_ONE_DESTINATION 3
# #define SND_MIXER_SWITCH3_ONE_DESTINATION 4
# typedef struct snd_mixer_element_switch3_info
# {
# uint32_t type;
# int32_t voices, voices_over, voices_size;
# snd_mixer_voice_t *pvoices;
# void *pzero; /* align pointers on 64-bits --> point to NULL */
# uint8_t reserved[128]; /* must be filled with zero */
# } snd_mixer_element_switch3_info;
# typedef struct snd_mixer_element_switch3
# {
# int32_t rsw, rsw_over, rsw_size;
# int32_t zero[3];
# uint32_t *prsw;
# void *pzero; /* align pointers on 64-bits --> point to NULL */
# uint8_t reserved[128]; /* must be filled with zero */
# } snd_mixer_element_switch3;
# typedef struct snd_mixer_element_volume1_range
# {
# int32_t min, max;
# int32_t min_dB, max_dB;
# uint8_t reserved[128]; /* must be filled with zero */
# } snd_mixer_element_volume1_range_t;
# typedef struct snd_mixer_element_volume1_info
# {
# int32_t range, range_over, range_size;
# uint8_t zero[4]; /* alignment -- zero fill */
# struct snd_mixer_element_volume1_range *prange;
# uint8_t reserved[128]; /* must be filled with zero */
# } snd_mixer_element_volume1_info;
# typedef struct snd_mixer_element_volume1
# {
# int32_t voices, voices_over, voices_size;
# uint8_t zero[4]; /* align on 64-bits */
# uint32_t *pvoices;
# void *pzero; /* align pointers on 64-bits --> point to NULL */
# uint8_t reserved[128]; /* must be filled with zero */
# } snd_mixer_element_volume1;
# typedef struct snd_mixer_element_volume2_range
# {
# uint8_t reserved[128]; /* must be filled with zero */
# } snd_mixer_element_volume2_range_t;
# typedef struct snd_mixer_element_volume2_info
# {
# int32_t svoices, svoices_over, svoices_size;
# int32_t range, range_over, range_size;
# snd_mixer_voice_t *psvoices;
# struct snd_mixer_element_volume2_range *prange;
# uint8_t reserved[128]; /* must be filled with zero */
# } snd_mixer_element_volume2_info;
# typedef struct snd_mixer_element_volume2
# {
# int32_t avoices, avoices_over, avoices_size;
# uint8_t zero[4]; /* alignment -- zero fill */
# int32_t *pavoices;
# void *pzero; /* align pointers on 64-bits --> point to NULL */
# uint8_t reserved[128]; /* must be filled with zero */
# } snd_mixer_element_volume2;
# typedef struct snd_mixer_element_accu1_info
# {
# int32_t attenuation;
# uint8_t reserved[124]; /* must be filled with zero */
# } snd_mixer_element_accu1_info;
# typedef struct snd_mixer_element_accu2_info
# {
# int32_t attenuation;
# uint8_t reserved[124]; /* must be filled with zero */
# } snd_mixer_element_accu2_info;
# typedef struct snd_mixer_element_accu3_range
# {
# int32_t min, max;
# int32_t min_dB, max_dB;
# uint8_t reserved[128]; /* must be filled with zero */
# } snd_mixer_element_accu3_range_t;
# typedef struct snd_mixer_element_accu3_info
# {
# int32_t range, range_over, range_size;
# uint8_t zero[4]; /* alignment -- zero fill */
# struct snd_mixer_element_accu3_range *prange;
# void *pzero; /* align pointers on 64-bits --> point to NULL */
# uint8_t reserved[128]; /* must be filled with zero */
# } snd_mixer_element_accu3_info;
# typedef struct snd_mixer_element_accu3
# {
# int32_t voices, voices_over, voices_size;
# uint8_t zero[4]; /* alignment -- zero fill */
# int32_t *pvoices;
# void *pzero; /* align pointers on 64-bits --> point to NULL */
# uint8_t reserved[128]; /* must be filled with zero */
# } snd_mixer_element_accu3;
# #define SND_MIXER_MUX1_NONE (0x1)
# typedef struct snd_mixer_element_mux1_info
# {
# uint32_t attrib;
# uint8_t reserved[124]; /* must be filled with zero */
# } snd_mixer_element_mux1_info;
# typedef struct snd_mixer_element_mux1
# {
# int32_t output, output_over, output_size;
# uint8_t zero[4]; /* alignment -- zero fill */
# snd_mixer_eid_t *poutput;
# void *pzero; /* align pointers on 64-bits --> point to NULL */
# uint8_t reserved[128]; /* must be filled with zero */
# } snd_mixer_element_mux1;
# #define SND_MIXER_MUX2_NONE (0x1)
# typedef struct snd_mixer_element_mux2_info
# {
# uint32_t attrib;
# uint8_t reserved[124]; /* must be filled with zero */
# } snd_mixer_element_mux2_info;
# typedef struct snd_mixer_element_mux2
# {
# snd_mixer_eid_t output;
# uint8_t reserved[128]; /* must be filled with zero */
# } snd_mixer_element_mux2;
# #define SND_MIXER_TC1_SW (0x1)
# #define SND_MIXER_TC1_BASS (0x2)
# #define SND_MIXER_TC1_TREBLE (0x4)
# typedef struct snd_mixer_element_tone_control1_info
# {
# uint32_t tc;
# int32_t min_bass, max_bass;
# int32_t min_bass_dB, max_bass_dB;
# int32_t min_treble, max_treble;
# int32_t min_treble_dB, max_treble_dB;
# uint8_t reserved[124]; /* must be filled with zero */
# } snd_mixer_element_tone_control1_info;
# typedef struct snd_mixer_element_tone_control1
# {
# uint32_t tc;
# uint32_t sw:1;
# int32_t treble;
# int32_t bass;
# uint8_t reserved[128]; /* must be filled with zero */
# } snd_mixer_element_tone_control1;
# #define SND_MIXER_EFF1_SW (1<<0)
# #define SND_MIXER_EFF1_MONO_SW (1<<1)
# #define SND_MIXER_EFF1_WIDE (1<<2)
# #define SND_MIXER_EFF1_VOLUME (1<<3)
# #define SND_MIXER_EFF1_CENTER (1<<4)
# #define SND_MIXER_EFF1_SPACE (1<<5)
# #define SND_MIXER_EFF1_DEPTH (1<<6)
# #define SND_MIXER_EFF1_DELAY (1<<7)
# #define SND_MIXER_EFF1_FEEDBACK (1<<8)
# #define SND_MIXER_EFF1_DEPTH_REAR (1<<9)
# typedef struct snd_mixer_element_3d_effect1_info
# {
# uint32_t effect;
# int32_t min_depth, max_depth;
# int32_t min_depth_rear, max_depth_rear;
# int32_t min_wide, max_wide;
# int32_t min_center, max_center;
# int32_t min_volume, max_volume;
# int32_t min_space, max_space;
# int32_t min_delay, max_delay;
# int32_t min_feedback, max_feedback;
# uint8_t reserved[124]; /* must be filled with zero */
# } snd_mixer_element_3d_effect1_info;
# typedef struct snd_mixer_element_3d_effect1
# {
# uint32_t effect;
# uint32_t sw:1;
# uint32_t mono_sw:1;
# int32_t depth;
# int32_t depth_rear;
# int32_t wide;
# int32_t center;
# int32_t volume;
# int32_t space;
# int32_t delay;
# int32_t feedback;
# uint8_t reserved[124]; /* must be filled with zero */
# } snd_mixer_element_3d_effect1;
# #define SND_MIXER_PAN_LEFT_RIGHT 1
# #define SND_MIXER_PAN_FRONT_REAR 2
# #define SND_MIXER_PAN_BOTTOM_UP 3
# typedef struct snd_mixer_element_pan_control1_range
# {
# int32_t pan_type;
# int32_t min, max;
# int32_t min_dB, max_dB;
# uint8_t reserved[124]; /* must be filled with zero */
# } snd_mixer_element_pan_control1_range_t;
# typedef struct snd_mixer_element_pan_control1_info
# {
# int32_t range, range_over, range_size;
# uint8_t zero[4]; /* alignment -- zero fill */
# struct snd_mixer_element_pan_control1_range *prange;
# void *pzero; /* align pointers on 64-bits --> point to NULL */
# uint8_t reserved[128]; /* must be filled with zero */
# } snd_mixer_element_pan_control1_info;
# typedef struct snd_mixer_element_pan_control1
# {
# int32_t pan, pan_over, pan_size;
# uint8_t zero[4]; /* alignment -- zero fill */
# int32_t *ppan;
# void *pzero; /* align pointers on 64-bits --> point to NULL */
# uint8_t reserved[128]; /* must be filled with zero */
# } snd_mixer_element_pan_control1;
# typedef struct snd_mixer_element_pre_effect1_info_item
# {
# uint8_t reserved[128]; /* must be filled with zero */
# } snd_mixer_element_pre_effect1_info_item_t;
# typedef struct snd_mixer_element_pre_effect1_info_parameter
# {
# uint8_t reserved[128]; /* must be filled with zero */
# } snd_mixer_element_pre_effect1_info_parameter_t;
# typedef struct snd_mixer_element_pre_effect1_info
# {
# int32_t items, items_over, items_size;
# int32_t parameters, parameters_over, parameters_size;
# struct snd_mixer_element_pre_effect1_info_item *pitems;
# struct snd_mixer_element_pre_effect1_info_parameter *pparameters;
# uint8_t reserved[128]; /* must be filled with zero */
# } snd_mixer_element_pre_effect1_info;
# typedef struct snd_mixer_element_pre_effect1
# {
# int32_t item;
# int32_t parameters, parameters_over, parameters_size;
# int32_t *pparameters;
# void *pzero; /* align pointers on 64-bits --> point to NULL */
# uint8_t reserved[128]; /* must be filled with zero */
# } snd_mixer_element_pre_effect1;
# typedef struct snd_mixer_element_info
# {
# snd_mixer_eid_t eid;
# union
# {
# snd_mixer_element_io_info io;
# snd_mixer_element_pcm1_info pcm1;
# snd_mixer_element_pcm2_info pcm2;
# snd_mixer_element_converter_info converter;
# snd_mixer_element_switch3_info switch3;
# snd_mixer_element_volume1_info volume1;
# snd_mixer_element_volume2_info volume2;
# snd_mixer_element_accu1_info accu1;
# snd_mixer_element_accu2_info accu2;
# snd_mixer_element_accu3_info accu3;
# snd_mixer_element_mux1_info mux1;
# snd_mixer_element_mux2_info mux2;
# snd_mixer_element_tone_control1_info tc1;
# snd_mixer_element_3d_effect1_info teffect1;
# snd_mixer_element_pan_control1_info pc1;
# snd_mixer_element_pre_effect1_info peffect1;
# uint8_t reserved[128]; /* must be filled with zero */
# } data;
# uint8_t reserved[128]; /* must be filled with zero */
# } snd_mixer_element_info_t;
# typedef struct snd_mixer_element
# {
# snd_mixer_eid_t eid;
# union
# {
# snd_mixer_element_switch1 switch1;
# snd_mixer_element_switch2 switch2;
# snd_mixer_element_switch3 switch3;
# snd_mixer_element_volume1 volume1;
# snd_mixer_element_volume2 volume2;
# snd_mixer_element_accu3 accu3;
# snd_mixer_element_mux1 mux1;
# snd_mixer_element_mux2 mux2;
# snd_mixer_element_tone_control1 tc1;
# snd_mixer_element_3d_effect1 teffect1;
# snd_mixer_element_pan_control1 pc1;
# snd_mixer_element_pre_effect1 peffect1;
# uint8_t reserved[128]; /* must be filled with zero */
# } data;
# uint8_t reserved[128]; /* must be filled with zero */
# } snd_mixer_element_t;
# typedef enum
# {
# SND_MIXER_CHN_FRONT_LEFT,
# SND_MIXER_CHN_FRONT_RIGHT,
# SND_MIXER_CHN_FRONT_CENTER,
# SND_MIXER_CHN_REAR_LEFT,
# SND_MIXER_CHN_REAR_RIGHT,
# SND_MIXER_CHN_WOOFER,
# SND_MIXER_CHN_SURR_LEFT,
# SND_MIXER_CHN_SURR_RIGHT,
# SND_MIXER_CHN_LAST = 31,
# } snd_mixer_channel_t;
# #define SND_MIXER_CHN_MASK_MONO (1<<SND_MIXER_CHN_FRONT_LEFT)
# #define SND_MIXER_CHN_MASK_FRONT_LEFT (1<<SND_MIXER_CHN_FRONT_LEFT)
# #define SND_MIXER_CHN_MASK_FRONT_RIGHT (1<<SND_MIXER_CHN_FRONT_RIGHT)
# #define SND_MIXER_CHN_MASK_FRONT_CENTER (1<<SND_MIXER_CHN_FRONT_CENTER)
# #define SND_MIXER_CHN_MASK_REAR_LEFT (1<<SND_MIXER_CHN_REAR_LEFT)
# #define SND_MIXER_CHN_MASK_REAR_RIGHT (1<<SND_MIXER_CHN_REAR_RIGHT)
# #define SND_MIXER_CHN_MASK_WOOFER (1<<SND_MIXER_CHN_WOOFER)
# #define SND_MIXER_CHN_MASK_SURR_LEFT (1<<SND_MIXER_CHN_SURR_LEFT)
# #define SND_MIXER_CHN_MASK_SURR_RIGHT (1<<SND_MIXER_CHN_SURR_RIGHT)
# #define SND_MIXER_CHN_MASK_STEREO (SND_MIXER_CHN_MASK_FRONT_LEFT|SND_MIXER_CHN_MASK_FRONT_RIGHT)
# #define SND_MIXER_CHN_MASK_4 (SND_MIXER_CHN_MASK_STEREO|SND_MIXER_CHN_MASK_REAR_LEFT|SND_MIXER_CHN_MASK_REAR_RIGHT)
# #define SND_MIXER_CHN_MASK_5_1 (SND_MIXER_CHN_MASK_4|SND_MIXER_CHN_MASK_FRONT_CENTER|SND_MIXER_CHN_MASK_WOOFER)
# #define SND_MIXER_CHN_MASK_7_1 (SND_MIXER_CHN_MASK_5_1|SND_MIXER_CHN_MASK_SURR_LEFT|SND_MIXER_CHN_MASK_SURR_RIGHT)
# #define SND_MIXER_GRPCAP_VOLUME (1<<0)
# #define SND_MIXER_GRPCAP_JOINTLY_VOLUME (1<<1)
# #define SND_MIXER_GRPCAP_MUTE (1<<2)
# #define SND_MIXER_GRPCAP_JOINTLY_MUTE (1<<3)
# #define SND_MIXER_GRPCAP_CAPTURE (1<<4)
# #define SND_MIXER_GRPCAP_JOINTLY_CAPTURE (1<<5)
# #define SND_MIXER_GRPCAP_EXCL_CAPTURE (1<<6)
# #define SND_MIXER_GRPCAP_PLAY_GRP (1<<29)
# #define SND_MIXER_GRPCAP_CAP_GRP (1<<30)
# #define SND_MIXER_GRPCAP_SUBCHANNEL (1<<31)
# #define SND_MIXER_GRP_MAX_VOICES 32
class snd_mixer_gid_t(Structure):
_fields_ = [
('type', int32_t),
('name', c_char * 32),
('index', int32_t),
('reserved', uint8_t * 124), # must be filled with zero
('weight', int32_t), # Reserved used for internal sorting oprations
]
class _mg_volume_names(Structure):
_fields_ = [
('front_left', uint32_t),
('front_right', uint32_t),
('front_center', uint32_t),
('rear_left', uint32_t),
('rear_right', uint32_t),
('woofer', uint32_t),
('reserved', uint8_t * 128),
]
class _union_mg_volume(Union):
_fields_ = [
('values', uint32_t * 32),
('names', _mg_volume_names),
]
class snd_mixer_group_t(Structure):
_fields_ = [
('gid', snd_mixer_gid_t),
('caps', uint32_t),
('channels', uint32_t),
('min', int32_t),
('max', int32_t),
('volume', _union_mg_volume),
('mute', uint32_t),
('capture', uint32_t),
('capture_group', int32_t),
('elements_size', int32_t),
('elements', int32_t),
('elements_over', int32_t),
('pelements', POINTER(snd_mixer_eid_t)),
('change_duration', uint16_t), # milliseconds
('spare', uint16_t),
('min_dB', int32_t),
('max_dB', int32_t),
('reserved', uint8_t * 120), # must be filled with zero
]
# typedef struct snd_mixer_info_s
# {
# uint32_t type;
# uint32_t attrib;
# uint32_t elements;
# uint32_t groups;
# char id[64];
# char name[64];
# uint8_t reserved[128]; /* must be filled with zero */
# } snd_mixer_info_t;
# /* Implied by.../ asound/lib/mixer/mixer.c */
# typedef struct snd_mixer_elements_s
# {
# int32_t elements, elements_size, elements_over;
# uint8_t zero[4]; /* alignment -- zero fill */
# snd_mixer_eid_t *pelements;
# void *pzero; /* align pointers on 64-bits --> point to NULL */
# uint8_t reserved[128]; /* must be filled with zero */
# } snd_mixer_elements_t;
# typedef struct snd_mixer_groups_s
# {
# int32_t groups, groups_size, groups_over;
# uint8_t zero[4]; /* alignment -- zero fill */
# snd_mixer_gid_t *pgroups;
# void *pzero; /* align pointers on 64-bits --> point to NULL */
# uint8_t reserved[128]; /* must be filled with zero */
# } snd_mixer_groups_t;
# typedef struct snd_mixer_routes_s
# {
# snd_mixer_eid_t eid;
# int32_t routes, routes_size, routes_over;
# uint8_t zero[4]; /* alignment -- zero fill */
# snd_mixer_eid_t *proutes;
# void *pzero; /* align pointers on 64-bits --> point to NULL */
# uint8_t reserved[128]; /* must be filled with zero */
# } snd_mixer_routes_t;
# #define SND_MIXER_READ_REBUILD 0
# #define SND_MIXER_READ_ELEMENT_VALUE 1
# #define SND_MIXER_READ_ELEMENT_CHANGE 2
# #define SND_MIXER_READ_ELEMENT_ADD 3
# #define SND_MIXER_READ_ELEMENT_REMOVE 4
# #define SND_MIXER_READ_ELEMENT_ROUTE 5
# #define SND_MIXER_READ_GROUP_VALUE 6
# #define SND_MIXER_READ_GROUP_CHANGE 7
# #define SND_MIXER_READ_GROUP_ADD 8
# #define SND_MIXER_READ_GROUP_REMOVE 9
# typedef struct snd_mixer_read
# {
# int32_t cmd;
# uint8_t zero[4]; /* alignment -- zero fill */
# union
# {
# snd_mixer_eid_t eid;
# snd_mixer_gid_t gid;
# uint8_t reserved[128]; /* must be filled with zero */
# } data;
# uint8_t reserved[128]; /* must be filled with zero */
# } snd_mixer_read_t;
# #define SND_MIXER_IOCTL_PVERSION _IOR ('R', 0x10, int)
# #define SND_MIXER_IOCTL_INFO _IOR ('R', 0x20, snd_mixer_info_t)
# #define SND_MIXER_IOCTL_ELEMENTS _IOWR('R', 0x30, snd_mixer_elements_t)
# #define SND_MIXER_IOCTL_ELEMENT_INFO _IOWR('R', 0x31, snd_mixer_element_info_t)
# #define SND_MIXER_IOCTL_ELEMENT_READ _IOWR('R', 0x32, snd_mixer_element_t)
# #define SND_MIXER_IOCTL_ELEMENT_WRITE _IOWR('R', 0x33, snd_mixer_element_t)
# #define SND_MIXER_IOCTL_GROUPS _IOWR('R', 0x40, snd_mixer_groups_t)
# #define SND_MIXER_IOCTL_GROUP_READ _IOWR('R', 0x41, snd_mixer_group_t)
# #define SND_MIXER_IOCTL_GROUP_WRITE _IOWR('R', 0x42, snd_mixer_group_t)
# #define SND_MIXER_IOCTL_ROUTES _IOWR('R', 0x50, snd_mixer_routes_t)
# #define SND_MIXER_IOCTL_GET_FILTER _IOR ('R', 0x60, snd_mixer_filter_t)
# #define SND_MIXER_IOCTL_SET_FILTER _IOW ('R', 0x61, snd_mixer_filter_t)
# #define SND_MIXER_IOCTL_AUDIO_DUCKING _IOWR('R', 0x62, int)
# typedef struct snd_mixer_filter
# {
# uint32_t enable; /* bitfield of 1<<SND_MIXER_READ_* */
# uint8_t reserved[124]; /* must be filled with zero */
# } snd_mixer_filter_t;
# /*****************/
# /*****************/
# /*** PCM ***/
# /*****************/
# /*****************/
SND_PCM_VERSION = SND_PROTOCOL_VERSION(ord('P'), 3, 0, 0)
# /*
# * Implied by.../ asound/lib/pcm/plugin/block.c(must be small as they
# * index into an array of fd 's)
# */
SND_PCM_CHANNEL_PLAYBACK = 0
SND_PCM_CHANNEL_CAPTURE = 1
SND_PCM_SFMT_U8 = 0
SND_PCM_SFMT_S8 = 1
SND_PCM_SFMT_U16_LE = 2
SND_PCM_SFMT_U16_BE = 3
SND_PCM_SFMT_S16_LE = 4
SND_PCM_SFMT_S16_BE = 5
SND_PCM_SFMT_U24_LE = 6
SND_PCM_SFMT_U24_BE = 7
SND_PCM_SFMT_S24_LE = 8
SND_PCM_SFMT_S24_BE = 9
SND_PCM_SFMT_U32_LE = 10
SND_PCM_SFMT_U32_BE = 11
SND_PCM_SFMT_S32_LE = 12
SND_PCM_SFMT_S32_BE = 13
SND_PCM_SFMT_A_LAW = 14
SND_PCM_SFMT_MU_LAW = 15
SND_PCM_SFMT_IEC958_SUBFRAME_LE = 16
SND_PCM_SFMT_IEC958_SUBFRAME_BE = 17
SND_PCM_SFMT_AC3 = 18
SND_PCM_SFMT_FLOAT_LE = 19
SND_PCM_SFMT_FLOAT_BE = 20
SND_PCM_SFMT_FLOAT64_LE = 22
SND_PCM_SFMT_FLOAT64_BE = 23
SND_PCM_SFMT_IMA_ADPCM = 24
SND_PCM_SFMT_GSM = 25
SND_PCM_SFMT_MPEG = 26
SND_PCM_SFMT_SPECIAL = 27
# #ifdef SND_LITTLE_ENDIAN
# #define SND_PCM_SFMT_U16 SND_PCM_SFMT_U16_LE
# #define SND_PCM_SFMT_S16 SND_PCM_SFMT_S16_LE
# #define SND_PCM_SFMT_U24 SND_PCM_SFMT_U24_LE
# #define SND_PCM_SFMT_S24 SND_PCM_SFMT_S24_LE
# #define SND_PCM_SFMT_U32 SND_PCM_SFMT_U32_LE
# #define SND_PCM_SFMT_S32 SND_PCM_SFMT_S32_LE
# #define SND_PCM_SFMT_IEC958_SUBFRAME SND_PCM_SFMT_IEC958_SUBFRAME_LE
# #define SND_PCM_SFMT_FLOAT SND_PCM_SFMT_FLOAT_LE
# #define SND_PCM_SFMT_FLOAT64 SND_PCM_SFMT_FLOAT64_LE
# #else
# #define SND_PCM_SFMT_U16 SND_PCM_SFMT_U16_BE
# #define SND_PCM_SFMT_S16 SND_PCM_SFMT_S16_BE
# #define SND_PCM_SFMT_U24 SND_PCM_SFMT_U24_BE
# #define SND_PCM_SFMT_S24 SND_PCM_SFMT_S24_BE
# #define SND_PCM_SFMT_U32 SND_PCM_SFMT_U32_BE
# #define SND_PCM_SFMT_S32 SND_PCM_SFMT_S32_BE
# #define SND_PCM_SFMT_IEC958_SUBFRAME SND_PCM_SFMT_IEC958_SUBFRAME_BE
# #define SND_PCM_SFMT_FLOAT SND_PCM_SFMT_FLOAT_BE
# #define SND_PCM_SFMT_FLOAT64 SND_PCM_SFMT_FLOAT64_BE
# #endif
# #define SND_PCM_FMT_U8 (1<<SND_PCM_SFMT_U8)
# #define SND_PCM_FMT_S8 (1<<SND_PCM_SFMT_S8)
# #define SND_PCM_FMT_U16_LE (1<<SND_PCM_SFMT_U16_LE)
# #define SND_PCM_FMT_U16_BE (1<<SND_PCM_SFMT_U16_BE)
# #define SND_PCM_FMT_S16_LE (1<<SND_PCM_SFMT_S16_LE)
# #define SND_PCM_FMT_S16_BE (1<<SND_PCM_SFMT_S16_BE)
# #define SND_PCM_FMT_U24_LE (1<<SND_PCM_SFMT_U24_LE)
# #define SND_PCM_FMT_U24_BE (1<<SND_PCM_SFMT_U24_BE)
# #define SND_PCM_FMT_S24_LE (1<<SND_PCM_SFMT_S24_LE)
# #define SND_PCM_FMT_S24_BE (1<<SND_PCM_SFMT_S24_BE)
# #define SND_PCM_FMT_U32_LE (1<<SND_PCM_SFMT_U32_LE)
# #define SND_PCM_FMT_U32_BE (1<<SND_PCM_SFMT_U32_BE)
# #define SND_PCM_FMT_S32_LE (1<<SND_PCM_SFMT_S32_LE)
# #define SND_PCM_FMT_S32_BE (1<<SND_PCM_SFMT_S32_BE)
# #define SND_PCM_FMT_A_LAW (1<<SND_PCM_SFMT_A_LAW)
# #define SND_PCM_FMT_MU_LAW (1<<SND_PCM_SFMT_MU_LAW)
# #define SND_PCM_FMT_IEC958_SUBFRAME_LE (1<<SND_PCM_SFMT_IEC958_SUBFRAME_LE)
# #define SND_PCM_FMT_IEC958_SUBFRAME_BE (1<<SND_PCM_SFMT_IEC958_SUBFRAME_BE)
# #define SND_PCM_FMT_AC3 (1<<SND_PCM_SFMT_AC3)
# #define SND_PCM_FMT_FLOAT_LE (1<<SND_PCM_SFMT_FLOAT_LE)
# #define SND_PCM_FMT_FLOAT_BE (1<<SND_PCM_SFMT_FLOAT_BE)
# #define SND_PCM_FMT_FLOAT64_LE (1<<SND_PCM_SFMT_FLOAT64_LE)
# #define SND_PCM_FMT_FLOAT64_BE (1<<SND_PCM_SFMT_FLOAT64_BE)
# #define SND_PCM_FMT_IMA_ADPCM (1<<SND_PCM_SFMT_IMA_ADPCM)
# #define SND_PCM_FMT_GSM (1<<SND_PCM_SFMT_GSM)
# #define SND_PCM_FMT_MPEG (1<<SND_PCM_SFMT_MPEG)
# #define SND_PCM_FMT_SPECIAL (1<<SND_PCM_SFMT_SPECIAL)
# #ifdef SND_LITTLE_ENDIAN
# #define SND_PCM_FMT_U16 SND_PCM_FMT_U16_LE
# #define SND_PCM_FMT_S16 SND_PCM_FMT_S16_LE
# #define SND_PCM_FMT_U24 SND_PCM_FMT_U24_LE
# #define SND_PCM_FMT_S24 SND_PCM_FMT_S24_LE
# #define SND_PCM_FMT_U32 SND_PCM_FMT_U32_LE
# #define SND_PCM_FMT_S32 SND_PCM_FMT_S32_LE
# #define SND_PCM_FMT_IEC958_SUBFRAME SND_PCM_FMT_IEC958_SUBFRAME_LE
# #define SND_PCM_FMT_FLOAT SND_PCM_FMT_FLOAT_LE
# #define SND_PCM_FMT_FLOAT64 SND_PCM_FMT_FLOAT64_LE
# #else
# #define SND_PCM_FMT_U16 SND_PCM_FMT_U16_BE
# #define SND_PCM_FMT_S16 SND_PCM_FMT_S16_BE
# #define SND_PCM_FMT_U24 SND_PCM_FMT_U24_BE
# #define SND_PCM_FMT_S24 SND_PCM_FMT_S24_BE
# #define SND_PCM_FMT_U32 SND_PCM_FMT_U32_BE
# #define SND_PCM_FMT_S32 SND_PCM_FMT_S32_BE
# #define SND_PCM_FMT_IEC958_SUBFRAME SND_PCM_FMT_IEC958_SUBFRAME_BE
# #define SND_PCM_FMT_FLOAT SND_PCM_FMT_FLOAT_BE
# #define SND_PCM_FMT_FLOAT64 SND_PCM_FMT_FLOAT64_BE
# #endif
# #define SND_PCM_INFO_PLAYBACK 0x001
# #define SND_PCM_INFO_CAPTURE 0x002
# #define SND_PCM_INFO_DUPLEX 0x010
# #define SND_PCM_INFO_DUPLEX_RATE 0x020
# #define SND_PCM_INFO_DUPLEX_MONO 0x040
# #define SND_PCM_INFO_SHARED 0x100
# #define SND_PCM_INFO_UNSECURE 0x200
# #define SND_PCM_INFO_RESTRICTED 0x400
SND_PCM_MODE_UNKNOWN = 0
SND_PCM_MODE_BLOCK = 1
SND_PCM_MODE_STREAM = 2
SND_PCM_MODE_MASK = 0x0000FFFF
SND_PCM_MODE_FLAG_PROTECTED_CONTENT = (1<<16)
SND_PCM_MODE_FLAG_ENABLE_PROTECTION = (1<<17)
SND_PCM_MODE_FLAG_REQUIRE_PROTECTION = (1<<18)
# #define SND_SRC_MODE_NORMAL 0
# #define SND_SRC_MODE_ACTUAL 1
# #define SND_SRC_MODE_ASYNC 2
# #define SND_SRC_MODE_PITCH 3
# #define SND_PCM_RATE_8000 (1<<1)
# #define SND_PCM_RATE_11025 (1<<2)
# #define SND_PCM_RATE_16000 (1<<3)
# #define SND_PCM_RATE_22050 (1<<4)
# #define SND_PCM_RATE_32000 (1<<5)
# #define SND_PCM_RATE_44100 (1<<6)
# #define SND_PCM_RATE_48000 (1<<7)
# #define SND_PCM_RATE_88200 (1<<8)
# #define SND_PCM_RATE_96000 (1<<9)
# #define SND_PCM_RATE_176400 (1<<10)
# #define SND_PCM_RATE_192000 (1<<11)
# #define SND_PCM_RATE_KNOT (1<<30)
# #define SND_PCM_RATE_CONTINUOUS (1<<31)
# #define SND_PCM_RATE_8000_44100 0x07E
# #define SND_PCM_RATE_8000_48000 0x0FE
# #define SND_PCM_CHNINFO_BLOCK 0x00001
# #define SND_PCM_CHNINFO_STREAM 0x00002
# #define SND_PCM_CHNINFO_MMAP 0x00010
# #define SND_PCM_CHNINFO_INTERLEAVE 0x00020
# #define SND_PCM_CHNINFO_NONINTERLEAVE 0x00040
# #define SND_PCM_CHNINFO_BLOCK_TRANSFER 0x00080
# #define SND_PCM_CHNINFO_PAUSE 0x00100
# #define SND_PCM_CHNINFO_MMAP_VALID 0x00200
# #define SND_PCM_CHNINFO_PROTECTED 0x00400
# #define SND_PCM_CHNINFO_SECURE 0x00800
# #define SND_PCM_CHNINFO_RESTRICTED 0x01000 /* channel data could be muted in circumstances that the driver choose to */
# #define SND_PCM_FILL_NONE 1
# #define SND_PCM_FILL_SILENCE 2
# #define SND_PCM_STATUS_NOTREADY 0
# #define SND_PCM_STATUS_READY 1
# #define SND_PCM_STATUS_PREPARED 2
# #define SND_PCM_STATUS_RUNNING 3
# #define SND_PCM_STATUS_UNDERRUN 4
# #define SND_PCM_STATUS_OVERRUN 5
# #define SND_PCM_STATUS_UNSECURE 6
# #define SND_PCM_STATUS_PAUSED 10
# #define SND_PCM_STATUS_ERROR 10000 /* HW error, need to prepare the stream */
# #define SND_PCM_STATUS_CHANGE 10001 /* stream change, need to param the stream */
# #define SND_PCM_STATUS_PREEMPTED 10002 /* stream was preempted by a higher priority stream */
SND_PCM_START_DATA = 1
SND_PCM_START_FULL = 2
SND_PCM_START_GO = 3
SND_PCM_STOP_STOP = 1
SND_PCM_STOP_ROLLOVER = 2
UINT_MAX = 2 ** (8 * sizeof(c_uint)) - 1
SND_PCM_BOUNDARY = UINT_MAX
SND_PCM_PARAMS_BAD_MODE = 1
SND_PCM_PARAMS_BAD_START = 2
SND_PCM_PARAMS_BAD_STOP = 3
SND_PCM_PARAMS_BAD_FORMAT = 4
SND_PCM_PARAMS_BAD_RATE = 5
SND_PCM_PARAMS_BAD_VOICES = 6
SND_PCM_PARAMS_NO_CHANNEL = 10
class snd_pcm_info_t(Structure):
_fields_ = [
('type', uint32_t), # soundcard type
('flags', uint32_t), # see SND_PCM_INFO_XXXX
('id', uint8_t * 64), # ID of this PCM device
('name', c_char * 80), # name of this device
('playback', int32_t), # playback subdevices=-1
('capture', int32_t), # capture subdevices=-1
('card', int32_t),
('device', int32_t),
('shared_card', int32_t),
('shared_device', int32_t),
('reserved', uint8_t * 128), # must be filled with zero
]
class snd_pcm_sync_t(Union):
_fields_ = [
('id', uint8_t * 16),
('id16', uint16_t * 8),
('id32', uint32_t * 4),
('id64', uint64_t * 2),
]
class snd_pcm_digital_t(Structure):
_fields_ = [
('dig_status', uint8_t * 24), # AES/EBU/IEC958 channel status bits
('dig_subcode', uint8_t * 147), # AES/EBU/IEC958 subcode bits
# Note: supposed to be a bitfield :1
('dig_valid', uint8_t), # must be non-zero to accept these values
('dig_subframe', uint8_t * 4), # AES/EBU/IEC958 subframe bits
('dig_reserved', uint8_t * 128), # must be filled with zero
]
class snd_pcm_channel_info_t(Structure):
_fields_ = [
('subdevice', int32_t), # subdevice number
('subname', c_char * 36), # subdevice name
('channel', int32_t), # channel information
('zero1', int32_t), # filler
('zero2', int32_t * 4), # filler
('flags', uint32_t), # see to SND_PCM_CHNINFO_XXXX
('formats', uint32_t), # supported formats
('rates', uint32_t), # hardware rates
('min_rate', int32_t), # min rate (in Hz)
('max_rate', int32_t), # max rate (in Hz)
('min_voices', int32_t), # min voices
('max_voices', int32_t), # max voices
('max_buffer_size', int32_t), # max buffer size in bytes
('min_fragment_size', int32_t), # min fragment size in bytes
('max_fragment_size', int32_t), # max fragment size in bytes
('fragment_align', int32_t), # align fragment value
('fifo_size', int32_t), # stream FIFO size in bytes
('transfer_block_size', int32_t), # bus transfer block size in bytes
('zero3', uint8_t * 4), # alignment -- zero fill
('dig_mask', snd_pcm_digital_t), # AES/EBU/IEC958 supported bits
('zero4', uint32_t), # filler
('mixer_device', int32_t), # mixer device
('mixer_eid', snd_mixer_eid_t), # mixer element identification
('mixer_gid', snd_mixer_gid_t), # mixer group identification
('reserved', uint8_t * 128), # must be filled with zero
]
# typedef struct snd_pcm_channel_info
# {
# int32_t subdevice; /* subdevice number */
# char subname[36]; /* subdevice name */
# int32_t channel; /* channel information */
# int32_t zero1; /* filler */
# int32_t zero2[4]; /* filler */
# uint32_t flags; /* see to SND_PCM_CHNINFO_XXXX */
# uint32_t formats; /* supported formats */
# uint32_t rates; /* hardware rates */
# int32_t min_rate; /* min rate (in Hz) */
# int32_t max_rate; /* max rate (in Hz) */
# int32_t min_voices; /* min voices */
# int32_t max_voices; /* max voices */
# int32_t max_buffer_size; /* max buffer size in bytes */
# int32_t min_fragment_size; /* min fragment size in bytes */
# int32_t max_fragment_size; /* max fragment size in bytes */
# int32_t fragment_align; /* align fragment value */
# int32_t fifo_size; /* stream FIFO size in bytes */
# int32_t transfer_block_size; /* bus transfer block size in bytes */
# uint8_t zero3[4]; /* alignment -- zero fill */
# snd_pcm_digital_t dig_mask; /* AES/EBU/IEC958 supported bits */
# uint32_t zero4; /* filler */
# int32_t mixer_device; /* mixer device */
# snd_mixer_eid_t mixer_eid; /* mixer element identification */
# snd_mixer_gid_t mixer_gid; /* mixer group identification */
# uint8_t reserved[128]; /* must be filled with zero */
# } snd_pcm_channel_info_t;
class snd_pcm_format_t(Structure):
_fields_ = [
('interleave', uint32_t),
('format', int32_t),
('rate', int32_t),
('voices', int32_t),
('special', int32_t),
('reserved', uint8_t * 124),
]
# typedef struct snd_pcm_voice_conversion
# {
# uint32_t app_voices;
# uint32_t hw_voices;
# uint32_t matrix[32];
# } snd_pcm_voice_conversion_t;
class _cp_stream(Structure):
_fields_ = [
('queue_size', int32_t),
('fill', int32_t),
('max_fill', int32_t),
('reserved', uint8_t * 124),
]
class _cp_block(Structure):
_fields_ = [
('frag_size', int32_t),
('frags_min', int32_t),
('frags_max', int32_t),
('frags_buffered_max', uint32_t), # When set, as many as this fragments will be used in long buffered mode
('reserved', uint8_t * 120), # must be filled with zero
]
class _union_cp(Union):
_fields_ = [
('stream', _cp_stream),
('block', _cp_block),
('reserved', uint8_t * 128), # must be filled with zero
]
class snd_pcm_channel_params_t(Structure):
_fields_ = [
('channel', int32_t),
('mode', int32_t),
('sync', snd_pcm_sync_t), # hardware synchronization ID
('format', snd_pcm_format_t),
('digital', snd_pcm_digital_t),
('start_mode', int32_t),
('stop_mode', int32_t),
('time', int32_t, 1),
('ust_time', int32_t, 1),
('why_failed', uint32_t), # SND_PCM_PARAMS_BAD_???
('buf', _union_cp),
('sw_mixer_subchn_name', c_char * 32), # sw_mixer subchn name override
('cross_core_handle', uint32_t), # Cross-core handle for the audio channel
('reserved', uint8_t * 92), # must be filled with zero
]
class _cs_stream(Structure):
_fields_ = [
('queue_size', int32_t),
('reserved', uint8_t * 124),
]
class _cs_block(Structure):
_fields_ = [
('frag_size', int32_t),
('frags', int32_t),
('frags_min', int32_t),
('frags_max', int32_t),
('max_frag_size', uint32_t), # When set, as many as this fragments will be used in long buffered mode
('reserved', uint8_t * 124), # must be filled with zero
]
class _union_cs(Union):
_fields_ = [
('stream', _cs_stream),
('block', _cs_block),
('reserved', uint8_t * 128), # must be filled with zero
]
class snd_pcm_channel_setup_t(Structure):
_fields_ = [
('channel', int32_t),
('mode', int32_t),
('format', snd_pcm_format_t),
('digital', snd_pcm_digital_t),
('buf', _union_cs),
('msbits_per_sample', int16_t),
('pad1', int16_t),
('mixer_device', int32_t), # mixer device
('mixer_eid', POINTER(snd_mixer_eid_t)), # pcm source mixer element
('mixer_gid', POINTER(snd_mixer_gid_t)), # lowest level mixer group subchn specific
('mmap_valid', uint8_t, 1), # channel can use mmapped access
('mmap_active', uint8_t, 1), # channel is using mmaped transfers
('mixer_card', int32_t), # mixer card
('reserved', uint8_t * 104), # must be filled with zero
]
class snd_pcm_channel_status_t(Structure):
_fields_ = [
('channel', int32_t), # channel information
('mode', int32_t), # transfer mode
('status', int32_t), # channel status-SND_PCM_STATUS_XXXX
('scount', uint32_t), # number of bytes processed from playback/capture start
('stime', timeval), # time when playback/capture was started
('ust_time', uint32_t), # UST time when playback/capture was started
('frag', int32_t), # current fragment
('count', int32_t), # number of bytes in queue/buffer
('free', int32_t), # bytes in queue still free
('underrun', int32_t), # count of underruns (playback) from last status
('overrun', int32_t), # count of overruns (capture) from last status
('overrange', int32_t), # count of ADC (capture) overrange detections from last status
('subbuffered', uint32_t), # bytes sub buffered in the pluggin interface
('reserved', uint8_t * 124), # must be filled with zero
]
# #define QNX_SHM_NAME_LEN (4+8+1+4+1+6) /* "/snd" + 8 pid in hex +
# * "-" + 4 cntr in dex + null + alignment on 64-bit */
# typedef struct snd_pcm_mmap_info_s
# {
# char dmabuf_name[QNX_SHM_NAME_LEN];
# char dmactl_name[QNX_SHM_NAME_LEN];
# int32_t size;
# int32_t ctl_size;
# uint32_t driver_flags; /* SEE ADO_SHMBUF_DMA_???? */
# uint32_t user_flags; /* SEE ADO_SHMBUF_DMA_???? */
# uint8_t reserved[112]; /* must be filled with zero */
# } snd_pcm_mmap_info_t;
# typedef struct
# {
# volatile int32_t status; /* read only */
# volatile uint32_t frag_io; /* read only */
# volatile uint32_t block; /* read only */
# volatile uint32_t expblock; /* read write */
# volatile int32_t voices; /* read only */
# volatile int32_t frag_size; /* read only */
# volatile int32_t frags; /* read only */
# uint8_t reserved[124]; /* must be filled with zero */
# } snd_pcm_mmap_io_status_t;
# typedef struct
# {
# volatile uint32_t number; /* read only */
# volatile int32_t addr; /* read only */
# volatile int32_t voice; /* read only */
# volatile int8_t data; /* read write */
# volatile int8_t io; /* read only */
# uint8_t res[2];
# } snd_pcm_mmap_fragment_t;
# typedef struct
# {
# snd_pcm_mmap_io_status_t status;
# snd_pcm_mmap_fragment_t fragments[0]; /* This array is dynamic. See the mmap_io_status.frags variable for its length. */
# } snd_pcm_mmap_control_t;
# #define SND_PCM_IOCTL_PVERSION _IOR ('A', 0x10, int)
# #define SND_PCM_IOCTL_INFO _IOR ('A', 0x20, snd_pcm_info_t)
# #define SND_PCM_IOCTL_CHANNEL_DRAIN _IO ('A', 0x30)
# #define SND_PCM_IOCTL_CHANNEL_FLUSH _IO ('A', 0x31)
# #define SND_PCM_IOCTL_CHANNEL_GO _IO ('A', 0x32)
# #define SND_PCM_IOCTL_CHANNEL_INFO _IOR ('A', 0x33, snd_pcm_channel_info_t)
# #define SND_PCM_IOCTL_CHANNEL_PARAMS _IOWR('A', 0x34, snd_pcm_channel_params_t)
# #define SND_PCM_IOCTL_CHANNEL_PAUSE _IOW ('A', 0x35, int)
# #define SND_PCM_IOCTL_CHANNEL_PREFER _IO ('A', 0x36)
# #define SND_PCM_IOCTL_CHANNEL_PREPARE _IO ('A', 0x37)
# #define SND_PCM_IOCTL_CHANNEL_SETUP _IOR ('A', 0x38, snd_pcm_channel_setup_t)
# #define SND_PCM_IOCTL_CHANNEL_STATUS _IOR ('A', 0x39, snd_pcm_channel_status_t)
# #define SND_PCM_IOCTL_CHANNEL_PARAM_FIT _IOWR('A', 0x40, snd_pcm_channel_params_t)
# #define SND_PCM_IOCTL_MMAP_INFO _IOR ('A', 0x50, snd_pcm_mmap_info_t)
# #define SND_PCM_IOCTL_SYNC_GO _IOW ('A', 0x60, snd_pcm_sync_t)
# #define SND_PCM_IOCTL_CHANNEL_AUDIOMAN_HANDLE _IOW ('A', 0x70, uint32_t)
# #define SND_PCM_IOCTL_LOGGING _IOWR('A', 0x71, uint32_t)
# #define SND_PCM_IOCTL_NEW_LINK_GROUP _IO ('A', 0x72)
# #define SND_PCM_IOCTL_ADD_LINK_GROUP _IO ('A', 0x73)
# #define SND_PCM_IOCTL_REMOVE_LINK_GROUP _IO ('A', 0x74)
# #define SND_PCM_IOCTL_END_LINK_GROUP _IO ('A', 0x75)
# #define SND_PCM_LB_VERSION SND_PROTOCOL_VERSION('L',3,0,0)
# #define SND_PCM_LB_STREAM_MODE_PACKET 100
# #define SND_PCM_LB_STREAM_MODE_RAW 101
# #define SND_PCM_LB_TYPE_DATA 301
# #define SND_PCM_LB_TYPE_FORMAT 302
# #define SND_PCM_LB_TYPE_POSITION 303
# #define SND_PCM_LB_IOCTL_PVERSION _IOR ('L', 0x10, int)
# #define SND_PCM_LB_IOCTL_FORMAT _IOR ('L', 0x30, snd_pcm_format_t)
# #define SND_PCM_LB_IOCTL_STREAM_MODE _IOW ('L', 0x40, int)
# typedef struct snd_pcm_loopback_header_s
# {
# int32_t size;
# int32_t type;
# uint8_t reserved[128]; /* must be filled with zero */
# } snd_pcm_loopback_header_t;
# typedef struct snd_pcm_loopback_status
# {
# snd_pcm_channel_status_t status;
# uint32_t lost;
# uint8_t reserved[124]; /* must be filled with zero */
# } snd_pcm_loopback_status_t;
# #define SND_PCM_LB_IOCTL_STATUS 212
# /*****************/
# /*****************/
# /*** RAW MIDI ***/
# /*****************/
# /*****************/
# #define SND_RAWMIDI_VERSION SND_PROTOCOL_VERSION('W',3,0,0)
# #define SND_RAWMIDI_CHANNEL_INPUT 0
# #define SND_RAWMIDI_CHANNEL_OUTPUT 1
# #define SND_RAWMIDI_INFO_OUTPUT 0x00000001 /* device is capable
# * rawmidi output */
# #define SND_RAWMIDI_INFO_INPUT 0x00000002 /* device is capable
# * rawmidi input */
# #define SND_RAWMIDI_INFO_DUPLEX 0x00000004 /* device is capable the
# * duplex module */
# typedef struct snd_rawmidi_info
# {
# int32_t type; /* soundcard type */
# uint32_t flags; /* see SND_RAWMIDI_INFO_XXXX */
# uint8_t id[64]; /* ID of this RawMidi device */
# char name[80]; /* name of this RawMidi device */
# uint8_t reserved[128]; /* must be filled with zero */
# } snd_rawmidi_info_t;
# typedef struct snd_rawmidi_params_s
# {
# int32_t channel;
# int32_t size;
# int32_t room;
# int32_t max;
# int32_t min;
# uint8_t reserved[132]; /* must be filled with zero */
# } snd_rawmidi_params_t;
# typedef struct snd_rawmidi_status_s
# {
# int32_t channel;
# int32_t size;
# int32_t count;
# int32_t queue;
# int32_t free;
# int32_t overrun;
# uint8_t reserved[128]; /* must be filled with zero */
# } snd_rawmidi_status_t;
# #define SND_RAWMIDI_IOCTL_PVERSION _IOR ('W', 0x10, int)
# #define SND_RAWMIDI_IOCTL_INFO _IOR ('W', 0x20, snd_rawmidi_info_t)
# #define SND_RAWMIDI_IOCTL_CHANNEL_PARAMS _IOW ('W', 0x30, snd_rawmidi_params_t)
# #define SND_RAWMIDI_IOCTL_CHANNEL_STATUS _IOWR('W', 0x40, snd_rawmidi_status_t)
# #define SND_RAWMIDI_IOCTL_CHANNEL_DRAIN _IOW ('W', 0x50, int)
# #define SND_RAWMIDI_IOCTL_CHANNEL_FLUSH _IOW ('W', 0x51, int)
# /*****************/
# /*****************/
# /** Vector Ops ***/
# /*****************/
# /*****************/
# typedef struct snd_v_args_s
# {
# /* .../ asound/lib/pcm/pcm.c:687 */
# int32_t count;
# uint8_t zero[4]; /* alignment -- zero fill */
# const struct iovec *vector;
# void *pzero; /* align pointers on 64-bits --> point to NULL */
# } snd_v_args_t;
# #define SND_IOCTL_READV _IOW ('K', 0x20, snd_v_args_t)
# #define SND_IOCTL_WRITEV _IOW ('K', 0x30, snd_v_args_t)
SND_PCM_OPEN_PLAYBACK = 0x0001
SND_PCM_OPEN_CAPTURE = 0x0002
SND_PCM_OPEN_DUPLEX = 0x0003
SND_PCM_OPEN_NONBLOCK = 0x1000
PLUGIN_DISABLE_BUFFER_PARTIAL_BLOCKS = 1 << 0
PLUGIN_DISABLE_MMAP = 1 << 1
snd_pcm_find = _func(c_int, c_uint, POINTER(c_int), POINTER(c_int), POINTER(c_int), c_int)
snd_pcm_open = _func(c_int, POINTER(POINTER(snd_pcm_t)), c_int, c_int, c_int)
snd_pcm_open_name = _func(c_int, POINTER(POINTER(snd_pcm_t)), c_char_p, c_int)
snd_pcm_open_preferred = _func(c_int, POINTER(POINTER(snd_pcm_t)), POINTER(c_int), POINTER(c_int), c_int)
snd_pcm_close = _func(c_int, POINTER(snd_pcm_t))
snd_pcm_file_descriptor = _func(c_int, POINTER(snd_pcm_t), c_int)
snd_pcm_nonblock_mode = _func(c_int, POINTER(snd_pcm_t), c_int)
snd_pcm_info = _func(c_int, POINTER(snd_pcm_t), POINTER(snd_pcm_info_t))
snd_pcm_channel_info = _func(c_int, POINTER(snd_pcm_t), POINTER(snd_pcm_channel_info_t))
snd_pcm_channel_params = _func(c_int, POINTER(snd_pcm_t), POINTER(snd_pcm_channel_params_t))
snd_pcm_channel_setup = _func(c_int, POINTER(snd_pcm_t), POINTER(snd_pcm_channel_setup_t))
snd_pcm_channel_status = _func(c_int, POINTER(snd_pcm_t), POINTER(snd_pcm_channel_status_t))
snd_pcm_playback_prepare = _func(c_int, POINTER(snd_pcm_t))
snd_pcm_capture_prepare = _func(c_int, POINTER(snd_pcm_t))
snd_pcm_channel_prepare = _func(c_int, POINTER(snd_pcm_t), c_int)
snd_pcm_playback_go = _func(c_int, POINTER(snd_pcm_t))
snd_pcm_capture_go = _func(c_int, POINTER(snd_pcm_t))
snd_pcm_channel_go = _func(c_int, POINTER(snd_pcm_t), c_int)
snd_pcm_playback_pause = _func(c_int, POINTER(snd_pcm_t))
snd_pcm_capture_pause = _func(c_int, POINTER(snd_pcm_t))
snd_pcm_channel_pause = _func(c_int, POINTER(snd_pcm_t), c_int)
snd_pcm_playback_resume = _func(c_int, POINTER(snd_pcm_t))
snd_pcm_capture_resume = _func(c_int, POINTER(snd_pcm_t))
snd_pcm_channel_resume = _func(c_int, POINTER(snd_pcm_t), c_int)
snd_pcm_playback_drain = _func(c_int, POINTER(snd_pcm_t))
snd_pcm_playback_flush = _func(c_int, POINTER(snd_pcm_t))
snd_pcm_capture_flush = _func(c_int, POINTER(snd_pcm_t))
snd_pcm_channel_flush = _func(c_int, POINTER(snd_pcm_t), c_int)
snd_pcm_transfer_size = _func(ssize_t, POINTER(snd_pcm_t), c_int)
snd_pcm_write = _func(ssize_t, POINTER(snd_pcm_t), c_void_p, size_t)
snd_pcm_read = _func(ssize_t, POINTER(snd_pcm_t), c_void_p, size_t)
# snd_pcm_mmap = _func(c_int, POINTER(snd_pcm_t), c_int, POINTER(POINTER(snd_pcm_mmap_control_t)), POINTER(c_void_p))
snd_pcm_munmap = _func(c_int, POINTER(snd_pcm_t), c_int)
snd_pcm_get_audioman_handle = _func(c_int, POINTER(snd_pcm_t), POINTER(c_uint))
snd_pcm_set_audioman_handle = _func(c_int, POINTER(snd_pcm_t), c_uint)
# snd_pcm_link = _func(c_int, POINTER(snd_pcm_t), POINTER(snd_pcm_t))
# snd_pcm_unlink = _func(c_int, POINTER(snd_pcm_t))
# misc
snd_pcm_format_signed = _func(c_int, c_int)
snd_pcm_format_unsigned = _func(c_int, c_int)
snd_pcm_format_linear = _func(c_int, c_int)
snd_pcm_format_little_endian = _func(c_int, c_int)
snd_pcm_format_big_endian = _func(c_int, c_int)
snd_pcm_format_width = _func(c_int, c_int) # in bits
snd_pcm_build_linear_format = _func(c_int, c_int, c_int, c_int)
snd_pcm_format_size = _func(ssize_t, c_int, size_t)
snd_pcm_get_format_name = _func(c_char_p, c_int)
snd_cards = _func(c_int)
snd_cards_list = _func(c_int, POINTER(c_int), c_int, POINTER(c_int))
snd_card_name = _func(c_int, c_char_p)
snd_card_get_name = _func(c_int, c_int, c_char_p, size_t)
snd_card_get_longname = _func(c_int, c_int, c_char_p, size_t)
snd_ctl_open = _func(c_int, POINTER(POINTER(snd_ctl_t)), c_int)
snd_ctl_open_name = _func(c_int, POINTER(POINTER(snd_ctl_t)), c_char_p)
snd_ctl_close = _func(c_int, POINTER(snd_ctl_t))
snd_ctl_driver_version = _func(c_int, POINTER(snd_ctl_t))
snd_ctl_file_descriptor = _func(c_int, POINTER(snd_ctl_t))
snd_ctl_hw_info = _func(c_int, POINTER(snd_ctl_t), POINTER(snd_ctl_hw_info_t))
snd_ctl_switch_list = _func(c_int, POINTER(snd_ctl_t), POINTER(snd_switch_list_t))
# snd_ctl_switch_read = _func(c_int, POINTER(snd_ctl_t), POINTER(snd_switch_t))
# snd_ctl_switch_write = _func(c_int, POINTER(snd_ctl_t), POINTER(snd_switch_t))
snd_ctl_pcm_info = _func(c_int, POINTER(snd_ctl_t), c_int, POINTER(snd_pcm_info_t))
snd_ctl_pcm_channel_info = _func(c_int, POINTER(snd_ctl_t), c_int, c_int, c_int, POINTER(snd_pcm_channel_info_t))
snd_ctl_pcm_channel_switch_list = _func(c_int, POINTER(snd_ctl_t), c_int, c_int, POINTER(snd_switch_list_t))
snd_ctl_pcm_playback_switch_list = _func(c_int, POINTER(snd_ctl_t), c_int, POINTER(snd_switch_list_t))
snd_ctl_pcm_capture_switch_list = _func(c_int, POINTER(snd_ctl_t), c_int, POINTER(snd_switch_list_t))
# snd_ctl_pcm_channel_switch_read = _func(c_int, POINTER(snd_ctl_t), c_int, c_int, POINTER(snd_switch_t))
# snd_ctl_pcm_playback_switch_read = _func(c_int, POINTER(snd_ctl_t), c_int, POINTER(snd_switch_t))
# snd_ctl_pcm_capture_switch_read = _func(c_int, POINTER(snd_ctl_t), c_int, POINTER(snd_switch_t))
# snd_ctl_pcm_channel_switch_write = _func(c_int, POINTER(snd_ctl_t), c_int, c_int, POINTER(snd_switch_t))
# snd_ctl_pcm_playback_switch_write = _func(c_int, POINTER(snd_ctl_t), c_int, POINTER(snd_switch_t))
# snd_ctl_pcm_capture_switch_write = _func(c_int, POINTER(snd_ctl_t), c_int, POINTER(snd_switch_t))
# snd_ctl_mixer_info = _func(c_int, POINTER(snd_ctl_t), c_int, POINTER(snd_mixer_info_t))
snd_ctl_mixer_switch_list = _func(c_int, POINTER(snd_ctl_t), c_int, POINTER(snd_switch_list_t))
# snd_ctl_mixer_switch_read = _func(c_int, POINTER(snd_ctl_t), c_int, POINTER(snd_switch_t))
#snd_ctl_mixer_switch_write = _func(c_int, POINTER(snd_ctl_t), c_int, POINTER(snd_switch_t))
# snd_ctl_read = _func(c_int, POINTER(snd_ctl_t), POINTER(snd_ctl_callbacks_t))
snd_mixer_open = _func(c_int, POINTER(POINTER(snd_mixer_t)), c_int, c_int)
snd_mixer_open_name = _func(c_int, POINTER(POINTER(snd_mixer_t)), c_char_p)
snd_mixer_close = _func(c_int, POINTER(snd_mixer_t))
snd_mixer_file_descriptor = _func(c_int, POINTER(snd_mixer_t))
# snd_mixer_info = _func(c_int, POINTER(snd_mixer_t), POINTER(snd_mixer_info_t))
# snd_mixer_elements = _func(c_int, POINTER(snd_mixer_t), POINTER(snd_mixer_elements_t))
# snd_mixer_routes = _func(c_int, POINTER(snd_mixer_t), POINTER(snd_mixer_routes_t))
# snd_mixer_groups = _func(c_int, POINTER(snd_mixer_t), POINTER(snd_mixer_groups_t))
# snd_mixer_group_read = _func(c_int, POINTER(snd_mixer_t), POINTER(snd_mixer_group_t))
# snd_mixer_group_write = _func(c_int, POINTER(snd_mixer_t), POINTER(snd_mixer_group_t))
# snd_mixer_element_info = _func(c_int, POINTER(snd_mixer_t), POINTER(snd_mixer_element_info_t))
# snd_mixer_element_read = _func(c_int, POINTER(snd_mixer_t), POINTER(snd_mixer_element_t))
# snd_mixer_element_write = _func(c_int, POINTER(snd_mixer_t), POINTER(snd_mixer_element_t))
# snd_mixer_get_filter = _func(c_int, POINTER(snd_mixer_t), POINTER(snd_mixer_filter_t))
# snd_mixer_set_filter = _func(c_int, POINTER(snd_mixer_t), POINTER(snd_mixer_filter_t))
# snd_mixer_read = _func(c_int, POINTER(snd_mixer_t), POINTER(snd_mixer_callbacks_t))
snd_pcm_plugin_transfer_size = _func(ssize_t, POINTER(snd_pcm_t), c_int, size_t)
snd_pcm_plugin_hardware_size = _func(ssize_t, POINTER(snd_pcm_t), c_int, size_t)
snd_pcm_plugin_info = _func(c_int, POINTER(snd_pcm_t), POINTER(snd_pcm_channel_info_t))
snd_pcm_plugin_set_disable = _func(c_uint, POINTER(snd_pcm_t), c_uint)
snd_pcm_plugin_set_src_method = _func(c_uint, POINTER(snd_pcm_t), c_uint)
snd_pcm_plugin_set_src_mode = _func(c_uint, POINTER(snd_pcm_t), c_uint, c_int)
snd_pcm_plugin_src_max_frag = _func(c_int, POINTER(snd_pcm_t), c_uint)
snd_pcm_plugin_update_src = _func(c_int, POINTER(snd_pcm_t), POINTER(snd_pcm_channel_setup_t), c_int)
snd_pcm_plugin_params = _func(c_int, POINTER(snd_pcm_t), POINTER(snd_pcm_channel_params_t))
snd_pcm_plugin_setup = _func(c_int, POINTER(snd_pcm_t), POINTER(snd_pcm_channel_setup_t))
snd_pcm_plugin_status = _func(c_int, POINTER(snd_pcm_t), POINTER(snd_pcm_channel_status_t))
snd_pcm_plugin_prepare = _func(c_int, POINTER(snd_pcm_t), c_int)
snd_pcm_plugin_playback_drain = _func(c_int, POINTER(snd_pcm_t))
snd_pcm_plugin_flush = _func(c_int, POINTER(snd_pcm_t), c_int)
snd_pcm_plugin_pointer = _func(c_int, POINTER(snd_pcm_t), c_int, POINTER(c_void_p), POINTER(size_t))
snd_pcm_plugin_write = _func(ssize_t, POINTER(snd_pcm_t), c_void_p, size_t)
snd_pcm_plugin_read = _func(ssize_t, POINTER(snd_pcm_t), c_void_p, size_t)
# snd_pcm_plugin_set_voice_conversion = _func(c_int, POINTER(snd_pcm_t), c_int, POINTER(snd_pcm_voice_conversion_t))
# snd_pcm_plugin_get_voice_conversion = _func(c_int, POINTER(snd_pcm_t), c_int, POINTER(snd_pcm_voice_conversion_t))
snd_strerror = _func(c_char_p, c_int)
#----------------------------
# apply argtypes/restype to all functions
#
_register_funcs('libasound.so', globals())
# EOF
|
[
"peter@engcorp.com"
] |
peter@engcorp.com
|
7dee759bea033347ff76f5b4a3103bb226993ca7
|
c5975130be45244b9e163b280cee80d76021be23
|
/Demo_exp_simulation.py
|
cc2a2077e2fac91902a18c599112a0448cc4c9c6
|
[] |
no_license
|
shaunak-desai/Laser-Tracker-simulation
|
ce3fed597b4603e030db2d6846ea5ab1066ab547
|
d3a3dea274c60ce94ccbf8e669383a550832cb2f
|
refs/heads/main
| 2023-03-11T15:21:48.358890
| 2021-02-18T12:28:48
| 2021-02-18T12:28:48
| 340,024,767
| 0
| 0
| null | null | null | null |
UTF-8
|
Python
| false
| false
| 7,010
|
py
|
'''
Simulation for demo experiment with 4 targtes placed at an approrpirate distance to simulate the 4 corners of the
ATHENA WFI CCD detector
'''
import numpy as np
import transformations as t
import transform
import matplotlib.pyplot as plt
import Uncertainities as un
import math as m
from Kabsch_best_fit import best_fit
'''
###############################################################################
'''
def demo_err(window_thickness, window_centring_theta, window_centring_phi, distance, mechanical_un, pressure, glass_index,iterations):
#Initilizing the 6-DOF arrays
X = []
Y = []
Z = []
Roll = []
Pitch = []
Yaw = []
scaling = False
for i in range(iterations):
#Target coordinates in the demo_CCD cordinate frame
CCD_target_1 = np.matrix([[-0.070],[0.070],[0],[1]])
CCD_target_2 = np.matrix([[0.070],[0.070],[0],[1]])
CCD_target_3 = np.matrix([[-0.070],[-0.070],[0],[1]])
CCD_target_4 = np.matrix([[0.070],[-0.070],[0],[1]])
targets_demo_CCD_frame = np.concatenate((CCD_target_1, CCD_target_2, CCD_target_3, CCD_target_4), axis = 1)
#transformation matrix to convert target coordinates from demo_CCD frame to LT frame
demo_CCD_to_LT = np.matrix([[1,0,0,0],[0,1,0,0],[0,0,1,-distance],[0,0,0,1]])
#Conversin of target coordinats to LT frame
Homogeneous_modelled_points = (demo_CCD_to_LT * targets_demo_CCD_frame).T
#removing the fourth dimension for ease of calculations
modelled_points = np.delete(Homogeneous_modelled_points, 3, 1)
#Various sources of errors
LT_err = un.LT_uncertainities(modelled_points, 1).T
Window_err = un.Window_uncertainities(modelled_points, window_thickness, window_centring_theta, window_centring_phi)
#print (Window_err)
#Vaccum_index_err = un.vaccum_pressure(modelled_points, window_centring, glass_index, pressure)
#print(Vaccum_index_err)
Mechanical_err = un.Mechanical_uncertainities(modelled_points, mechanical_un, mechanical_un, mechanical_un)
modelled_points_1 = modelled_points
#Adding mechanical uncertainities of mounting the targets
modelled_points_1 = modelled_points + Mechanical_err
#Converting the modelled points into Laser tracker's spherical coordinate system
spherical_points = transform.cartesian_to_spherical(modelled_points_1)
#Adding the uncertainities from the Laser Tracker
spherical_points = spherical_points + LT_err
#Adding uncertainities from the window
spherical_points = spherical_points + Window_err
#Adding angular shifts due to vaccum from Snell's law
#spherical_points = spherical_points + Vaccum_index_err
#Converting back to the cartesian coordinate system
cartesian_points = transform.spherical_to_cartesian(spherical_points)
Homogeneous_transform = t.superimposition_matrix(modelled_points, cartesian_points, usesvd=True)
Rotation = np.matrix(Homogeneous_transform[0:3, 0:3])
Translation = np.matrix(Homogeneous_transform[0:3,3]).T
#[Rotation, Translation] = best_fit(cartesian_points, modelled_points)
#calculating homogeneous transformation function
#Homogeneous_transform = np.zeros((4,4))
#Homogeneous_transform[:3,:3] = Rotation
#Homogeneous_transform[0,3] = Translation[0]
#Homogeneous_transform[1,3] = Translation[1]
#Homogeneous_transform[2,3] = Translation[2]
#Homogeneous_transform[3,3] = 1
#Finding the euler angles from the homogeneous transformation matrix
euler_angles = np.matrix(t.euler_from_matrix(Homogeneous_transform)).T
for i in range(3):
euler_angles[i,0] = m.degrees(euler_angles[i,0]) * 3600
#Appending strings for 6-DOF values in each iteration
X.append((Translation[0,0])*1e6)
Y.append((Translation[1,0])*1e6)
Z.append((Translation[2,0])*1e6)
Roll.append(euler_angles[0,0])
Pitch.append(euler_angles[1,0])
Yaw.append(euler_angles[2,0])
#calculating the standard deviation for the 6-DOF values
X = np.std(X)
Y = np.std(Y)
Z = np.std(Z)
Roll = np.std(Roll)
Pitch = np.std(Pitch)
Yaw = np.std(Yaw)
return X, Y, Z, Roll, Pitch, Yaw
'''
###############################################################################
'''
X = []
Y = []
Z = []
Roll = []
Pitch = []
Yaw = []
# changing LT distance
for LT_distance in np.linspace(1.7, 2.9, num = 5):
[x, y, z, roll, pitch, yaw] = demo_err(4e-5, 135, 135, LT_distance, 0, 0.01, 1.51289,10000)
X.append(x)
Y.append(y)
Z.append(z)
Roll.append(roll)
Pitch.append(pitch)
Yaw.append(yaw)
'''
#changing pressure in vaccum
for pressure in np.linspace(1e-6, 0.01, num = 100):
[x, y, z, roll, pitch, yaw] = demo_err(0, 0.00016968, 2, 0, pressure, 1.51289, 100)
X.append(x)
Y.append(y)
Z.append(z)
Roll.append(roll)
Pitch.append(pitch)
Yaw.append(yaw)
'''
'''
###############################################################################
'''
distance = np.linspace(1.7, 2.9, num =5)
plt.figure()
plt.plot(distance, X, 'r')
plt.plot(distance, Y, 'b')
plt.plot(distance, Z, 'g')
plt.legend(['X-err', 'Y-err', 'Z-err'], loc = 2)
plt.xlabel('Distance of LT from the Target plane (meters)')
plt.ylabel('Errors ($\mu$m)')
plt.title('Translational error propagation with increasing distance')
plt.grid()
plt.show
plt.figure()
plt.plot(distance, Roll, 'r')
plt.plot(distance, Pitch, 'b')
plt.plot(distance, Yaw, 'g')
plt.legend(['Roll', 'Pitch', 'Yaw'], loc =2)
plt.xlabel('Distance of LT from the Target plane (meters)')
plt.ylabel('Errors (arcecond)')
plt.title('Rotational error propagation with increasing distance')
plt.grid()
plt.show
'''
###############################################################################
'''
'''
pressure = np.linspace(1e-6, 0.01, num = 100)
plt.figure()
plt.plot(pressure, X, 'r')
plt.plot(pressure, Y, 'b')
plt.plot(pressure, Z, 'g')
plt.legend(['X-err', 'Y-err', 'Z-err'], loc = 2)
plt.xlabel('Prressure in vaccum chamber (mBar)')
plt.ylabel('Errors ($\mu$m)')
plt.title('Translational error')
plt.grid()
plt.show
plt.figure()
plt.plot(pressure, Roll, 'r')
plt.plot(pressure, Pitch, 'b')
plt.plot(pressure, Yaw, 'g')
plt.legend(['Roll', 'Pitch', 'Yaw'], loc =2)
plt.xlabel('Prressure in vaccum chamber (mBar)')
plt.ylabel('Errors (arcecond)')
plt.title('Rotational error')
plt.grid()
plt.show
'''
|
[
"noreply@github.com"
] |
shaunak-desai.noreply@github.com
|
c80abae38d1dabb5dfaa1cc1b9606faa528421bd
|
13b72e5c48f5f7213d9a569f699dc1554bc363dd
|
/demo/libdemo/list_git__repos.py
|
35f238aa8d2d69b1030b7d8cfefa92fded15d932
|
[] |
no_license
|
srikanthpragada/PYTHON_02_MAR_2021
|
6997fece4ad627bb767c0bca5a5e166369087e68
|
5dfd0c471378bd22379ac0d66f8785d4d315017b
|
refs/heads/master
| 2023-04-04T20:28:22.738042
| 2021-04-17T14:19:48
| 2021-04-17T14:19:48
| 344,498,123
| 0
| 0
| null | null | null | null |
UTF-8
|
Python
| false
| false
| 362
|
py
|
import requests
user = "srikanthpragada"
response = requests.get(f"https://api.github.com/users/{user}/repos")
if response.status_code != 200:
print(f"Sorry! Could not get details for {user} from github!")
exit()
repos = response.json() # Convert JSON to dict
for repo in repos:
print(repo['name'])
print(repo['description'])
print('-' * 50)
|
[
"srikanthpragada@gmail.com"
] |
srikanthpragada@gmail.com
|
4b1ecbe8bfc1dfb288e7e30b8ba859c26d6a53c9
|
b13ca274b4463c9900840ee6516094b7509b6041
|
/empower/lvapp/lvaphandler.py
|
b5c42204161a1e3cc0f451116cafa63d304a1803
|
[
"Apache-2.0"
] |
permissive
|
imec-idlab/sdn_wifi_manager
|
09d206f2f649aa715752d3c44e011d3f54faf592
|
eda52649f855722fdec1d02e25a28c61a8fbda06
|
refs/heads/master
| 2021-06-23T08:03:22.482931
| 2020-12-03T11:30:10
| 2020-12-03T11:30:10
| 162,106,793
| 0
| 0
|
Apache-2.0
| 2019-03-27T16:23:31
| 2018-12-17T09:33:47
|
Python
|
UTF-8
|
Python
| false
| false
| 3,522
|
py
|
#!/usr/bin/env python3
#
# Copyright (c) 2016 Roberto Riggio
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing,
# software distributed under the License is distributed on an
# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
# KIND, either express or implied. See the License for the
# specific language governing permissions and limitations
# under the License.
"""LVAPs Handerler."""
import tornado.web
import tornado.httpserver
from empower.datatypes.etheraddress import EtherAddress
from empower.restserver.apihandlers import EmpowerAPIHandler
from empower.core.resourcepool import ResourceBlock
from empower.main import RUNTIME
class LVAPHandler(EmpowerAPIHandler):
"""LVAP handler. Used to view LVAPs (controller-wide)."""
HANDLERS = [r"/api/v1/lvaps/?",
r"/api/v1/lvaps/([a-zA-Z0-9:]*)/?"]
def get(self, *args, **kwargs):
""" Get all LVAPs or just the specified one.
Args:
lvap_id: the lvap address
Example URLs:
GET /api/v1/lvaps
GET /api/v1/lvaps/11:22:33:44:55:66
"""
try:
if len(args) > 1:
raise ValueError("Invalid URL")
if not args:
self.write_as_json(RUNTIME.lvaps.values())
else:
lvap = EtherAddress(args[0])
self.write_as_json(RUNTIME.lvaps[lvap])
except KeyError as ex:
self.send_error(404, message=ex)
except ValueError as ex:
self.send_error(400, message=ex)
self.set_status(200, None)
def put(self, *args, **kwargs):
""" Set the WTP for a given LVAP, effectivelly hands-over the LVAP to
another WTP
Args:
lvap_id: the lvap address
Request:
version: the protocol version (1.0)
Example URLs:
PUT /api/v1/lvaps/11:22:33:44:55:66
"""
try:
if len(args) != 1:
raise ValueError("Invalid URL")
request = tornado.escape.json_decode(self.request.body)
if "version" not in request:
raise ValueError("missing version element")
lvap_addr = EtherAddress(args[0])
lvap = RUNTIME.lvaps[lvap_addr]
if "wtp" in request:
wtp_addr = EtherAddress(request['wtp'])
wtp = RUNTIME.wtps[wtp_addr]
lvap.wtp = wtp
elif "blocks" in request:
pool = []
for block in request["blocks"]:
wtp_addr = EtherAddress(block['wtp'])
wtp = RUNTIME.wtps[wtp_addr]
hwaddr = EtherAddress(block['hwaddr'])
channel = int(block['channel'])
band = int(block['band'])
r_block = ResourceBlock(wtp, hwaddr, channel, band)
pool.append(r_block)
lvap.blocks = pool
if "encap" in request:
encap = EtherAddress(request["encap"])
lvap.encap = encap
except KeyError as ex:
self.send_error(404, message=ex)
except ValueError as ex:
self.send_error(400, message=ex)
self.set_status(204, None)
|
[
"pedroisolani@gmail.com"
] |
pedroisolani@gmail.com
|
a2be10df8f5d1ea4680193e91049bd3d406fbf87
|
747ef767341c4e6ecd1cad733fc84781951f2424
|
/pq/2020/p7/C/solC1.py
|
0318815ab20422cb93a5a75f01f4b1e8317f0281
|
[] |
no_license
|
islami00/csc102
|
4cdfd761814bab0d7ef45f2270aa70bf411d233b
|
93c89f7b087ff90c34b508b12d654c42857adbd2
|
refs/heads/main
| 2023-06-19T06:27:47.620859
| 2021-07-15T16:25:10
| 2021-07-15T16:25:10
| 348,317,933
| 0
| 0
| null | null | null | null |
UTF-8
|
Python
| false
| false
| 57
|
py
|
n = [12, 3, 3, 4, 4, 5]
ls = [x for x in n if x % 2 == 0]
|
[
"islam-machine@gmail.com"
] |
islam-machine@gmail.com
|
4ec929b17336e778d3d9066994a98845264bb1ae
|
e168e0dde760fc5fac8f96f8ac9f6d86ba471685
|
/ex41.py
|
84577b13ad6e3f8c529a9cbbc79c903904ce4212
|
[] |
no_license
|
Kalinitch/LPTHW
|
a9db1d61d3497c374670fc38f58123216aa32d01
|
a005cbd09a5bae1cd77a7f1c7777ad277d078510
|
refs/heads/master
| 2021-09-06T11:13:17.876476
| 2018-02-05T21:36:19
| 2018-02-05T21:36:19
| 111,433,628
| 1
| 0
| null | null | null | null |
UTF-8
|
Python
| false
| false
| 2,419
|
py
|
import random
from urllib import urlopen
import sys
WORD_URL = "http://learncodethehardway.org/words.txt"
WORDS = []
PHRASES = {
"class %%%(%%%):":
"Make a class named %%% that is-a %%%.",
"class %%%(object):\n\tdef __init__(self, ***)" :
"class %%% has-a __init__ that takes self and *** parameters.",
"class %%%(object):\n\tdef ***(self, @@@)":
"class %%% has-a function named *** that takes self and @@@ parameters.",
"*** = %%%()":
"Set *** to an instance of class %%%.",
"***.***(@@@)":
"From *** get the *** function, and call it with parameters self, @@@.",
"***.*** = '***'":
"From *** get the *** attribute and set it to '***'."
}
# do they want to drill phrases first
if len(sys.argv) == 2 and sys.argv[1] == "english":
PHRASE_FIRST = True
else:
PHRASE_FIRST = False
#load up the words from the website
for word in urlopen(WORD_URL).readlines():
WORDS.append(word.strip())
def convert(snippet, phrase):
class_names = [w.capitalize() for w in
random.sample(WORDS, snippet.count("%%%"))]
other_names = random.sample(WORDS, snippet.count("***"))
results = []
param_names = []
for i in range(0, snippet.count("@@@")):
param_count = random.randint(1,3)
param_names.append(', '.join(random.sample(WORDS, param_count)))
for sentence in snippet, phrase:
result = sentence[:]
# fake class class_names
for word in class_names:
result = result.replace("%%%", word, 1)
# fake class class_names
for word in other_names:
result = result.replace("***", word, 1)
# fake class class_names
for word in class_names:
result = result.replace("@@@", word, 1)
results.append(result)
return results
# keep going untill they hit CTRL-D
try:
while True:
snippets = PHRASES.keys()
random.shuffle(snippets)
for snippet in snippets:
phrase = PHRASES[snippet]
question, answer = convert(snippet, phrase)
if PHRASE_FIRST:
question, answer = answer, question
print question
raw_input("> ")
print "ANSWER: %s\n\n" % answer
except EOFError:
print "\nBye"
|
[
"31504128+Kalinitch@users.noreply.github.com"
] |
31504128+Kalinitch@users.noreply.github.com
|
f4c8214ad10722f296a55d35be50738047f802c1
|
0ffaf53637a5c2f776c8b65957007c2a59667ca4
|
/access_presentation_data/flow_logic.py
|
bd34acdb4b20e7c38c21d7946681777846884335
|
[] |
no_license
|
henriquekfmaia/LTSimApi
|
a4667de6c871be10c68fdfb1762c78e086ce3ec2
|
f77f5a8dfa44522496bb16cd613c888bdfe134e7
|
refs/heads/master
| 2020-03-20T06:08:29.875501
| 2018-10-23T12:09:02
| 2018-10-23T12:09:02
| 124,909,083
| 0
| 0
| null | null | null | null |
UTF-8
|
Python
| false
| false
| 1,632
|
py
|
import math
import logic.comparer as co
def add_flows(flowA, flowB):
flowA.waterFlow.value = float(flowA.waterFlow.value) + float(flowB.waterFlow.value)
totalMass = float(flowA.massFlow.value) + float(flowB.massFlow.value)
if totalMass == 0:
totalMass = 1
wA = float(flowA.massFlow.value)/totalMass
wB = float(flowB.massFlow.value)/totalMass
flowA.massFlow.value = totalMass
arraySum = 0
for i in range(1, len(flowA.sizeDistribution.value.array)):
flowA.sizeDistribution.value.array[i].value = wA*float(flowA.sizeDistribution.value.array[i].value) + wB*float(flowB.sizeDistribution.value.array[i].value)
arraySum += flowA.sizeDistribution.value.array[i].value
flowA.sizeDistribution.value.array[0].value = 100 - arraySum
return flowA
def restart_flow(flow):
flow.waterFlow.value = 0
flow.massFlow.value = 0
for i in flow.sizeDistribution.value.array:
i.value = 0
flow.sizeDistribution.value.array[0].value = 100
return flow
def is_flow_known(flow):
return(hasattr(flow.waterFlow, 'value') and hasattr(flow.massFlow, 'value') and hasattr(flow.sizeDistribution, 'value'))
#return (flow.waterFlow.value == None or flow.massFlow.value == None or flow.sizeDistribution.value == None)
def get_flow_errors(flowA, flowB):
errors = []
errors.append(co.get_errors_number(flowA.waterFlow.value, flowB.waterFlow.value))
errors.append(co.get_errors_number(flowA.massFlow.value, flowB.massFlow.value))
errors.append(co.get_errors_distribution(flowA.sizeDistribution, flowB.sizeDistribution))
return max(errors)
|
[
"henrique.maia@poli.ufrj.br"
] |
henrique.maia@poli.ufrj.br
|
72207f113df60216c97329d5cb3193a0ab4d1938
|
8edf69b3e354c09fb0950ea7be19e92b20b8d208
|
/pyzkaccess/ctypes.py
|
321cb0062313080fdf84b7beef9c96569ebc2934
|
[
"Apache-2.0"
] |
permissive
|
redjanym/pyzkaccess
|
41f57a42026b67da97f6ba4081f9fbdd0c30813d
|
051236992d5b3f6bdb01fe64b7a787b0c22369a9
|
refs/heads/master
| 2022-12-10T04:24:21.328624
| 2020-09-07T22:11:34
| 2020-09-07T22:11:34
| null | 0
| 0
| null | null | null | null |
UTF-8
|
Python
| false
| false
| 705
|
py
|
"""This module is intended to safety import Windows-specific features
from `ctypes` stdlib module on non-windows platform -- they are
replaced by mock objects. Despite the code which uses that features
becomes partially unoperable in this case, we can import it and
generate documentation for instance
"""
import warnings
from ctypes import * # noqa
from unittest.mock import Mock
try:
from ctypes import WinDLL # noqa
except ImportError:
warnings.warn(
'ctypes.WinDLL is not available on non-Windows system. The code is not functional on '
'current platform, but in order to be able import it we mock WinDLL with '
'unittest.mock.Mock object'
)
WinDLL = Mock()
|
[
"gosha753951@gmail.com"
] |
gosha753951@gmail.com
|
9e4155600c0214f59a5e96c16b9db5ae5e42af68
|
c4a0e8e5707f95c934a82e92a32cf749e35ab8d1
|
/Chapter3/numberprint.py
|
b96674777e5b1719dec3f306c7d998f7b31627a7
|
[] |
no_license
|
Blattarex/PublicFolder
|
2c42c5b760f03f53501c3a28834b91cf1eea721f
|
46b7d57fb92796fb0ac533b12125505c65b35889
|
refs/heads/master
| 2020-04-23T11:47:12.165028
| 2014-02-28T02:04:06
| 2014-02-28T02:04:06
| null | 0
| 0
| null | null | null | null |
UTF-8
|
Python
| false
| false
| 245
|
py
|
file = open("binary.dat", "rb")
size = struct.calcsize("i")
bytes_read = file.read(size)
while bytes_read:
value = struct.unpack("i", bytes_read)
value = value[0]
print(value, end=" ")
bytes_read = file.read(size)
file.close()
|
[
"pidgetmidget@gmail.com"
] |
pidgetmidget@gmail.com
|
bf439e9862b4ae08f44e047b1d51ff58c9ae6f67
|
c6666d0235d1d03ed9a5a2d1a3cfa9ccc9d9e88c
|
/webcore/migrations/0001_initial.py
|
bc167afd1bfb17d3738481c8cc02fc4ac5b3fcf0
|
[] |
no_license
|
boiyelove/savingsensei
|
67dc8a5690c7599dd126159837af6e567852aa73
|
8acd46ebd770b9e18f64e14ff08bfd2ddbcc0edc
|
refs/heads/master
| 2021-10-20T01:32:10.775234
| 2019-02-25T03:27:31
| 2019-02-25T03:27:31
| 172,426,033
| 0
| 0
| null | null | null | null |
UTF-8
|
Python
| false
| false
| 1,812
|
py
|
# -*- coding: utf-8 -*-
# Generated by Django 1.9.7 on 2016-06-20 12:21
from __future__ import unicode_literals
from django.db import migrations, models
class Migration(migrations.Migration):
initial = True
dependencies = [
]
operations = [
migrations.CreateModel(
name='Banner',
fields=[
('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')),
('title', models.CharField(max_length=30)),
('desc', models.CharField(max_length=60)),
('btn_link', models.URLField()),
('btn_title', models.CharField(max_length=18)),
('created', models.DateTimeField(auto_now=True)),
('updated', models.DateTimeField(auto_now_add=True)),
],
),
migrations.CreateModel(
name='Contact',
fields=[
('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')),
('email', models.EmailField(max_length=254)),
('subject', models.CharField(max_length=30)),
('content', models.TextField()),
('created', models.DateTimeField(auto_now=True)),
('updated', models.DateTimeField(auto_now_add=True)),
],
),
migrations.CreateModel(
name='Newsletter',
fields=[
('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')),
('email', models.EmailField(max_length=254)),
('created', models.DateTimeField(auto_now=True)),
('updated', models.DateTimeField(auto_now_add=True)),
],
),
]
|
[
"daahrmmieboiye@gmail.com"
] |
daahrmmieboiye@gmail.com
|
40dbe20a67504c37c5be2abfab99add67569df21
|
585bac463cb1919ac697391ff130bbced73d6307
|
/36_ValidSudoku/solution1.py
|
91569e66f6dee678a57f5dd30000308804bcd22a
|
[] |
no_license
|
llgeek/leetcode
|
ce236cf3d3e3084933a7a4a5e8c7766f7f407285
|
4d340a45fb2e9459d47cbe179ebfa7a82e5f1b8c
|
refs/heads/master
| 2021-01-22T23:44:13.318127
| 2020-03-11T00:59:05
| 2020-03-11T00:59:05
| 85,667,214
| 1
| 0
| null | null | null | null |
UTF-8
|
Python
| false
| false
| 769
|
py
|
class Solution:
def isValidSudoku(self, board: List[List[str]]) -> bool:
seen = set()
for i in range(len(board)):
for j in range(len(board[0])):
val = board[i][j]
if val == '.': continue
if str(val) + 'in row ' + str(i) in seen:
return False
seen.add(str(val) + 'in row ' + str(i))
if str(val) + 'in column ' + str(j) in seen:
return False
seen.add(str(val) + 'in column ' + str(j))
if str(val) + 'in grid ' + str(i // 3) + ' ' + str(j // 3) in seen:
return False
seen.add(str(val) + 'in grid ' + str(i // 3) + ' ' + str(j // 3))
return True
|
[
"angleflycll@gmail.com"
] |
angleflycll@gmail.com
|
7050fce62d11317def18ad9ac77a6d3dfed8819d
|
b4ca970112c5eefeb5863535aefbea8b51ac4daa
|
/app.py
|
73706eb3be17717dc19fcbe4f946b3e5b618c363
|
[] |
no_license
|
ramitsharma94/financial_summarizer
|
59c84d2a670605059840a21c5bac36848a076e42
|
13a75ba5bc648c2ede9d0f8f9595ed7cf8e7c439
|
refs/heads/main
| 2023-07-06T09:22:40.173135
| 2021-08-09T21:56:34
| 2021-08-09T21:56:34
| 314,419,824
| 0
| 0
| null | null | null | null |
UTF-8
|
Python
| false
| false
| 3,622
|
py
|
from flask import Flask, render_template, url_for, request, redirect, jsonify, Response
# from flask_sqlalchemy import SQLAlchemy
import flask_excel as excel
from datetime import datetime
import pandas as pd
from werkzeug.utils import secure_filename
from flask_uploads import UploadSet, configure_uploads, DOCUMENTS, IMAGES
import pandas as pd
import torch
import time
from models_summarizer import predict_summaries, scrape_web_data
from models_QnA import answer
app = Flask(__name__)
docs = UploadSet('datafiles', DOCUMENTS)
app.config['UPLOADED_DATAFILES_DEST'] = 'static/uploads'
configure_uploads(app, docs)
@app.route("/", methods = ['GET', "POST"])
def index():
if(request.method == "POST"):
option = request.form["links"]
if(option == "single_link"):
try:
link = request.form["Enter_Link"]
input_text = scrape_web_data(link)
df2 = pd.DataFrame()
df2["input_text"] = [input_text]
df2["Predicted_Summaries"] = [predict_summaries(input_text)]
df2.drop("input_text", axis =1, inplace = True)
sub_q = "Who is the subject?"
obj_q = "Who is the object?"
df2['Subject_Predicted'] = df2['Predicted_Summaries'].apply(lambda x: answer(x, sub_q))
df2['Object_Predicted'] = df2['Predicted_Summaries'].apply(lambda x: answer(x, obj_q))
html = df2.to_html()
text_file = open("./templates/results.html", "w", encoding = "utf8")
text_file.write(html)
text_file.close()
return render_template("results.html")
except:
return "Please enter the correct link!"
else:
try:
filename = request.files['file']
data = pd.read_excel(filename)
links = data["Link"]
dict_links = {}
for link in links:
if(link not in dict_links):
dict_links[link] = scrape_web_data(link)
else:
dict_links[link] = 0
df2 = pd.DataFrame()
df2["input_text"] = [v for k, v in dict_links.items()]
print("Web Scraping Done. Prediction Start!")
summ = []
for i, text in enumerate(df2["input_text"]):
summ.append(predict_summaries(text))
print("Done: {}".format(i))
df2["Predicted_Summaries"] = summ
df2.drop("input_text", axis =1, inplace = True)
sub_q = "Who is the subject?"
obj_q = "Who is the object?"
df2['Subject_Predicted'] = df2['Predicted_Summaries'].apply(lambda x: answer(x, sub_q))
df2['Object_Predicted'] = df2['Predicted_Summaries'].apply(lambda x: answer(x, obj_q))
## Everything will be written to a html file
html = df2.to_html()
text_file = open("./templates/results.html", "w", encoding = "utf8")
text_file.write(html)
text_file.close()
return render_template("results.html")
except:
return "Either the input link is incorrect or the column name is incorrect!"
else:
return render_template("index.html")
if __name__=="__main__":
app.run(debug = True)
|
[
"noreply@github.com"
] |
ramitsharma94.noreply@github.com
|
f8711cdf3509fa00badbd49057ade5fedad53403
|
94b2834ded570173806e7c49b9bbf0e36d57d2cf
|
/weather_api/users/migrations/0001_initial.py
|
bc1d65974c906974e97b3ce328a099a8eb5de3a8
|
[
"MIT"
] |
permissive
|
dimka2014/weather-api
|
2bb2ef02357ab1f498f7469b3dfe4d46c9be6772
|
14456b87c2da2edf4350f822788e04287d62e744
|
refs/heads/master
| 2021-01-20T05:32:40.930109
| 2017-05-04T11:21:38
| 2017-05-04T11:21:38
| 89,790,449
| 0
| 0
| null | null | null | null |
UTF-8
|
Python
| false
| false
| 2,360
|
py
|
# -*- coding: utf-8 -*-
# Generated by Django 1.10 on 2017-04-28 10:57
from __future__ import unicode_literals
from django.db import migrations, models
import django.utils.timezone
class Migration(migrations.Migration):
initial = True
dependencies = [
('auth', '0008_alter_user_username_max_length'),
]
operations = [
migrations.CreateModel(
name='User',
fields=[
('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')),
('password', models.CharField(max_length=128, verbose_name='password')),
('last_login', models.DateTimeField(blank=True, null=True, verbose_name='last login')),
('is_superuser', models.BooleanField(default=False, help_text='Designates that this user has all permissions without explicitly assigning them.', verbose_name='superuser status')),
('email', models.EmailField(max_length=255, unique=True, verbose_name='email address')),
('is_staff', models.BooleanField(default=False, help_text='Designates whether the user can log into this admin site.', verbose_name='staff status')),
('is_active', models.BooleanField(default=True, help_text='Designates whether this user should be treated as active. Unselect this instead of deleting accounts.', verbose_name='active')),
('date_joined', models.DateTimeField(default=django.utils.timezone.now, verbose_name='date joined')),
('confirmation_token', models.CharField(blank=True, default=None, max_length=32, null=True)),
('reset_password_token', models.CharField(blank=True, default=None, max_length=32, null=True)),
('groups', models.ManyToManyField(blank=True, help_text='The groups this user belongs to. A user will get all permissions granted to each of their groups.', related_name='user_set', related_query_name='user', to='auth.Group', verbose_name='groups')),
('user_permissions', models.ManyToManyField(blank=True, help_text='Specific permissions for this user.', related_name='user_set', related_query_name='user', to='auth.Permission', verbose_name='user permissions')),
],
options={
'swappable': 'AUTH_USER_MODEL',
},
),
]
|
[
"dimabelyaev27@gmail.com"
] |
dimabelyaev27@gmail.com
|
ec4a35b2fe0721b18cf3992900a6e0b2644a3bc7
|
a5525982bbb505a1d0e6ede817938eb87281d447
|
/tests/smoketest/scripts/create_account_list.py
|
9da9abf73638ad74ed295229da77c4bbd3054769
|
[
"MIT",
"LicenseRef-scancode-warranty-disclaimer"
] |
permissive
|
Oskii/Ezira
|
09966cf37da91fa18c7392f5d6acb0b943ae97fc
|
e8e8f0648940298850ff33136fcd5cd2030539d3
|
refs/heads/master
| 2020-03-09T00:59:32.724717
| 2019-04-26T02:19:37
| 2019-04-26T02:19:37
| 128,502,128
| 0
| 0
| null | null | null | null |
UTF-8
|
Python
| false
| false
| 1,563
|
py
|
#!/usr/bin/env python3
"""
Create list of all steem accounts in file.
Usage: create_account_list.py <server_address> <output_filename>
"""
import sys
import json
import requests
def main():
if len( sys.argv ) != 3:
exit( "Usage: create_account_list.py <server_address> <output_filename>" )
url = sys.argv[1] + "/rpc"
print( url )
filename = sys.argv[2]
try: file = open( filename, "w" )
except: exit( "Cannot open file " + filename )
headers = { 'content-type': 'application/json' }
last_account = ""
end = False
accounts_count = 0
while end == False:
request = {
"jsonrpc": "2.0",
"id": 0,
"method": "database_api.list_accounts",
"params": { "start": last_account, "limit": 1000, "order": "by_name" }
}
try:
response = requests.post( url, data=json.dumps(request), headers=headers).json()
accounts = response["result"]["accounts"]
except:
print( "rpc failed for last_account: " + last_account )
print( response )
end = True
continue
if last_account != "":
assert accounts[0]["name"] == last_account
del accounts[0]
if len( accounts ) == 0:
end = True
continue
last_account = accounts[-1]["name"]
accounts_count += len( accounts )
for account in accounts:
file.write( account["name"] + "\n" )
# while end == False
file.close()
print( str(accounts_count) + " accounts")
if __name__ == "__main__":
main()
|
[
"eu-be@tlen.pl"
] |
eu-be@tlen.pl
|
7840c7e3a7666d10832cfc9de216c5f356783730
|
2fbdf22081ea33ad23e246ffd390438b89e4f6b6
|
/PerceptronClassifier/main.py
|
589676e77fcbb3c8cd32e691a25212b9dfab6e31
|
[] |
no_license
|
Vladimir1994/MachineLearning
|
bb13836abc8ea523b10890a9208009ffa4e9db84
|
7fd8a9b10fe97b1df80309e92e6cd6e6c5f75cb6
|
refs/heads/master
| 2021-01-10T15:23:05.173110
| 2016-04-09T21:14:22
| 2016-04-09T21:14:22
| 48,575,524
| 0
| 0
| null | null | null | null |
UTF-8
|
Python
| false
| false
| 1,306
|
py
|
import numpy as np
from sklearn.linear_model import Perceptron
from sklearn.metrics import accuracy_score
from sklearn.preprocessing import StandardScaler
def count_accuracy(train_markers, train_features, test_markers,
test_features, is_standard):
if is_standard:
scaler = StandardScaler()
train_features = scaler.fit_transform(train_features)
test_features = scaler.transform(test_features)
clf = Perceptron()
clf.fit(train_features, train_markers)
predictions = clf.predict(test_features)
accuracy = accuracy_score(test_markers, predictions)
return accuracy
def main():
train_data = np.loadtxt('perceptron-train.csv', delimiter=',')
test_data = np.loadtxt('perceptron-test.csv', delimiter=',')
train_markers = train_data[:, 0]
train_features = train_data[:, 1:]
test_markers = test_data[:, 0]
test_features = test_data[:, 1:]
acc_standard = count_accuracy(train_markers, train_features, test_markers,
test_features, True)
acc_not_standard = count_accuracy(train_markers, train_features,
test_markers, test_features, False)
acc_dif = acc_standard - acc_not_standard
print(acc_dif)
if __name__ == "__main__":
main()
|
[
"vladimir.matveev.1994@mail.com"
] |
vladimir.matveev.1994@mail.com
|
8ac537b07be9198d7e763729701e1d2f9534ce38
|
456f6b8e7a8df0ea3e7c80a183a665f4066ef11d
|
/mulTwoInt.py
|
5eeb78d1e241c012f8e9cd4b081829ed913f8d77
|
[] |
no_license
|
diksha002/TD3_Diksha_Dwijesh
|
441118500865aed4908032b99e04501f0717160b
|
36609fd73f4d7edaf2f0ba38825e3a77932812c5
|
refs/heads/master
| 2021-02-14T19:23:51.276272
| 2020-03-10T20:20:49
| 2020-03-10T20:20:49
| 244,827,955
| 0
| 0
| null | null | null | null |
UTF-8
|
Python
| false
| false
| 599
|
py
|
import sys
def mul(a,b):
multiply= a * b
return multiply
if __name__=="__main__":
if ( len(sys.argv) > 3 ):
print ("Erreur! Entrez que deux valeurs")
elif ( len(sys.argv) == 1 ):
print ("Erreur! Entrez deux valeurs")
a = input("premier valeur: ")
b = input("deuxieme valeur: ")
a = int(a)
b = int(b)
print ( mul(a,b) )
elif ( len(sys.argv) == 2 ):
print ("Erreur! Entrez encore une valeur")
y = input("entrez la deuxieme valeur: ")
x = int( sys.argv[1] )
y = int(y)
print ( mul(a,b) )
else:
x = int( sys.argv[1] )
y = int( sys.argv[2] )
print ( mul(a,b) )
|
[
"dwi.rag123@gmail.com"
] |
dwi.rag123@gmail.com
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.